CumInCAD is a Cumulative Index about publications in Computer Aided Architectural Design
supported by the sibling associations ACADIA, CAADRIA, eCAADe, SIGraDi, ASCAAD and CAAD futures

PDF papers
References
id caadria2024_166
authors Li, Jinmin, Luo, Yilu, Lu, Shuai, Zhang, Jingyun, Wang, Jun, Guo, Rizen and Wang, ShaoMing
year 2024
title ChatDesign: Bootstrapping Generative Floor Plan Design With Pre-trained Large Language Models
source Nicole Gardner, Christiane M. Herr, Likai Wang, Hirano Toshiki, Sumbul Ahmad Khan (eds.), ACCELERATED DESIGN - Proceedings of the 29th CAADRIA Conference, Singapore, 20-26 April 2024, Volume 1, pp. 99–108
doi https://doi.org/10.52842/conf.caadria.2024.1.099
summary Large language models (LLMs) have achieved remarkable success in various domains, revolutionizing tasks such as language translation, text generation, and question-answering. However, generating floor plan designs poses a unique challenge that demands the fulfilment of intricate spatial and relational constraints. In this paper, we propose ChatDesign, an innovative approach that leverages the power of pre-trained LLMs to generate floor plan designs from natural language descriptions, while incorporating iterative modifications based on user interaction. By processing user input text through a pre-trained LLM and utilizing a decoder, we can generate regression parameters and floor plans that are precisely tailored to satisfy the specific needs of the user. Our approach incorporates an iterative refinement process, optimizing the model output by considering the input text and previous results. Throughout these interactions, we employ many strategic techniques to ensure the generated design images align precisely with the user's requirements. The proposed approach is extensively evaluated through rigorous experiments, including user studies, demonstrating its feasibility and efficacy. The empirical results consistently demonstrate the superiority of our method over existing approaches, showcasing its ability to generate floor plans that rival those created by human designer. Our code will be available at https://github.com/THU-Kingmin/ChatDesign.
keywords floor plan generation, large language models, user interactions, automatic design, deep learning, pre-train models
series CAADRIA
email ljm22@mails.tsinghua.edu.cn
full text file.pdf (841,760 bytes)
references Content-type: text/plain
Details Citation Select
100%; open Chen, Q., Wu, Q., Tang, R., Wang, Y., Wang, S., & Tan, M. (2020) Find in CUMINCAD Intelligent , home 3d: Automatic 3d-house design from linguistic descriptions only. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 12625-12634)

100%; open Ding, M., Yang, Z., Hong, W., Zheng, W., Zhou, C., Yin, D., & Tang, J. (2021) Find in CUMINCAD Cogview: Mastering text-to-image generation via transformers , Advances in Neural Information Processing Systems, 34, 19822-19835

100%; open Li, W., Zhang, P., Zhang, L., Huang, Q., He, X., Lyu, S., & Gao, J. (2019) Find in CUMINCAD Ob ject-driven text-to-image synthesis via adversarial training , Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 12174-12182)

100%; open Raffel, C., Shazeer, N., Roberts, A., Lee, K., Narang, S., Matena, M., & Liu, P. J. (2020) Find in CUMINCAD Exploring the limits of transfer learning with a unified text-to-text transformer , The Journal of Machine Learning Research, 21(1), 5485-5551

100%; open Saharia, C., Chan, W., Saxena, S., Li, L., Whang, J., Denton, E. L., & No rouzi, M. (2022) Find in CUMINCAD Photorealistic text-to-image diffusion models with deep language understanding , Advances in Neural Information Processing Systems, 35, 36479-36494

100%; open Wu, W., Fu, X. M., Tang, R., Wang, Y., Qi, Y. H., & Liu, L. (2019) Find in CUMINCAD Predicting and steering performance in architectural materials , ACM Transactions on Graphics (TOG), 38(6), 1-12

last changed 2024/11/17 22:05
pick and add to favorite papersHOMELOGIN (you are user _anon_516252 from group guest) CUMINCAD Papers Powered by SciX Open Publishing Services 1.002