id |
caadria2024_166 |
authors |
Li, Jinmin, Luo, Yilu, Lu, Shuai, Zhang, Jingyun, Wang, Jun, Guo, Rizen and Wang, ShaoMing |
year |
2024 |
title |
ChatDesign: Bootstrapping Generative Floor Plan Design With Pre-trained Large Language Models |
source |
Nicole Gardner, Christiane M. Herr, Likai Wang, Hirano Toshiki, Sumbul Ahmad Khan (eds.), ACCELERATED DESIGN - Proceedings of the 29th CAADRIA Conference, Singapore, 20-26 April 2024, Volume 1, pp. 99–108 |
doi |
https://doi.org/10.52842/conf.caadria.2024.1.099
|
summary |
Large language models (LLMs) have achieved remarkable success in various domains, revolutionizing tasks such as language translation, text generation, and question-answering. However, generating floor plan designs poses a unique challenge that demands the fulfilment of intricate spatial and relational constraints. In this paper, we propose ChatDesign, an innovative approach that leverages the power of pre-trained LLMs to generate floor plan designs from natural language descriptions, while incorporating iterative modifications based on user interaction. By processing user input text through a pre-trained LLM and utilizing a decoder, we can generate regression parameters and floor plans that are precisely tailored to satisfy the specific needs of the user. Our approach incorporates an iterative refinement process, optimizing the model output by considering the input text and previous results. Throughout these interactions, we employ many strategic techniques to ensure the generated design images align precisely with the user's requirements. The proposed approach is extensively evaluated through rigorous experiments, including user studies, demonstrating its feasibility and efficacy. The empirical results consistently demonstrate the superiority of our method over existing approaches, showcasing its ability to generate floor plans that rival those created by human designer. Our code will be available at https://github.com/THU-Kingmin/ChatDesign. |
keywords |
floor plan generation, large language models, user interactions, automatic design, deep learning, pre-train models |
series |
CAADRIA |
email |
ljm22@mails.tsinghua.edu.cn |
full text |
file.pdf (841,760 bytes) |
references |
Content-type: text/plain
|
Chen, Q., Wu, Q., Tang, R., Wang, Y., Wang, S., & Tan, M. (2020)
Intelligent
, home 3d: Automatic 3d-house design from linguistic descriptions only. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 12625-12634)
|
|
|
|
Ding, M., Yang, Z., Hong, W., Zheng, W., Zhou, C., Yin, D., & Tang, J. (2021)
Cogview: Mastering text-to-image generation via transformers
, Advances in Neural Information Processing Systems, 34, 19822-19835
|
|
|
|
Li, W., Zhang, P., Zhang, L., Huang, Q., He, X., Lyu, S., & Gao, J. (2019)
Ob ject-driven text-to-image synthesis via adversarial training
, Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 12174-12182)
|
|
|
|
Raffel, C., Shazeer, N., Roberts, A., Lee, K., Narang, S., Matena, M., & Liu, P. J. (2020)
Exploring the limits of transfer learning with a unified text-to-text transformer
, The Journal of Machine Learning Research, 21(1), 5485-5551
|
|
|
|
Saharia, C., Chan, W., Saxena, S., Li, L., Whang, J., Denton, E. L., & No rouzi, M. (2022)
Photorealistic text-to-image diffusion models with deep language understanding
, Advances in Neural Information Processing Systems, 35, 36479-36494
|
|
|
|
Wu, W., Fu, X. M., Tang, R., Wang, Y., Qi, Y. H., & Liu, L. (2019)
Predicting and steering performance in architectural materials
, ACM Transactions on Graphics (TOG), 38(6), 1-12
|
|
|
|
last changed |
2024/11/17 22:05 |
|