Discover the SciOpen Platform and Achieve Your Research Goals with Ease.
Search articles, authors, keywords, DOl and etc.
With advancements in digital technology, the field of architectural design has increasingly embraced data and algorithms to enhance design efficiency and quality. Recent advancements in text-to-image (T2I) generation models have enabled the creation of images that correspond to textual descriptions. However, textual descriptions struggle to capture essential style characteristics in style images. In this study, we proposed a method for architectural facade design based on the stable diffusion model (SDM) that combined stylistic images or keywords as input with the structural conditions of content images to generate images with both stylistic and architectural features. By employing the constrastive language-image pre-training (CLIP) image encoder to convert the style image into its initial image embedding and feature extraction from multilayer cross-attention and training optimization to obtain a pretrained image embedding, the proposed method extracts stylistic features from style images and converts them into corresponding embeddings. This process enables the generated images to embody stylistic features and artistic semantic information. Furthermore, the T2I adapter model is employed to use the architectural structure of content images as conditional guidance, thereby ensuring that the generated images exhibit the corresponding structural features. By leveraging these two aspects, the proposed method can decorate architecture with stylistic features from stylistic images while preserving the architectural structure features of content images, resulting in images that reflect the content images after style transformation. Our method is mainly used in architectural design applications. It was capable of generating facade images from flat design drawings, three-dimensional (3D) architectural models, and hand-drawn sketches and has achieved commendable results.
Y. Yang, J. Q. Yang, R. Bao, et al. Corporate relative valuation using heterogeneous multi-modal graph neural network. IEEE Trans Knowl Data Eng, 2021, 35: 211–224.
B. Bölek, O. Tutal, H. Özbaşaran. A systematic review on artificial intelligence applications in architecture. J Des Resilience Archit Plann, 2023, 4: 91–104.
I. Goodfellow, J. Pouget-Abadie, M. Mirza, et al. Generative adversarial networks. Commun ACM, 2020, 63: 139–144.
A. K. Ali, O. J. Lee. Facade style mixing using artificial intelligence for urban infill. Architecture, 2023, 3: 258–269.
L. Zhang, L. Zheng, Y. L. Chen, et al. CGAN-assisted renovation of the styles and features of street facades—A case study of the Wuyi area in Fujian, China. Sustainability, 2022, 14: 16575.
J. Betker, G. Goh, L. Jing, et al. Improving image generation with better captions. Comput Sci, 2023, 2: 8.
The articles published in this open access journal are distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits use, distribution and reproduction in any medium, provided the original work is properly cited.