From Concept to Consistent Multi-View Renders
Instructor:Athanassios Economou, PhD
Team:Chunlan Wang,Yichao Shi,Jiaxiang Luo
Period:2024 Fall
Software:Rhino/Shape Machine/Stable diffusion/Python
The incorporation of artificial intelligence (AI) in architectural design has progressed markedly. Nonetheless, current workflows are constrained. A major issue is the inconsistency in AI-generated renderings from different perspectives of the same architectural design. Moreover, there is the problem of insufficient integration with the conceptual design phase. This paper introduces a unified workflow that combines shape grammars with Stable Diffusion to enhance architectural visualization. The proposed methodology employs a three-stage computational process: first, upon inputting the design parti, it produces diverse and feasible floorplans via shape grammars implemented in the Shape Machine; next, it extrudes these floorplans into three-dimensional models; finally, it renders these models into high-quality, cohesive interior visuals from various perspectives utilizing Stable Diffusion, enhanced with ControlNet and LoRA. The results demonstrate that our autonomous and efficient workflow substantially reduces design and rendering time, while enhancing control and flexibility in design generation. This workflow streamlines the design process from initial concepts to final presentations, improving the quality and consistency of renders relative to conventional AI workflows
The proposed workflow incorporating SG and SD in interior design for apartments
The shape rules used for floor plan configurations generation
The operations of SG in SM: (a) Control flow (b) Initial shapes (c)productions
3d Building model productions with SG and Elefront
Productions from SD based on the outcomes of SG