Uploaded by Jojo Inadventure

ArchiGAN : Artificial Intelligence x Architecture

advertisement
ArchiGAN: Artificial Intelligence x
Architecture
Stanislas Chaillou
AI will soon massively empower architects in their day-to-day practice. This article
provides a proof of concept. The framework used here offers a springboard for
discussion, inviting architects to start engaging with AI, and data scientists to consider
Architecture as a field of investigation. In this article, we summarize a part of our
thesis, submitted at Harvard in May 2019, where Generative Adversarial Neural
Networks (or GANs) get leveraged to design floor plans and entire buildings (Fig. 1).
We believe that a statistical approach to design conception will shape AI’s potential
for Architecture. This approach is less deterministic and more holistic in character.
Rather than using machines to optimize a set of variables, relying on them to extract
significant qualities and mimicking them all along the design process represents a
paradigm shift.
We can unpack floor plan design into three distinct steps:
(I) building footprint massing
(II) program repartition
(III) furniture layout
Each step corresponds to a Pix2Pix GAN-model trained to perform one of the
three tasks above. By nesting these models one after the other, we create an entire
apartment building “generation stack” while allowing for user input at each step.
Additionally, by tackling multi-apartment processing, this project scales beyond the
simplicity of single-family houses.
Beyond the mere development of a generation pipeline, this attempt aims at
demonstrating the potential of GANs for any design process, whereby nesting
GAN models, and allowing user input between them, we try to achieve a back and
forth between humans and machines, between disciplinarian intuition and technical
innovation.
S. Chaillou (B)
Spacemaker, Oslo, Norway
e-mail: stan@spacemaker.ai
© Springer Nature Singapore Pte Ltd. 2020
P. F. Yuan et al. (eds.), Architectural Intelligence,
https://doi.org/10.1007/978-981-15-6568-7_8
117
118
S. Chaillou
Fig. 1 GAN-Generated masterplan. Source Author
1 Representation, Learning, and Framework
Pix2Pix uses a conditional Generative Adversarial Network (cGAN) to learn a
mapping from an input image to an output image. The network consists of two main
pieces, the Generator and the Discriminator. The Generator transforms the input
image to an output image; the Discriminator tries to guess if the image was produced
by the generator or if it is the original image. The two parts of the network challenge
each other resulting in higher quality outputs which are difficult to differentiate from
the original images.
We use this ability to learn image mappings which lets our models learn topological features and space organization directly from floor plan images. We control the
type of information that the model learns by formatting images. As an example, just
showing our model the shape of a parcel and its associated building footprint yields
a model able to create typical building footprints given a parcel’s shape.
We used Christopher Hesse’s implementation of Pix2Pix. Figure 2 displays the
results of a typical training. We show there how one of my GAN-models progressively
learns how to layout rooms and the position of doors and windows in space–also
called fenestration–for a given apartment unit in the sequence in Fig. 2. Although the
initial attempts proved imprecise, the machine builds some form of intuition after
250 iterations.
2 Precedents
The early work of Isola et al. in November 2018 enabling image-to-image translation
with their model Pix2Pix has paved the way for our research.
ArchiGAN: Artificial Intelligence x Architecture
119
Fig. 2 Apartment architectural sequence. Source Author
Zheng and Huang in 2018 first studied floor plan analysis using GAN. The authors
proposed to use GANs for floor plan recognition and generation using Pix2PixHD. [1]
Floor plan images processed by their GAN architecture get translated into programmatic patches of colors. Inversely, patches of colors in their work turn into drawn
rooms. If the user specifies the position of openings and rooms, the network elements
laid out become furniture.
Nathan Peters’ thesis [2] at the Harvard Graduate School of Design in the same
year tackled the possibility of laying out rooms across a single-family home footprint.
Peters’ work turns an empty footprint into programmatic patches of color without
specified fenestration.
Regarding GANs as design assistants, Nono Martinez’ thesis [3] at the Harvard
GSD in 2017 investigated the idea of a loop between the machine and the designer
to refine the very notion of “design process”.
3 Stack and Models
We build upon the previously described precedents to create a three-step generation
stack. As described in Fig. 3, each model of the stack handles a specific task of the
workflow: (I) footprint massing, (II) program repartition, (III) furniture layout.
An architect is able to modify or fine-tune the model’s output between each step,
thereby achieving the expected machine-human interaction.
120
S. Chaillou
Fig. 3 Generation stack in three models. Source Author
3.1 Model I: Footprint
Building footprints significantly define the internal organization of floor plans. Their
shape is heavily conditioned by their surroundings and, more specifically, the shape
of their parcel. Since the design of a housing building footprint can be inferred from
the shape of the piece of land it stands on, we have trained a model to generate
typical footprints, using GIS-data (Geographic Information System) from the city of
Boston. We feed pairs of images to the network during training in a format suitable
for Pix2Pix, displaying the raw parcel (left image) and the same parcel with a given
building drawn over it (right image). We show some typical results in Fig. 4.
Fig. 4 Results of Model I. Source Author
ArchiGAN: Artificial Intelligence x Architecture
121
Fig. 5 Results of model II. Source Author
3.2 Model II: Program
Model II handles repartition and fenestration. The network takes as input the footprint
of a given housing unit produced by Model I, the position of its entrance door (green
square), and the position of the main windows specified by the user (red lines). The
plans used to train the network to derive from a database of 800 + plans of apartments,
properly annotated and given in pairs to the model during training. In the output, the
program encodes rooms using colors while representing the wall structure and its
fenestration using a black patch. Some typical results are displayed in Fig. 5.
3.3 Model III: Furnishing
Finally, Model III tackles the challenge of furniture layout using the output of model
II. This model trains on pairs of images, mapping room programs in color to adequate
furniture layouts. The program retains wall structure and fenestration during image
translation while filling the rooms with relevant furniture, specified by each room’s
program. Figure 6 displays some typical results.
4 UI and Experience
We provide the user with a simple interface for each step throughout our pipeline. On
the left, users can input a set of constraints and boundaries to generate the resulting
122
S. Chaillou
Fig. 6 Results of model III. Source Author
plan on the right. The designer can then iteratively modify the input on the left to
refine the result on the right. Figure 7 showcases this type of interface & process, set
up for Model II.
5 Model Chaining and Apartment Building Generation
We scale the utilization of GANs in this part to the entire apartment building design.
The project uses an algorithm to chain Models I, II, and III, one after the other,
processing multiple units as single images at each step. Figure 8 shows this pipeline.
The challenge of drawing floor plates hosting multiple units marks the difference
between single-family houses and apartment buildings. Strategically, the ability to
control the position of windows and units’ entrances is key to enable unit placement
while ensuring each apartment’s quality. Since Model II takes doors and windows
position as input, the generation stack described above can scale to entire floor plates
generation.
The user is invited to specify the unit split between Model I and Model II, in other
words, specifying how each floor plate divides into apartments and to position each
unit entrance door and windows, as well as potential vertical circulations (staircases,
cores, etc.). The proposed algorithm then feeds each resulting unit to Model II (results
shown in Fig. 9), and then III (result in Fig. 10), to finally reassemble each floor plate
of the initial building. The algorithm finally outputs as individual images, all floor
plates of the generated building.
ArchiGAN: Artificial Intelligence x Architecture
123
Fig. 7 Design iterations using model II interface. Source Author
Fig. 8 Apartment building generation pipeline. Source Author
6 Going Further
If generating standard apartments can be achieved using this technique, pushing the
boundaries of our models is the natural next step. GANs offer remarkable flexibility
to solve seemingly highly constrained problems. In the case of floor plan layout,
partitioning and furnishing the space by hand can be a challenging process as the
124
S. Chaillou
Fig. 9 Model II output, program for each floor plate. Source Author
Fig. 10 Model III output, furnishing of each individual unit. Source Author
footprint changes in dimension and shape. Our models prove to be quite “smart” in
their ability to adapt to changing constraints, as evidenced in Fig. 11.
The ability to control the units’ entrance door and windows position, coupled
with the flexibility of my models, allows us to tackle space planning at a larger scale,
beyond the logic of a single unit. In Fig. 12, we scale our pipeline to entire buildings
generation while investigating our model’s reaction to odd apartment shapes and
contextual constraints.
7 Limitations and Future Improvements
If the above results lay down the premise of GANs’ potential for Architecture, some
clear limitations will drive further investigations in the future.
First, as apartment units stack up in a multi-story building, we cannot guarantee for now the continuity of load-bearing walls from one floor to the next.
Since all the internal structure is laid out differently for each unit, load-bearing walls
ArchiGAN: Artificial Intelligence x Architecture
125
Fig. 11 Flexible building layouts. Source Author
might not be aligned. For now, we consider the façades to be load bearing. However,
the ability to specify load-bearing elements’ position in the input of Model II could
potentially help address this issue.
Additionally, increasing the size of the output layer by obtaining larger
images which offer better definition is a natural next step. We want to deploy the
Pix2Pix HD project developed by NVIDIA in August 2018 to achieve this.
Finally, a major challenge comes from the data format of our outputs. GANs
like Pix2Pix handle only pixel information. The resulting images produced in our
pipeline cannot, for now, be used directly by architects & designers. Transforming
this output from a raster image to a vector format is a crucial step for allowing
the above pipeline to integrate with common tools & practices.
8 Conclusion
AI will soon massively empower architects in their day-to-day practice. As such
potential is about to be demonstrated, our work participates to the proof of concept
while our framework offers a springboard for discussion, inviting architects to
start engaging with AI, and data scientists to consider Architecture as a field of
investigation. However, today, our manifesto could be summarized in four major
points.
Conceptually first, our belief is that a statistical approach to design conception
shapes AI’s potential for Architecture. Its less-deterministic and more-holistic character is undoubtedly a chance for our field. Rather than using machines to optimize a
set of variables, relying on them to extract significant qualities and mimicking them
all along the design process is a paradigm shift.
Second, we are directionally convinced that our ability to design the right pipeline
will condition AI’s success as a new architectural toolset. Our choice for the “Grey
126
S. Chaillou
Fig. 12 GAN-enabled building layouts. Source Author
Boxing” approach, as introduced by Prof. Andrew Witt in Log, will likely secure the
best potential results. This method contrasts with the “black box” approach, that only
allows users to input information upfront, and to get finished design options at the
end of the process, without any control over the successive generation steps. To the
contrary, by breaking out our pipeline into discrete steps, “Grey Boxing” permits the
user to intervene all along the way. His tight control over the machine is his ultimate
guarantee of the design process quality.
Third, we technically believe that the sequential nature of the application will facilitate its manageability and foster its development. The ability to intervene throughout
ArchiGAN: Artificial Intelligence x Architecture
127
the generating process is a fundamental dimension: as each step of the pipeline represents a distinct portion of architectural expertise, each model can be trained independently, opening the way to significant improvements and experimentation in the
near future. Indeed, improving this entire pipeline end-to-end could be a long and
cumbersome task, while amending it step by step remains a manageable process,
within the reach of most architects and engineers in the industry.
Finally, we hope our framework will help address the endless breadth and
complexity of the models to be trained and those used in any generation pipeline.
Tackling parcels-footprint-room split-etc., as we do is one possible approach among,
we believe, a large set of options. To encapsulate the necessary steps of space planning, the key is more the principle than the method. And with the growing availability
of architectural data, we encourage further work and open-minded experimentation.
Far from thinking about AI as the new dogma in Architecture, we conceive this
field as a new challenge, full of potential, and promises. We see here the possibility
for rich results, that will complement our practice and address some blind spots of
our discipline.
References
1. Zheng, H., & Huang, W. (2018). Architectural drawings recognition and generation through
machine learning. Cambridge: MA, ACADIA.
2. Peters, N. (2017). Master thesis: “Enabling alternative architectures: Collaborative frameworks
for participatory design.” Cambridge, MA: Harvard Graduate School of Design.
3. Martinez, N. (2016). Suggestive drawing among human and artificial intelligences. Cambridge,
MA: Harvard Graduate School of Design.
Download