Augmented Computational Design: Methodical Application of Artificial Intelligence in Generative Design
Augmented Computational Design: Methodical Application of Artificial Intelligence in Generative Design
Abstract
This chapter presents methodological reflections on the necessity and
utility of artificial intelligence in generative design. Specifically, the chap-
ter discusses how generative design processes can be augmented by AI
to deliver in terms of a few outcomes of interest or performance indica-
tors while dealing with hundreds or thousands of small decisions. The
core of the performance-based generative design paradigm is about mak-
ing statistical or simulation-driven associations between these choices and
consequences for mapping and navigating such a complex decision space.
This chapter will discuss promising directions in Artificial Intelligence for
augmenting decision-making processes in architectural design for mapping
and navigating complex design spaces.
Keywords: Artificial Intelligence, Design Space Exploration, Generative
Design, Augmented Intelligence, Probabilistic Design 1
1 Introduction
The core of the performance-driven computational design is to trace the sensitiv-
ity of variations of some performance indicators to the differences between design
alternatives. Therefore any argument about the utility of AI for performance-
based design must necessarily discuss the representation of such differences, as
explicitly as possible. The existing data models and data representations in the
field of Architecture, Engineering, and Construction (AEC), such as CAD and
BIM are heavily focused on geometrically representing building elements and
facilitating the process of construction management. Unfortunately, the field of
AEC does not currently have a structured discourse based on an explicit repre-
sentation of decision variables and outcomes of interest. Specifically, the notion
of design representation and the idea of data modelling for representing “what
needs to be attained from buildings” is rather absent in the literature.
1. This is the author version of the book chapter “Augmented Computational Design: Me-
thodical Application of Artificial Intelligence in Generative Design.” In Artificial Intelligence
in Performance-Driven Design: Theories, Methods, and Tools Towards Sustainability, edited
by Narjes Abbasabadi and Mehdi Ashayeri. Wiley, 2023
1
This treatise proposes to systematically view the differences between design
alternatives in terms of decision variables, be they spatial or non-spatial. Based
on such an explicit formulation of decision variables, we set forth a framework
for building and utilizing AI in [architectural] generative design processes for
associating decision variables and outcomes of interest as performance indicators
in a reciprocal relationship. This reciprocity is explained in terms of the duality
between two quintessential problems to be addressed in generative design, i.e.
the problem of evaluation of design alternatives (mapping), and the problem of
derivation of design alternatives (navigation).
Starting from an explicit representation of a design space as an ordered
pair of two vectors respectively denoting decision variables and performance in-
dicators, we put forth a mathematical framework for structuring data-driven
approaches to generative design in the field of AEC. This framework highlights
two major types of applications for AI in performance-driven design and their
fusion: those capable of augmenting design evaluation procedures and those
capable of augmenting design derivation procedures. Moreover, we introduce
the reciprocity of “flows” and “manifolds” as an intermediary notion for going
beyond the so-called form-function dichotomy. Discussing these notions neces-
sitates the introduction of a mathematical foundation for the framework rooted
in multi-variate calculus.
The main advantage of this explicit formulation is to enhance the explain-
ability of AI when utilized in generative design by introducing meaningful and
interpretable latent spaces based on the reciprocal relationship between mani-
folds and flows. The balance of predictive/deterministic power and interpretabil-
ity/explainability is discussed in the concrete context of an illustrative example.
A chain of key concepts will be introduced in this chapter, starting from the
notion of decision-making in design, the nature of design variables, the specifics
of spatial decision variables, the notion of design space, and the two dual actions
in the exploration of design spaces: mapping and navigating.
Whilst the introduced framework is quite general, a particular class of Prob-
abilistic Graphical Models (PGM), Bayesian Belief Networks (BBN), is intro-
duced to provide a concrete illustrative example of the utility of AI in AEC. For
a deeper insight into this particular approach to data-driven design, the read-
ers are referred to the two classical books about PGM: Pearl (1988) & Koller
and Friedman (2009). The illustrative example is a BBN trained for making a
data-driven replica of the building energy model used by the Dutch government
in order to obtain a rough meta-model to be used in mass-scale policy analysis,
e.g. for advising the government on the relative utility of energy transition sub-
sidies and planning measures. This example is chosen not because the BBNs are
the most advanced models or the most accurate models for approximating such
large functions. The choice is rather pragmatic in that this model has proven to
be promising from the stance of predictive power while retaining a basic level
of theoretical interpretability and intuitive appeal.
The chapter is structured as follows: we first present a historical context
to establish the necessity of such a data-driven generative design framework;
continue with conceptualizing and mathematically formulating the structure of
the framework, dubbed as Augmented Computational Design (ACD); present
an illustrative example demonstrating the utility of the framework; and con-
clude with a discussion on its outlook, open questions, and avenues for further
research.
2
2 Background
Here we revisit the utility of AI for data-driven generative design by highlighting
some key gaps of knowledge in the field of AEC and briefly mentioning overar-
ching frameworks in computational design and AI that can be used to address
these gaps.
3
forms the basis of the generative design paradigm as extensively articulated by
Nourian, Azadi, and Oval (2023) & Veloso and Krishnamurti (2021). Similarly,
the challenges, opportunities, and promising ways of utilization of AI (particu-
larly deep-learning and generative models) for goal-oriented design explorations
have been discussed extensively in (Regenwetter, Nobari, and Ahmed 2022) &
(Regenwetter and Ahmed 2022).
4
Morphogenesis 1952
Alan Turing, Artificial Intelligence Paul Werbos
Mathematics, Back Propagation
Computer-Science 1974
+ Theoretical Biology Machine Learning
GPU
Any Computing
Technology that
Soft Computing
enables Statistical
mimicing Methods
human Enabling AlphaGo 2017
behaviour Learning Fuzzy, Neural,
Deep Learning
without Explicit Evolutionary, &
Programming Probabilistic Multi-Layered
Computing Artificial Neural
Networks
Jeff Hinton,
50
60
70
80
90
00
10
0
AlexNet, CNN,
2
19
19
19
19
19
20
20
20
Demis Hassabis,
AlphaGo, DRL...
Early AI, focused on
Building Expert
Systems, Symbolic Slow Take Off of Machine
AI, Prolog Learning after the AI Winter,
Aliases: Knowledge-Based
PyTorch
Systems, Computational
Intelligence
Accord.NET
TensorFlow
Scikit-Learn
NumPy
5
High-level Task Description Interpolation & Extrapolation Policy Formulation & Prescription
Low-level Task Description Clustering Classification Regression Dim. Reduction State Space Search
Technical Purpose Function Approximation with Imprecision & Uncertainty Optimization & Control
Collective Intelligence
&
Natural Language Processing Probabilistic
Dimensionality Reduction
(NLP) Graphical Models
6
Decision Space Performance Space Performance Space Decision Space
Mapping Navigating
of from
Choices Consequences
to to
Consequences Choices
7
A Spectrum of Generative Design Methodologies & Generative Models for Content Generation
Formulation of Design Rules Mathematical Derivation Devising Design Games Systematic Data Collection
Computational Design
as from & &
Practice
Topology Rewriting Methods Governing Equations Scoring Mechanisms Black-Box Training
Rule-Based Models White Box Models Grey Box Models Black Box Models
8
3 Framework
The emphasis on the decision-making approach to design entails that design
tasks can be formulated as a set of [typically unstructured] questions about the
form and materialization/construction of an object (a building) to be answered.
In this chapter, we focus on the questions that pertain to form.
We propose a mathematical framework for generative design that relates
multiple strands of work together. We use design space for referring to an or-
dered pair of two vector spaces: a decision space containing vectors or data
points representing design configurations in the form of x ∈ (0, 1]n and a perfor-
mance/quality space containing vectors or data-point representing combinations
of outcomes of interest in the form of o ∈ [0, 1]q .
The mathematical lens that we shall put on the issue is to redefine both of
these notions to provide a much more specific and workable idea for discussing
the utility or futility of applying AI to design problem-solving. It is hopefully
easy for the reader to accept that a regular discretization of a so-called design
space (which is an unfortunately common misnomer, but here somewhat prag-
matically useful) provides a straightforward and simple discretization of design
decisions as vectors in the form of x := [xi ]n×1 ∈ (0, 1]n or x ∈ {0, 1}n , where
n is the number of discrete cells in the design space in which virtually any
conceivable shape can be constructed at a certain level of resolution. Without
loss of generality, however, the decision variables are not necessarily spatial and
can be assumed to be relativized float variables in the range of minimum and
maximum admissible parameter values of the functions that together result in
the shape and configuration of a building. Even if a multi-colour (multi-label,
multi-functional) space is the subject of the design problem, then multiple cat-
egories/colours of such vectors can be seen together as a matrix of decision
variables, whose rows have to add up to 1 (see Figure 5).
Once this terminology is established, it is easy to observe that, in Machine
Learning (ML) terms, the problem of performance-based design can be seen
as two problems that are dual to one another, a multi-variate regression prob-
lem for figuring out an approximation function that can map a few outputs
to many inputs (here we call this mapping or the evaluation problem), and a
pseudo-inversion problem for finding the combination of inputs that could re-
sult in desired output data points (here we call this navigation or the derivation
problem), see Figure 3.
When approached as a data-driven problem-solving task, both problems are
somewhat hard and impossible to solve in the absolute sense of the word, unless
we think about them as loss minimization or approximation problems. The nav-
igation/design problem is much harder than the mapping/evaluation problem.
The main idea here is to advocate for training (fitting) meta-models (neural
networks) to sets of sampled pairs of inputs and outputs to first approximate a
complex design space as a map between decision data points and performance
data points, and then find the pseudo-inverse of this map or navigate it in the
reverse direction to be able to find designs (decision data points) that perform
in a desired way. In other words, mathematically, we look at the performance-
based design process as a pairing between a decision space and a performance
space, where a map is conceptualized as a function f : (0, 1]n 7→ [0, 1]q ) such
that o = f (x). The pseudo-inverse map is thus dubbed as f −1 : [0, 1]q 7→ (0, 1]n ,
such that x = f −1 (o).
9
For the sake of brevity and also generalizability to non-spatial design prob-
lems, we shall focus on massing problems and leave coloured configuration prob-
lems out of the picture momentarily (see Figure 5 for the distinction).
Furthermore, by considering two abstract and high-level descriptions of a
design task in our proposed regular discretization frameworks, we can formulate
two mathematical tasks:
• Mapping Design Spaces: approximating the function that can model the
associations between the many input design variables and a few outcomes
of interest;
• Navigating Design Spaces: approximating the inverse function that can
guide the generation of valid configurations in the decision space given
desired data points in the performance space.
10
The Jacobian approximation is numerically computable provided the under-
lying function is smooth and differentiable. For brevity, as commonly done, we
have omitted the fact that the Jacobian can be evaluated at a certain input
data point and that it is expected to be the best linear approximation of the
function in question in the vicinity of that point. If we abbreviate the notation
for the Jacobian as such a functional, then we can denote the approximate lin-
ear function at any given data point as follows (as a first-order Taylor Series
expansion): o(x)|x∼xo ≃ J(x0 ) (x − x0 ), or simply put, as o ≃ Jx, if we assume
x to represent the vector of difference between the input data point with the
centre of the neighbourhood.
The Jacobian approximation is also illuminating for another important rea-
son: it allows us to approximate the Jacobian in a different sense, i.e. in the
sense of dimensionality/rank reduction using the Singular Value Decomposition
(SVD) to see a clearer picture of the main factors playing the most significant
roles in attaining the outcomes of interest, in other words identifying the inputs
variables to which the outcomes of interest are most sensitive. Even though
we do not explicitly perform this operation in our demonstrative example us-
ing the SVD, it is still illuminating to see what SVD can do for this insightful
approximation and dimensionality reduction for two reasons:
• The SVD approximation of the Jacobian allows us to make a cognitive and
interpretable map of the most important causes of the effects of interest
• The SVD approximation of the Jacobian allows us to conceptualize a
pseudo-inverse function to navigate the design space from the side of per-
formance data points.
The SVD (low-rank) approximation of the Jacobian Matrix can be denoted
as below:
J := UΣVT , (1)
where Uq×q := [uk ]1×q and Vn×n := [vi ]1×n are orthogonal matrices (i.e.
UUT = Iq×q and VVT = In×n ), and Σ is a matrix of size q × n with only
p = min{q, n} non-zero diagonal entries denoted as σc and called singular val-
ues, which are the square roots of the eigenvalues of both JT J and JJT , sorted
in descending order, q.v. (Martin and Porter 2012).
X
J≃ σc uc vcT , (2)
c∈[0,r)
11
SVD factorized matrix:
J† := VΣ† UT , (3)
where Σ† is simply formed as a diagonal matrix of size n×q with the reciprocals
of the singular values. Similarly, the approximate pseudo-inverse of the Jacobian
can be computed as: X
J† ≃ σc−1 vc uTc . (4)
c∈[0,r)
However, in the same way, a minimal loss approximate solution exists for such
equations when the matrix is rectilinear, J† o is expected to be the least-square
solution to the linearized Jacobian approximate of a navigation problem. Even
though the system might in theory have a solution, the odds of finding a unique
solution are practically very skewed towards having an indeterminate system
with many more inputs than outputs, and thus the system will have many
approximate solutions rather than a unique exact solution.
This is of course in line with the intuition of most human beings about
the inherent difficulty of design problems for which there is no unique solution.
Note that in all these theoretical treatments we implicitly assumed that all data
points within the decision space correspond to valid designs whereas in reality,
it might be more difficult to ensure finding valid solutions (feasible in the sense
of complying with constraints) rather than good solutions. In other words,
constraint solving tends to be more difficult than optimization in a feasible
region of decision space.
12
0 Morton Binary Index
Morton Code= (d)
1 (ZYXZYX) Problem Elements
[decimal]
2
assuming XX=00
3
4 voxels:
1010 1011 1110 1111 10
36 38
11 14
52 15
54
11
11
100100 100110 110100 110110
10
10
100000 100010 110000 110010
01
01
000100 000110 010100 010110
Z-Axis
00
00
000000 000010 010000 010010
0 1 2 3 4
34
9
Massing Problem (Binary [3D] Image) Zoning Problem (Coloured [3D] Image)
3
6
Y-Axis
1
2
s
xi
-A
0 1 2 3 4
Figure 5: An illustrative discrete design domain and its associated decision space
distinguished for shaping/massing and zoning/configuring problems
13
3.3 Statistical Approaches to Design
Amongst the statistical approaches to design, we can distinguish the possibilistic
approaches from the probabilistic ones.
• probabilistic approaches: BBN, Variational Auto-Encoders (VAE), and
Diffusion Models
• possibilistic approaches: Markovian Design Machines, Fuzzy Design (see
MAGMA below)
14
4 Demonstration
In this section, we will present a demonstration of the utility of the proposed
framework to indicate how a discrete decision-making approach can facilitate
generative design processes. As a disclaimer, it must be noted that this example
is not chosen for technical reasons related to AI but rather due to its real and
societal and environmental important purpose for policy analysis concerning
energy transition planning actions at the level of a country, and sustainability
strategies at the building level.
15
influence of each design decision; and (3) cannot provide detailed information
yet about later stage design choices, such as the technical systems.
The framework of Augmented Computational Design (ACD) is particularly
useful here as it allows us to relate the aggregated performance changes of the
few NTA 8800 outputs of interest to the changes in the many design deci-
sion parameters of its input. In this particular case, we adopt a probabilistic
meta-modelling (function approximation) approach based on the methodology
suggested by (Conti and Kaijima 2021).
4.2 Methodology
Meta-Modelling Meta models are models that describe the structure, be-
haviours, or other characteristics of related models, providing a higher-level
abstraction for constructing and interpreting complex numerical models that
approximate more sophisticated models often based on simulations. A meta-
model is to serve as a simplified, computationally efficient model of the model
(Conti and Kaijima 2021), also referred to as a surrogate model (J. P. C. Kleijnen
1975). The process of creating a meta-model is referred to as meta-modelling
(Gigch 1991). Some alternative meta-modelling techniques include interpola-
tion methods such as spline models (Barton 1998), polynomial regression (J. P.
Kleijnen 2008), or Krigging (Ankenman, Nelson, and Staum 2010).
Within the ACD framework, such meta-models provide structured ways to
perform the two most important tasks of the generative design: mapping and
navigation.
In general, a standard meta-model can be described as: o = f (x) ≃ g(x),
where o is the aggregated simulation response, f denotes a computational
simulation-based model conceptualized as a vector function and g is the
approximated model function (cf. Figure 3.) With this notation, the objective
of meta-modelling is to build the g in such a way that it produces reasonably
close values of o. In the case of augmented computational design, meta-
modelling can be adopted as a methodology of design mapping that provides
a differentiable and ideally reversible g that can be used in the navigating
process. In other words, the meta-modelling should structurally relate the
choices and consequences in such a way that the choices can be derived from
the desired consequences; hence providing a data-driven basis for generative
design. The next part demonstrates a probabilistic meta-modelling approach
to navigation tasks in high-dimensional design decision spaces, based on the
methodology introduced by Conti and Kaijima (2021).
16
of the nodes. This structure is typically set by the modeller based on their knowl-
edge of the process; while the conditional probability distributions (transition
probability matrices) are learnt from the experimental data. BBNs can help
us semi-automatically reason about uncertain knowledge or data (Peng, Zhang,
and Pan 2010). This makes it possible to perform probabilistic inference, such
as computing the JPDs of some outputs (effects) given some inputs (causes).
The name of these ANN comes from the idea of updating beliefs or hypotheses
posterior to observing evidence; more precisely, utilizing the Bayes theorem for
updating conditional probabilities in network structures, in a fashion similar to
modelling and evaluating Markov Chains, albeit with the difference that Markov
Chains operate as uni-partite networks but each neuron of a BBN is a bipartite
network coupled with an outer product calculator.
The neurons of a BBN consist of two layers, the first of which can be
dubbed a presynaptic layer that combines input discrete probability distribu-
tions (through an outer product) and forms a JPD and then flattens the JPD
to form a vector-shaped probability distribution. The second, i.e. the synaptic
layer then is a CPD, i.e. practically a rectangular probability transition matrix
that maps this flattened JPD to the output probability distribution. A BBN
then consists of such neurons connected in a DAG. Training a BBN means find-
ing the entries of the CPD in such a way as to minimize the loss of recovery
of the output probability distribution from the input distributions. The appeal
of BBNs is twofold: on the one hand, they allow the inclusion of expert knowl-
edge and intuition in the architecture of the network and on the other hand the
training of the network makes the network adapted to the objective data. In
this case, we limit the architecture of the network to a single layer of neurons
to keep the network invertible.
Network Architecture:
probability of sending emails to customer
A Directed Acyclic Graph (DAG)
probability of having meetings with customer
probability of sales happening
probability of revenue brackets
presynaptic layer:
form a Joint Probability Distribution (JPD)
as
a flattened outer product of inputs
synaptic layer:
form a Conditional Probability Distribution (CPD) Nodes with a column sum equal to 0 are input nodes
as
Nodes with a row sum equal to 0 are output nodes
a rectangular transition probability matrix
Workflow Research from Conti and Kaijima (2018) illustrates the four pro-
cess steps involved in developing a BBN meta-model. In this use case, we alter
this methodology as creating a BBN with all 269 input parameters is infeasi-
ble. We add an intermediary step of sensitivity analysis to identify the most
influential input parameters before constructing the BBN. Thus we follow these
steps in order (see Figure 7): (1) sample the input parameter space, (2) run
simulations to generate the output values, (3) sensitivity analysis and selection
17
of influential input parameters, (4) train the BBN, and finally (5) evaluate the
model’s robustness. As Conti and Kaijima (2021) highlights, it is important to
model a shallow BBN as a complete bipartite graph connecting all input nodes
to all output nodes, effectively limiting the topology to two layers. This would
allow us to make a reversible approximation that can be used to derive the nec-
essary input configuration for any desired performance output. Additionally,
the fixed values can also include some of the input variables turning them into
design constraints.
18
Step 2: Run NTA 8800 Simulation Model Vectorization is an important
part of the ACD, we represent the decision variables and outputs of interests
as vectors (See Section 3.) Each sample point can be interpreted as a vector
of scalar input values x. Each batch of such vectors is fed to the NTA 8800
model to generate the vector of corresponding building performance outputs
o. After running the primary simulation model for the sampled input data
points, the response data is collated and linked to the input samples to form an
input-output dataset for regression modelling (as in Machine Learning).
19
The CPDs for the nodes can be learnt directly from the input-output simula-
tion data generated in steps 1 and 2, using the Maximum Likelihood algorithm.
Additionally, we discretize each variable range into a fixed number of intervals.
All numerical input distributions generated using a space-filling approach, like
Sobol’s sequence or Latin Hypercube, are sampled based on continuous ranges,
and should therefore be discretized. Discretization is done by dividing the in-
terval of the parameter over a fixed number of ranges between the minimum
and maximum values.
Figure 8: The single layer BBN; Right: BENG 1, BENG 2, and BENG 3
parameters are the output nodes; Left: Most sensitive input parameters as input
nodes.
Step 5: Validating the Meta-Model To assess how our trained BBN ap-
proximates the original NTA 8800 model, we use a cross-validation approach
in combination with Normalized Root Mean Square Error (NRMSE) and Mean
Absolute Percentage Error (MAPE) (James et al. 2013). Cross-validation splits
the generated input-output dataset (step 2) into a training set and testing set,
before building the BBN (step 3). The BBN is trained on the training set and
assessed based on the testing set. However, to obtain a more reliable estimate of
the model’s performance, the dataset is split into several subsets or folds, with
each fold used as both a training set and a testing set. This research adopts a
k-fold cross-validation technique, where k refers to the number of groups that
the data set is split into. We set k = 10 based on experimentation to ensure a
low bias and a modest variance. The model is then trained on k − 1 of the folds,
and the remaining fold is used for testing. This process is repeated k times, with
each fold used for testing once. The performance of the model is then evaluated
by averaging the performance across all k runs.
Following the approach suggested by Conti and Kaijima (2021), we computed
the mean difference of the predicted and actual output values and normalize the
RMSE values through division by standard deviation to achieve NRMSE.
To calculate the accuracy of the meta-model, it is recommended to use mul-
tiple metrics to get a comprehensive evaluation of the model’s accuracy. Hence,
NRMSE is combined with the MAPE metric. MAPE measures the average
absolute percentage difference between the predicted and actual values. It is
a measure of the magnitude of the errors in the model’s predictions. Lower
NRMSE and MAPE values indicate better model performance. The larger the
20
error between the two, the higher the NRMSE and MAPE values will become.
Therefore, the NRMSE and MAPE results will indicate how dispersed the pre-
diction data is compared to the actual model response.
4.3 Results
This section presents the numerical results obtained from the experiment of
NTA 8800 meta-model.
21
Figure 9: Histogram of BBN prediction and NTA 8800 outputs (top) and their
comparison (bottom) for BENG1, BENG2, and BENG3, in order from lft to right.
In this toy problem, the meta-model predicts the BENG 1 value to be within
the range of (0 − 50)kW h/m2 .y, with a confidence level of 100% (see Figure
10). To validate this result, we cross-checked the predicted result with the final
configuration of the dwelling using the original NTA 8800 model. The NTA 8800
model returns a value of 39.8kW h/m2 .y, confirming the prediction capability
of the meta-model.
The second problem reverses the first problem and involves the ex-ante de-
termination of the most probable design configuration that satisfies a specific
energy performance goal. In this example, the BENG 3 value of a typical Dutch
dwelling design (35%) does not satisfy the minimal requirements (50%). Since
this problem arises in the final design stage, some input parameters cannot be
changed anymore. In this case, the architects and engineers are limited to mod-
ifying only the area (AreaPV) and power (PPV) of the PV panels. Since the
minimum required performance goal for BENG 3 is 50%, we set the goal value
to a range of 60 − 80%. Given the binning approach employed, it should be
noted that the AreaPV value of 5 depicted in the figure corresponds to a range
of [40, 50]m2 , while the PPV value of 5 corresponds to a PV Power range of
[200, 250]W/m2 . Accordingly, the meta-model advises increasing the PV area
to [40, 50]m2 , and the PV power to [200, 250]W/m2 (See Figure 10).
This discretisation allows for a clearer representation of the recommended
parameter values within the specified ranges, facilitating the interpretation and
practical implementation of the BBN meta-model outputs. Since these ranges
are the maximum of both scales: the meta-models advise can be interpreted as
maximizing the PV area and PV power to reach the goal BENG 3 value of 60%
to 80%.
4.4 Discussion
In the end, to validate this result, we finish the loop by calculating the final
configuration of the dwelling with the original NTA 8800 model. The NTA
8800 returns a value of 71%, confirming the reverse inference capability of the
meta-model. These results illustrate in a simple and digestible example how
the BBN meta-model is capable of providing valuable insights and assisting
architects and engineers in navigating the multidimensional decision space.
By using the numerical Design of Experiments and Sensitivity Analysis we
are effectively conducting a dimensionality reduction task similar to the low rank
SVD as introduced before. As mentioned earlier, the mention of this particular
22
Figure 10: Output Recommendations of the BBN Meta-model for Achieving BENG
3 Compliance.
approach of ACD was to illustrate the utility of the framework with a concrete
example in a societally relevant context where a Machine Learning approach to
modelling can help make an otherwise very complicated simulation procedure
to be approximately scaled up massively for policy analysis. Here we discuss
the potentials and the shortcomings of the model and note the issues with this
large-scale black-box approximation that require further investigation.
The existence of categorical variables in the inputs of the BBN limits the
general applicability of ACD as it affects the smoothness and differentiability
of the underlying function that is being approximated. However, for pragmatic
reasons, we have ignored this issue to demonstrate the idea in a large-scale case.
Foreward Inference The trained BBN is now capable of inferring the outputs
of interest given certain input configurations. This inference uses the learned
CPDs to predict the most likely values for the outputs. In this way, we can
predict the energy performance of buildings, in a quick and intuitive way for ex-
ante assesment, based on a certain design configuration. In particular, inference
demonstrates the potentials of a mapping described in Section 3.
23
Backward Inference Since our BBN had only two layers in its network,
it can be reversed. This means that instead of presenting evidence to it, we
can present the desired performance values and ask for the derivation of the
particular configuration of inputs that will produce such an output. This can
be done through the Variable Elimination module of the pgmpy (Ankan and
Panda 2015). The same is also true for a combination of given input-outputs;
meaning that evidence can be given for both inputs and outputs of the BBN.
In such cases, the given inputs can also function as the design constraints.
The reverse inference demonstrates how we can utilize probabilistic models to
navigate a decision space as explained in Section 3.
Binning The variables are divided into a specified number of bins based on
the frequency of the values. In equal frequency binning, the data is first sorted
in ascending order and then divided into the specified number of bins, with
each bin containing an equal number of observations. Equal frequency binning
can be useful for analyzing data that has a skewed distribution or contains
outliers. By dividing the data into equal-frequency bins, the impact of outliers
may be reduced, and the distribution of the data may be more balanced across
the bins. Considering the prediction robustness of the discretized values, each
bin should contain a fair amount of data points. A fair amount is a bit vague
description because there is no method or rule of thumb for deciding the number
of bins. However, we need to keep in mind that the number of bins corresponds
with the number of states a parameter can have, and there with the number
of parameters that are used to learn the network. Increasing the number of
bins, results in an exponential increase of computational demand. Decreasing
the number of bins, however, results in too few states to gain the desired insight
accuracy in the input-output probabilities. This means that the number of bins
should be chosen carefully, keeping in mind the trade-offs, and satisfying the
research accuracy and the robustness of the Bayesian meta-model.
24
performance in dwellings, and (2) to develop a proposal for an adapted repre-
sentation of the NTA 8800 model tool, embedded in a computational model,
to support intelligent decision making. The research workflow that has been
described is a valid approach to reaching the stated design goals.
Toy problem The results of the toy problem show the implementation pos-
sibilities, and show, of course, simplified, the possibilities for professionals to
utilize the tool in the practical field. As shown in the previous chapter, the tool
works as expected, and is capable of making preliminary predictions with very
few input variables. It appeals to the imagination when the tool is connected
to spatial configurations instead of unmeaning variables in a computer script.
5 Outlook
The ACD framework and its constituent concepts can be best positioned within
the context of performance-driven computational design and generative design.
In particular, the idea of approximating complex and non-linear functions for
estimating measurable performance indicators from configurations of decision
variables, even if referring to non-spatial decision variables, is generalisable to
all areas of computer-aided design. However, such surrogate models are not
to replace simulation models based on first principles as they can not match
their transparency and explainability. Nevertheless, in cases where one needs to
estimate the effects of design decisions on human factors, ergonomics, or com-
binations of many different types of governing equations, an estimation model
trained from actual data can be of utility in that it provides a basis for compar-
isons in the absence of analytical knowledge. In other words, the utility of ANNs
for mapping the associations between decision data points and performance data
points is apparent.
The navigation problem, on the other hand, is much harder, philosophically,
technically, and mathematically for being solved in any sense. The real advan-
tage of an AI framework in dealing with a design space navigation problem can
be attained if the latent space of the model reveals interpretable information
or if it is at least coupled with a sensible low-dimensional space. If the latent
space of e.g. an Auto Encoder (Marin et al. 2021) is understandable as a low-
dimensional vector space (as an endpoint of the mapping and the start point of
the navigating processes) then it can be used not only to guide the navigation
process but also to gain insight as to which design variables are more important
in determining the attainment levels of outcomes of interest. In other words,
even though it appears that in the mapping process, the information content
of the decision data points is reducing gradually, one can think of this process
as a distillation of an elixir from a large data point that makes the information
richer from a human perspective.
In this light, the major advantages of the proposed framework are twofold:
Firstly, providing an elegant framework for applying AI in computational de-
sign in the presence of many complex quality criteria; and secondly, providing an
elegant framework for designing spatial manifolds very much like the method-
ology of electrical engineering in designing electronic circuits and systems for
signal processing. The latter point requires much more space for discussing the
theoretical minimum for such an approach to design from a signal processing
25
standpoint. In short, however, we can briefly mention that the idea of defining
a central representation of a configuration as a discrete manifold provides for
directly modelling the functionality of the spatial manifold w.r.t. the flows of
walks, light rays, or forces not only from the point of view of spatial movement
but also much more elegantly and efficiently in the frequency or spectral domain
(which can be attained using Discrete Fourier Transform or Spectral Mesh Anal-
ysis). One fundamental idea of analogue circuit design from a signal processing
point of view is that of designing passive “filters” whose properties can much
better be understood in the so-called frequency domain analyses put forward by
Fourier and Laplace transforms of the so-called transfer functions of the RLC
(Resistor, Self-Induction Loop, Capacitor) circuits. This approach to circuit
design can be traced back to the ideas and propositions of Oliver Heaviside
(1850-1925), a self-educated pioneer of electrical engineering. Arguably, this
frequency-based outlook (relating to the spectrum of eigen frequencies of vibra-
tion of shapes, thus also identifiable as a spectral approach), has revolutionized
the formation of the field of electronics and thus contributed significantly to the
development of AI as we know it today. Identifying spectral latent spaces and
associating them with low-dimensional performance spaces and latent spaces of
ANNs is a topic that calls for further theoretical research and computational
experimentation.
References
Ankan, Ankur, and Abinash Panda. 2015. “pgmpy: Probabilistic graphical models
using python.” In Proceedings of the 14th python in science conference (scipy
2015), vol. 10. Citeseer.
Ankenman, Bruce, Barry L. Nelson, and Jeremy Staum. 2010. “Stochastic Kriging for
Simulation Metamodeling.” Operations Research 58, no. 2 (April): 371–382. https:
//doi.org/10.1287/opre.1090.0754. https://doi.org/10.1287%2Fopre.1090.0754.
Azadi, Shervin, and Pirouz Nourian. 2021. “GoDesign: A modular generative design
framework for mass-customization and optimization in architectural design.” In
Towards a new, configurable architecture, 1:285–294. Novi Sad, Serbia: CUMIN-
CAD, August.
Bai, Nan, Shervin Azadi, Pirouz Nourian, and Ana Pereira Roders. 2020. “Decision-
Making as a Social Choice Game” [in en]. In Proceedings of the 38th eCAADe
Conference, 2:10.
Barton, R.R. 1998. “Simulation metamodels.” In 1998 Winter Simulation Conference.
Proceedings (Cat. No.98CH36274). IEEE. https://doi.org/10.1109/wsc.1998.
744912. https://doi.org/10.1109%2Fwsc.1998.744912.
Batty, M. 1974. “A Theory of Markovian Design Machines.” Environment and Plan-
ning B: Planning and Design, https://doi.org/10.1068/b010125.
Bui, Van, Nam Tuan Le, Van Hoa Nguyen, Joongheon Kim, and Yeong Min Jang.
2021. “Multi-Behavior with Bottleneck Features LSTM for Load Forecasting in
Building Energy Management System.” Electronics 10, no. 9 (April): 1026. https:
//doi.org/10.3390/electronics10091026. https://doi.org/10.3390%2Felectronics
10091026.
26
Conti, Zack Xuereb, and Sawako Kaijima. 2018. “A Flexible Simulation Metamodel
for Exploring Multiple Design Spaces” [in en]. International Association for Shell
and Spatial Structures (IASS). https://doi.org/10.13140/RG.2.2.23313.53600.
http://rgdoi.net/10.13140/RG.2.2.23313.53600.
. 2021. “Explainable ML: Augmenting the interpretability of numerical simula-
tion using Bayesian networks.” In The Routledge Companion to Artificial Intelli-
gence in Architecture, 315–335. Routledge.
Fang, Kai-Tai, Runze Li, and Agus Sudjianto. 2005. Design and Modeling for Computer
Experiments. Chapman / Hall/CRC, October. https://doi.org/10.1201/9781420
034899. https://doi.org/10.1201%2F9781420034899.
Friedman, Yona. 1975. Toward a scientific architecture [in English]. First American
Edition. Cambridge, Mass: MIT Press, January. isbn: 978-0-262-06058-5.
NTA 8800. 2022. Accessed on Feb 28th, 2023, December. https://www.gebouwenergi
eprestatie.nl/bepalingsmethode.
Gero, John S, and Udo Kannengiesser. 2004. “The situated function–behaviour–
structure framework.” Design studies 25 (4): 373–391.
Giedion, Sigfried. 2009. Space, time and architecture: the growth of a new tradition.
Harvard University Press.
Gigch, J.P. van. 1991. System Design Modeling and Metamodeling. Language of science.
Plenum. isbn: 9780306437403. https://books.google.nl/books?id=M5mD0ZZcw
aEC.
Herman, Jon, and Will Usher. 2017. “SALib: An open-source Python library for Sen-
sitivity Analysis.” The Journal of Open Source Software 2, no. 9 (January): 97.
https://doi.org/10.21105/joss.00097. https://doi.org/10.21105%2Fjoss.00097.
Hicks, Charles Robert. 1964. Fundamental Concepts in the Design of Experiments.
New York: Holt, Rinehart / Winston.
Hillier, Bill. 2007. Space is the machine: a configurational theory of architecture. Space
Syntax.
James, G., D. Witten, T. Hastie, and R. Tibshirani. 2013. An Introduction to Statistical
Learning: with Applications in R. Springer Texts in Statistics. Springer New York.
isbn: 9781461471387. https://books.google.nl/books?id=qcI%5C AAAAQBAJ.
Jia, Zhuoran, Pirouz Nourian, Peter Luscuere, and Cor Wagenaar. 2023. “Spatial de-
cision support systems for hospital layout design: A review.” Journal of Building
Engineering, 106042.
Khan, Anam Nawaz, Naeem Iqbal, Rashid Ahmad, and Do-Hyeun Kim. 2021. “En-
semble Prediction Approach Based on Learning to Statistical Model for Effi-
cient Building Energy Consumption Management.” Symmetry 13, no. 3 (March):
405. https : / / doi . org / 10 . 3390 / sym13030405. https : / / doi . org / 10 . 3390 %
2Fsym13030405.
Kleijnen, Jack P. C. 1975. “A Comment on Blannings Metamodel for Sensitivity Anal-
ysis: The Regression Metamodel in Simulation.” Interfaces 5, no. 3 (May): 21–23.
https://doi.org/10.1287/inte.5.3.21. https://doi.org/10.1287%2Finte.5.3.21.
. 2008. “Low-order polynomial regression metamodels and their designs: basics.”
In Design and Analysis of Simulation Experiments, 15–71. Boston, MA: Springer
US. isbn: 978-0-387-71813-2. https://doi.org/10.1007/978-0-387-71813-2 2.
27
Koller, Daphne, and Nir Friedman. 2009. Probabilistic graphical models: principles and
techniques [in en]. Adaptive computation and machine learning. Cambridge, MA:
MIT Press. isbn: 978-0-262-01319-2.
Kroes, Peter. 2010. “Engineering and the dual nature of technical artefacts.” Cam-
bridge journal of economics 34 (1): 51–62.
Kroes, Peter, and Anthonie Meijers. 2006. “The dual nature of technical artefacts.”
Studies in History and Philosophy of Science 37 (1): 1–4.
Marin, Riccardo, Arianna Rampini, Umberto Castellani, Emanuele Rodolà, Maks Ovs-
janikov, and Simone Melzi. 2021. “Spectral Shape Recovery and Analysis Via
Data-driven Connections” [in en]. Int J Comput Vis 129, no. 10 (October): 2745–
2760. issn: 0920-5691, 1573-1405, accessed July 24, 2022. https://doi.org /10.
1007/s11263-021-01492-6. https://link.springer.com/10.1007/s11263-021-01492-
6.
Martin, Carla D, and Mason A Porter. 2012. “The extraordinary SVD.” The American
Mathematical Monthly 119 (10): 838–851.
Nen 7120+C2:2012/A1:2017 NL. 2017. Accessed on Feb 28th, 2023, June. https :
//www.nen.nl/nen-7120-c2-2012-a1-2017-nl-229670.
Nourian, Pirouz. 2016. Configraphics: Graph Theoretical Methods for Design and Anal-
ysis of Spatial Configurations [in en]. TU Delft Open, September. isbn: 978-
94-6186-720-9, accessed September 23, 2021. https : / / doi . org / 10 . 7480 / isb
n . 9789461867209. https : / / books . bk . tudelft . nl / press / catalog / book / isbn .
9789461867209.
Nourian, Pirouz, Shervin Azadi, and Robin Oval. 2023. “Generative Design in Archi-
tecture: From Mathematical Optimization to Grammatical Customization.” In
Computational Design and Digital Manufacturing, edited by Panagiotis Kyratsis,
Athanasios Manavis, and J. Paulo Davim, 1–43. Cham: Springer International
Publishing. isbn: 978-3-031-21167-6. https://doi.org/10.1007/978-3-031-21167-
6 1.
Nourian, Pirouz, Romulo Gonçalves, Sisi Zlatanova, Ken Arroyo Ohori, and Anh Vu
Vo. 2016. “Voxelization algorithms for geospatial applications: Computational
methods for voxelating spatial datasets of 3D city models containing 3D surface,
curve and point data models” [in en]. MethodsX 3 (January): 69–86. issn: 2215-
0161, accessed April 28, 2020. https : / / doi . org / 10 . 1016 / j . mex . 2016 . 01 . 001.
http://www.sciencedirect.com/science/article/pii/S2215016116000029.
O’Shaughnessy, Connor, Enrico Masoero, and Peter D. Gosling. 2021. Topology Op-
timization using the Discrete Element Method. Part 1: Methodology, Validation,
and Geometric Nonlinearity [in en]. Preprint. engrXiv, July. Accessed Decem-
ber 20, 2021. https://doi.org/10.31224/osf.io/c6ymn. https://osf.io/c6ymn.
Pearl, Judea. 1988. Probabilistic reasoning in intelligent systems: networks of plausible
inference. Morgan kaufmann.
Peng, Yun, Shenyong Zhang, and Rong Pan. 2010. “Bayesian Network Reasoning
with Uncertain Evidences.” International Journal of Uncertainty, Fuzziness and
Knowledge-Based Systems 18, no. 05 (October): 539–564. https://doi.org /10.
1142/s0218488510006696. https://doi.org/10.1142%2Fs0218488510006696.
Regenwetter, Lyle, and Faez Ahmed. 2022. Towards Goal, Feasibility, and Diversity-
Oriented Deep Generative Models in Design. ArXiv:2206.07170 [cs], June. Ac-
cessed July 22, 2022. http://arxiv.org/abs/2206.07170.
28
Regenwetter, Lyle, Amin Heyrani Nobari, and Faez Ahmed. 2022. “Deep Generative
Models in Engineering Design: A Review” [in en]. Journal of Mechanical Design
144, no. 7 (July): 071704. issn: 1050-0472, 1528-9001, accessed July 24, 2022.
https://doi.org /10.1115/1.4053859. https://asmedigitalcollection.asme.org /
mechanicaldesign/article/144/7/071704/1136676/Deep- Generative- Models- in-
Engineering-Design-A.
Sacks, Jerome, William J. Welch, Toby J. Mitchell, and Henry P. Wynn. 1989. “Design
and Analysis of Computer Experiments.” Statistical Science 4, no. 4 (November).
https : / / doi . org / 10 . 1214 / ss / 1177012413. https : / / doi . org / 10 . 1214 % 2Fss %
2F1177012413.
Simon, Herbert A. 1973. “The Structure of Ill Structured P coblems” [in en]. Artificial
Intelligence, 21.
. 2019. The Sciences of the Artificial, reissue of the third edition with a new
introduction by John Laird. MIT press.
Sobol’, Il’ya Meerovich. 1990. “On sensitivity estimation for nonlinear mathematical
models.” Matematicheskoe modelirovanie 2 (1): 112–118.
Soman, Aditya, Azadi, Shervin, and Nourian, Pirouz. 2022. “DeciGenArch: A Gener-
ative Design Methodology for Architectural Configuration via Multi-Criteria De-
cision Analysis” [in en]. In Proceedings of eCAADe 2022, 459–468. Education /
research in Computer Aided Architectural Design in Europe.
Union, Europian. 2021. Proposal for a Directive of the European Parliament and of
the Council on the Energy Performance of Buildings (Recast). Accessed on Feb
28th, 2023, December. https://eur-lex.europa.eu/legal-content/EN/ALL/?uri=
CELEX%3A52021PC0802.
Veloso, Pedro, and Ramesh Krishnamurti. 2020. “An Academy of Spatial Agents: Gen-
erating Spatial Configurations with Deep Reinforcement Learning.” September.
. 2021. “Mapping generative models for architectural design.” In The Routledge
Companion to Artificial Intelligence in Architecture, 29–58. Routledge.
Weng, Lilian. 2021. “What are diffusion models?” lilianweng.github.io (July). https:
//lilianweng.github.io/posts/2021-07-11-diffusion-models/.
Zeng, Mingliang, Xu Tan, Rui Wang, Zeqian Ju, Tao Qin, and Tie-Yan Liu. 2021.
“MusicBERT: Symbolic Music Understanding with Large-Scale Pre-Training.”
In ACL-IJCNLP 2021. June. https : / / www . microsoft . com / en - us / research /
publication / musicbert - symbolic - music - understanding - with - large - scale - pre -
training/.
29
A Acronyms
Table 1: Acronyms
Acronym Term
ACD Augmented Computational Design
AEC Architecture, Engineering, and Construction
AI Artificial Intelligence
ANN Artificial Neural Networks
BBN Bayesian Belief Networks
BEM Building Energy Modelling
BENG Bijna Energie Neutrale Gebouwen: Nearly Zero-Energy Buildings
BIM Building Information Model
CAD Computer-Aided Design
CEN Comité Européen de Normalisation: European Committee of
Normalization
CPD Conditional Probability Distribution
DAG Directed Acyclic Graph
DoE Design of Experiment
EPBD European Energy Performance of Buildings Directive
JPD Joint Probability Distributions
MAPE Mean Absolute Percentage Error
MAGMA Multi-Attribute Gradient-Driven Mass Aggregation
ML Machine Learning
NEN Nederlandse Norm: Royal Dutch Standardization Institute
NRMSE Normalized Root Mean Square Error
NTA 8800 Nederlandse Technische Afspraak (Dutch Technical Agreement)
PGM Probabilistic Graphical Models
SVD Singular Value Decomposition
VAE Variational Auto-Encoders
30
B Notation
Table 2: Notations
31