<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Modular Surrogate-based Optimization Framework for Expensive Computational Simulations</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Antonín Panzo</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Boyang Chen</string-name>
          <email>B.Chen-2@tudelft.nl</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Delft University of Technology, Department of Aerospace Structures and Materials Kluyverweg 1</institution>
          ,
          <addr-line>2629 HS Delft</addr-line>
          ,
          <country country="NL">The Netherlands</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>In practical applications, the use of computational modeling has been industry-wide adopted to speed up product development as well as reduce physical testing costs. Such models of complex or large systems are, however, often computationally expensive, hence solution times of hours or more are not uncommon. Additionally, as these models are typically evaluated using blackbox solvers, the direct study of relations between design parameters renders demanding in terms of computational time and provides poor engineering insight and understanding. To address this, a modular framework integrating computation automation with the use of surrogate-based modeling, optimization and visualization techniques is presented. The framework is built in the Python programming language. Its use is illustrated on a study of the side impact response of a car body using an artificial neural network as a surrogate together with the NSGA-III genetic algorithm for optimization.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1 Introduction</title>
      <p>
        Nowadays, solving of many engineering problems, such
as the development of transport vehicles, involves the use
of computationally expensive simulations, such as
computational fluid dynamics (CFD), the finite element method
(FEM), or others, as well as their combinations, e.g. the
fluid-structure interaction (FSI). These will be further
addressed as simulations. In practice, typically, such
simulations suffer from two major drawbacks. Firstly, even
thought the computational power available to an engineer
for running such simulations has been gradually
increasing over years, in practice the requirements on the model
accuracy had increased as well [
        <xref ref-type="bibr" rid="ref52">48</xref>
        ], and as such, there is
still a large quantity of simulations that take hours, days
and occasionally even more, just to evaluate a single
design [
        <xref ref-type="bibr" rid="ref46">43</xref>
        ]. Secondly, they are often of a black-box nature,
meaning that given an input ~X , an output ~Y is returned
to the user without an accompanying explicit relation
between the two [
        <xref ref-type="bibr" rid="ref44">41</xref>
        ].
      </p>
      <p>Within a practically infinite space of design choices for
the system under study, the task of an engineer is to
identify the most influential design parameters and understand
their degree of influence. Through the use of simulations,
often the next goal is to meet a set of threshold
requirements on selected targets, while minimizing or
maximizing a certain property of the overall system, such as cost,
weight, etc. Mathematically, this can be formulated as a
multi-objective optimization problem with the goal to:
subject to
minimize fm(~X ) m = 1; :::; P
g j(~X )</p>
      <p>0 j = 1; :::; Q
hk(~X ) = 0 k = 1; :::; R
X L
i</p>
      <p>Xi</p>
      <p>
        XU
i
i = 1; :::; nin
(1)
(2)
(3)
(4)
where f are the objective functions, g and h are the
inequality and equality constraints, respectively, and ~X is
the vector of design variables in the design search space
bounded by XiL and XiU from the lower and upper side,
respectively [47]. In case of multi-objective optimization
when m 2, the notion of singular optimality needs to be
augmented as there is typically not a single solution X~
minimizing all of the objectives of Equation 1 at the same
time. In such cases, the concept of so-called Pareto
optimality is considered [36], which can be loosely described
as: “for each solution that is contained in the Pareto set,
one can only improve one objective by accepting a
tradeoff in at least one other objective” [
        <xref ref-type="bibr" rid="ref42">39</xref>
        ].
      </p>
      <p>
        Due to said high computational costs and the black-box
nature of simulations, multi-objective parametric studies
and optimizations often become too lengthy and
impractical for actual application [
        <xref ref-type="bibr" rid="ref44 ref45">42, 41</xref>
        ]. The lack of
derivatives required for derivative-based optimization methods,
such as in Liu and Reynolds [35] or Peitz and Dellnitz
[
        <xref ref-type="bibr" rid="ref42">39</xref>
        ], can be addressed by the use of derivative-free
optimization (DFO) methods, such as genetic, evolutionary
or swarm algorithms [27]. However, these methods
compensate for the lack of gradients by evaluating on larger
sets of candidate solutions, which leaves the issue of high
computational cost unresolved. One way to resolve that
is by using a model of lower fidelity requiring less
computational resources [
        <xref ref-type="bibr" rid="ref36">33</xref>
        ]. However, the accuracy of such
lower fidelity models can become unacceptably low. As
an alternative, a surrogate-based approach can be adopted
[
        <xref ref-type="bibr" rid="ref12 ref20 ref5">20, 5, 12</xref>
        ].
      </p>
      <p>
        A surrogate model, also addressed as a metamodel, is
a mathematical model that upon training approximates the
response of the original model based on a sample set from
the design space [31]. Such computationally cheap
surrogate can then be called during the optimization defined by
Equation 1-4 instead of the original, expensive simulation.
In addition, the use of a surrogate model opens up several
new opportunities, which include:
1. using the surrogate as a substructure of a larger model
[
        <xref ref-type="bibr" rid="ref37">34</xref>
        ]
2. cheap numerical evaluation of gradients using the
surrogate
3. repeating the optimization for various formulations of
the problem (within the bounds of validity of the
surrogate) - compare various formulations of the same
car side impact problem: Youn et al. [54], Guo,
Wang, and Wang [26], and Tanabe and Ishibuchi [46]
4. design exploration through sensitivity studies and
visualization [
        <xref ref-type="bibr" rid="ref41 ref53">49, 38</xref>
        ]
On the other hand, a disadvantage is the added
complexity of the overall task due to introduction of extra tuning
parameters related to the surrogate model and its training.
      </p>
      <p>
        Due to the often black-box nature of simulations, it is
not possible to a-priori quantify the sample set size
required for an accurate training of the surrogate [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
Therefore, for a purely optimization-focused approach, in
general it is not guaranteed that such surrogate-based
approach is indeed computationally cheaper than
optimizing directly using the original model. Nevertheless, it has
been established that for budget-limited tasks with
expensive simulations, it is indeed the case [
        <xref ref-type="bibr" rid="ref41 ref47">44, 24, 38</xref>
        ]. This is
the case especially for multi-objective optimization, when
the exploration of the full Pareto front is required.
Confirming results of an empirical comparison can be found
e.g. in Voutchkov and Keane [
        <xref ref-type="bibr" rid="ref54">50</xref>
        ] on benchmark problems
and in Bessa and Pellegrino [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] on a practical problem of
the design of ultra-thin carbon fiber deployable shells.
      </p>
      <p>Example uses of surrogate-based optimization (SBO)
from the area of structural optimization include the
design of 3D weaving composite stiffened panels by Fu,
Ricci, and Bisagni [21], hypersonic vehicle’s metallic
thermal protection by Guo et al. [25] or a wellhead
connector for a subsea Christmas tree by Zeng et al. [55]. The
SBO methodology has been already integrated into
various commercial software packages, such as OptiSLang1
1ANSYS OptiSLang, https://www.dynardo.de/en/
software/optislang/ansys-optislang.html [Accessed on
03/08/2020]
or HEEDS2. This can be often practical, but not always.
Firstly, the user must rely only on the available capabilities
of such software packages, which can sometimes be
insufficient for the application at hand. Secondly, especially for
individuals or smaller businesses, the license price for such
packages can be too expensive, thus inaccessible. Last but
not least, for research purposes, commercial solutions
often offer only a limited insight into the used methods and
source code, leading to limited customabilizity, a feature
often required for research.</p>
      <p>In this paper, the core of a developed Python
opensource SBO framework is presented. In Section 2, the
different elements of the proposed framework are presented.
Next, in Section 3, the use of the framework is
demonstrated on a benchmark problem. Finally, the capabilities
of the framework are summarized in Section 4 together
with an outline of suggested further developments.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Proposed method</title>
      <p>This framework integrates elements of surrogate-based
modeling and automated model selection together with
derivative-free optimization. One of the driving ideas
behind it is modularity, such that it could be
readily customized for the user’s particular needs. Upon
completion of the master thesis project, the framework
will be made publicly available at github.com/apanzo/
optimization.</p>
      <p>The top-level illustration of the framework is shown in
Figure 1. As a starting point, the simulation is prepared,
and the framework settings, such as the selection of a
surrogate model and optimization algorithm, are set. Both
the use of a surrogate and performing optimization are
optional. The surrogate loop contains sample selection,
model evaluation, results storage and retrieval,
optimization and training of the surrogate, and finally accessing its
convergence. The optimization loop consists of the
actual optimization, and, in case that a surrogate model is
used, a verification of the obtained results against the
original model. In the following subsections, key parts of the
framework are discussed in the order from optimization
start to end as indicated in Figure 1.
2.1</p>
      <sec id="sec-2-1">
        <title>Sampling strategy</title>
        <p>Considering the expensiveness of the simulations
addressed by this framework, one of the keys of success is
the smart selection of the sample points within the
explored design space, such that the surrogate model obtains
a sufficiently representative sample to be trained on. For
this purpose, simpler sampling strategies such as the grid
search or random search are not suitable. The former is
expensive as it does not benefit from sampling in multiple
dimensions, as the sample projection on each of the input’s
2HEEDS MDO, https://www.redcedartech.com/index.
php/solutions/heeds-software [Accessed on 03/08/2020]
axes is independent of the number of inputs dimensions,
and in the case of re-sampling with a finer sample, either
the new sample discards all of the previous samples, or the
re-sampling is to be done on an integer multiple of
samples per input dimension of the current sample. The
random search does not suffer from those, but its weakness
lies in its poor space-filling property, thus a large amount
of samples is required to generate a sufficiently
representative sample.</p>
        <p>
          As a baseline within this framework, quasi-random
sampling strategies are implemented. In particular, the two
available sampling methods are the cell centered Latin
hypercube sampling (LHS) [
          <xref ref-type="bibr" rid="ref14 ref18">14, 18</xref>
          ] and the original Halton
sequence sampling [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]. The principal logic behind the
LHS method is that it is a sparse grid method where the
projection of the sample on each of the input’s axes
contains only one sample per uniform interval. As such, the
sample size is significantly reduced compared to the full
grid. On the contrary, the disadvantage of the poor
resampling pertains. This is not the case for the Halton sequence,
which is a multi-dimensional version of the van Der
Corput sequence. It is defined as
        </p>
        <p>
          P = ffq1 (Z); :::; fqn (Z)g
with bases q1; :::; qn that are mutually coprime and in
practice taken as the first s primes [
          <xref ref-type="bibr" rid="ref13 ref23">13, 22</xref>
          ]. fq(Z), the inverse
radix number, is obtained from an integer
        </p>
        <p>r
Z = å aiqi</p>
        <p>
          i=0
where q is the basis and the largest exponent is r =
int llnnZq , and 0
ai &lt; q are the coefficients [
          <xref ref-type="bibr" rid="ref11 ref13">13, 11</xref>
          ], as
        </p>
        <p>r
fq(Z) = å aiq (i 1)</p>
        <p>
          i=0
The advantage is that each point of the sequence is
generated independently of the previous point. An example
is shown in Figure 2. It is noted however that the
original method is suitable only up to 8 input dimensions, as in
higher dimensions, spurious correlations between the
inputs occur, considerably deteriorating the sample’s quality
[
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]. This can be leveraged by using its modified versions
that address this issue [
          <xref ref-type="bibr" rid="ref19">19, 37</xref>
          ].
        </p>
        <p>
          To take it one step further, considering that these, so
called static sampling methods, only take into account the
space-filling criteria, that is the input’s quality, and not the
nature of the response, that is the output’s quality, an
adaptive sampling method that does so is implemented as well.
As an example, if S is the full design space and the
response of a model is flat everywhere apart from the region
D 2 S, where there is a valley in the response, only a few
samples are required outside D, while in D more samples
are required to train the surrogate. The selected method
is from Eason and Cremaschi [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ], which determines the
new samples based on a balance of the variance in
predictions of the K models from cross-validation, explained
typically used and the task of learning is to obtain the
unknown relation between input and output data.
        </p>
        <p>The core concept of an ANN is a single neuron. Inspired
by the function of a real neuron inside the living brain, the
mathematical model from input x to output y consists of
performing a weighted summation of all the inputs
and applying a non-linear transformation
with g the so-called activation function. Common
activation functions are the logistic sigmoid</p>
        <p>nin
z = å wixi</p>
        <p>i=1
y = g(z)
g(z) =</p>
        <p>1
1 + e (z) ;
g(z) = max(0; z)
g(z) = z</p>
        <p>1
1 + e (z)
(5)
further in subsection 2.4, for exploitation and the distance
to the nearest sample for exploration, each normalized by
its maximum. The new sample is chosen where the sum of
the two criteria is the largest.
2.2</p>
      </sec>
      <sec id="sec-2-2">
        <title>External evaluator</title>
        <p>For the practical use of the framework, it has to be able to
interact with external solvers that evaluate the simulations.
This independence is one of the main strong points of the
framework, as it allows to integrate results from different
solvers. As an example, for a FSI study, different FEM
and CFD solvers can be used and their results integrated
within this framework.</p>
        <p>The implementation of different evaluators is
application dependent, therefore the interaction with each new
solver has to be customized according to its specific
interface. In the course of development of this framework,
a custom ANSYS evaluator has been integrated. For its
use, the procedure is to firstly build a parametric model
inside the ANSYS Workbench environment3, with the
input and output parameters defined. These parameters are
passed over by name to the framework together with the
path to the prepared model. The rest of the interaction is
automated. After determining the sample points, firstly the
framework checks that there are available free licenses for
carrying out the computations. If there are available
licenses, a request is sent to ANSYS Workbench to evaluate
the selected sample points. Once all the samples have been
evaluated, the framework retrieves the defined output
parameters and integrates them within its internal database.
2.3</p>
      </sec>
      <sec id="sec-2-3">
        <title>Surrogate model</title>
        <p>
          Among others, the most commonly used
surrogatemodeling techniques are kriging, radial basis functions
(RBF), artificial neural networks (ANN) or support vector
machines (SVM) [
          <xref ref-type="bibr" rid="ref12 ref17">17, 12</xref>
          ]. The former two are integrated
within the framework using the Surrogate Modeling
Toolbox (SMT) library [
          <xref ref-type="bibr" rid="ref10 ref59">10</xref>
          ] natively, while the ANN is
integrated as a custom class into SMT using the TensorFlow
library [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]. In the course of the reported research, the ANN
has been used due to previous experience in the research
group, therefore it is elaborated upon closer in this section.
        </p>
        <p>
          In the general sense, ANNs are one of the subclasses
of the broader field of machine learning (ML), which in
turn is a subclass of artificial intelligence (AI). Most
commonly, the ANNs are trained using a form of supervised
learning, that is by providing both the input and output
data during the training process. They can be used both
for classification, that is categorization of the input content
into different sub-classes with common features, as well as
regression. For the purpose of optimization, regression is
3ANSYS Workbench Platform, https://www.ansys.com/
-/media/Ansys/corporate/resourcelibrary/brochure/
workbench-platform-121.pdf [Accessed on 05/08/2020]
rectified linear unit (ReLU)
[
          <xref ref-type="bibr" rid="ref24">23</xref>
          ], or Swish
[
          <xref ref-type="bibr" rid="ref43">40</xref>
          ] functions.
        </p>
        <p>
          To provide a flexible range of outputs based on the
neuron’s activations, the neurons are organized into
layers. The typical neural network architecture for
regression is the multi-layer perceptron (MLP), consisting of
the first and last layers containing the amount of neurons
equal to the number of input and output dimensions,
respectively, and a selected number of intermediate,
hidden layers, which can each contain an arbitrary amount of
neurons. Building upon the Kolmogorov’s theorem [
          <xref ref-type="bibr" rid="ref31">32</xref>
          ],
Hecht-Nielsen, Ne, and Corporation [28] proved that using
specific activation functions, even an ANN with a single
hidden layer of neurons can approximate any continuous
function. However, it is not guaranteed that such ANN
can actually learn such representation [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ], and in practice,
using multiple hidden layers of neurons can simplify the
learning process [29].
        </p>
        <p>The training of the network is equivalent to the
determination of appropriate weights w from Equation 5 for each
neuron, such that the desired mapping x ! y from the input
data to the output data is obtained. The common training
method is the backpropagation method, where the
prediction error E after the forward pass is propagated back to
the contribution to that error from each neuron, and their
weights are updated in each training iteration. This is
schematically illustrated in Figure 3.</p>
        <p>Independent of the selected surrogate model, to obtain
an accurate model, a step of crucial importance is the
selection of the model’s hyperparameters, that is the
parameters that are not directly learned through the training
process. This is specific for each of the surrogate
models. As an example, in the case of the ANN, these are
the amount of hidden layers and neurons in each of them,
the learning rate parameter, regularization parameters or
the activation function choice. For this purpose,
modelspecific methods, such as the neuron pruning [56], where
the neurons of low significance are removed, as well as
model-independent, such as a random search, Bayesian
optimization or meta-heuristic optimization approaches,
can be used.</p>
        <p>
          At this stage, a random search and Bayesian
optimization (BO) [
          <xref ref-type="bibr" rid="ref48">45</xref>
          ] are included for the ANN using Keras
Tuner4. A hypermodel is defined with default
hyperparameters stored in a configuration file. Within, these
hyperparameters can be either changed manually, or
specified to be subject of automatic tuning. For ANNs, the
tunable hyperparameters are: number of hidden layers,
number of neurons in each layer, initial learning rate, activation
function and the regularization parameter. The current
implementation of the BO allows only setting a fixed
optimization budget, therefore the selected option is to start
with an initial random sample of 3x times the number
of optimized hyperparameters, and stop at 10x times the
number. The acquisition function is the upper-confidence
bound [
          <xref ref-type="bibr" rid="ref48">45</xref>
          ]. The frequency of hyperparameter
optimization is set as at the first iteration, and then every time the
sample set size increases by 50%.
2.4
        </p>
      </sec>
      <sec id="sec-2-4">
        <title>Surrogate’s convergence</title>
        <p>A metric of the surrogate’s accuracy of choice is to be
tracked to determine the convergence of the sample set
size. In our approach, the tracked metric is the mean
absolute error (MAE)</p>
        <p>EMAE =</p>
        <p>N
åi=1 jypred</p>
        <p>
          N
ytrainj
where ypred is the prediction of the surrogate, ytrain is the
training value and N is the total number of samples. Since
the amount of data is relatively low, to perform validation
of the trained surrogate model, the K-fold cross-validation
approach is adopted [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ]. In each iteration, the data is split
into a set of K-folds, and K surrogates are trained each
time leaving out a different fold of the data. This
holdout set is used to calculate the generalization error of the
model on previously unseen data. With increasing size of
the data set, leaving out the holdout set will affect the
surrogate’s prediction less and less, so the average EMAE over
the K surrogates will decrease. The resampling loop is
terminated when a satisfactory level of EMAE is attained.
        </p>
        <p>
          4keras-tuner, https://github.com/keras-team/
keras-tuner [Accessed on 06/08/2020]
Once the surrogate model has been trained to a desired
level of accuracy, the optimization run can be started.
Within this framework, various population-based
optimization algorithms are provided by the Pymoo library
[
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]. Pymoo provides implementations of a series of
population-based algorithms, including:
        </p>
        <sec id="sec-2-4-1">
          <title>Differential Evolution</title>
        </sec>
        <sec id="sec-2-4-2">
          <title>Genetic Algorithm</title>
        </sec>
        <sec id="sec-2-4-3">
          <title>BRKGA</title>
        </sec>
        <sec id="sec-2-4-4">
          <title>Nelder Mead</title>
        </sec>
        <sec id="sec-2-4-5">
          <title>Pattern Search</title>
        </sec>
        <sec id="sec-2-4-6">
          <title>CMAES</title>
        </sec>
        <sec id="sec-2-4-7">
          <title>NSGA-II and NSGA-III</title>
        </sec>
        <sec id="sec-2-4-8">
          <title>RNSGA-II and RNSGA-III</title>
        </sec>
        <sec id="sec-2-4-9">
          <title>UNSGA-III</title>
        </sec>
        <sec id="sec-2-4-10">
          <title>MOEA/D.</title>
          <p>
            Among those, the most commonly used in the industry is
the NSGA-II method [
            <xref ref-type="bibr" rid="ref15">15</xref>
            ] and its updated NSGA-III
version [
            <xref ref-type="bibr" rid="ref16">16</xref>
            ] adapted for the many-objective cases of m &gt; 2,
where an even exploration of the Pareto-front is
accomplished by the use of reference directions. A generic
scheme of a genetic algorithm is shown in Figure 4. The
idea behind is similar to Darwinian evolution, that is
that once the initial population of solutions is generated
(Step 1), the fitness of the individuals is evaluated (Step
2) and only the fittest survive in each generation (Step
3). From those, a selection parent selection takes places
(Step 4), from which new offsprings are generated through
crossover (Step 5). Additionally, a random mutation is
applied to their genome (Step 6) to maintain diversity within
the population and explore the design space better. At the
end of each iteration, the convergence criterion is
evaluated, and either the optimization stops, or continues with a
new generation [52].
          </p>
          <p>The available termination criteria are the maximum:
number of evaluations
number of generations
time
design space tolerance
objective space tolerance.</p>
          <p>Due to the computational cheapness of the surrogate, the
optimization process is not limited by the computational
cost of evaluating the simulation, such as the maximal
number of its evaluations. Therefore, the selection of the
termination criterion can focus purely on the convergence
of the optimization. In this case the last two criteria are
of main importance. A threshold tolerance on the metrics
tracking the evolution of solutions in the design or
objective space is defined, and if the solutions in the specified
amount of last iterations do not improve over this
threshold, the optimization is terminated. For robustness
purposes, also the maximal amount of evaluations or
generations can be specified, e.g. for cases when the optimal
solutions oscillate around some value, and thus in the
defined window would otherwise never have converged.</p>
          <p>It is also possible to enter the optimization directly using
the original model without constructing the surrogate. In
this case however, for an expensive simulation, the choices
on the optimization algorithm and its components,
especially the population size and number of generated
offsprings in each generation, have to be more carefully
selected with respect to the number of required simulations.
2.6</p>
        </sec>
      </sec>
      <sec id="sec-2-5">
        <title>Result verification</title>
        <p>Once an optimum (set) has been found, given that this
optimum has been found based on the surrogate, it is
important to verify it’s accuracy. In particular, it could be that the
algorithm has found an optimum that lies in a spurious
extremum caused by the surrogate’s inaccuracy which does
not pertain to the original model. Therefore, by submitting
a sample from the Pareto optimal set back to an evaluation
using the original model, it can be verified whether the
obtained solutions indeed pertain to the original model. If
the mean or maximal error within this verification exceeds
a pre-defined tolerance, the surrogate is retrained on an
enlarged sample, and the optimization occurs again, as seen
in Figure 1.</p>
      </sec>
      <sec id="sec-2-6">
        <title>2.7 Implementation notes</title>
        <p>
          As an important implementation note, the data is
normalized within the internal data flow. This is done such that
common settings can be used across different problems,
as well as for a single problem to treat inputs and
outputs of different magnitudes comparably. Not doing so
would mean that for example when the prediction error is
summed, the weight in kilograms would completely
dominate deformation in millimeters, which is undesirable.
Therefore, all the inputs and outputs are normalized by
dividing each quantity with its absolute maximal value. In
such a way, all the data is guaranteed to lie within the
[
          <xref ref-type="bibr" rid="ref1 ref1">1,1</xref>
          ] range, while maintaining its sign, conforming to the
constraint violation formulation in Equation 2.
3
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Illustrative example</title>
      <p>As an illustrative example of a practical problem, the
results on a parametrized nonlinear FEM simulation of a car
side impact problem [30] ran using the proposed
framework are presented in this section.
3.1</p>
      <sec id="sec-3-1">
        <title>Model description</title>
        <p>The design space is defined by 7 input parameters
0:45</p>
        <p>x2
0:875</p>
        <p>x5
0:5
0:5
0:5
0:4
0:4
x1
x3
x4
x6
x7
1:5
1:35
1:5
1:5
2:625
1:2
1:2
that represent the thicknesses of structural members such
as B-pillars, beams or the floor. A single solution was,
at the time of the problem’s publication, reported to take
about 20h [54]. Even thought this time has likely reduced
with today’s hardware, it is still a good illustrative example
of surrogate-based modeling.</p>
        <p>For the purpose of benchmarking, the response has been
parametrized by analytical expressions, which are
presented hereafter [30]. The 3 objectives of the design study
are to minimize the structural weight, impact force on the
passenger and average velocity of the side members that
absorbs the impact load, represented as
f1(~x)</p>
        <p>
          W = 1:98 + 4:9x1 + 6:67x2 + 6:98x3 + 4:01x4
+1:78x5 + 0:00001x6 + 2:73x7
f2(~x) = F
f3(~x) = 0:5(VMBP + VFD)
To translate it, the carside problem is evaluated from the
benchmark problems set. The initial sample is 10x the
number of input dimensions, thus 70, and in every
additional iteration, the sample set is increased by 20%.
The convergence metric for the model is the MAE
metric, with a maximal value of 0.03. The ANN is the used
surrogate model and its performance metric is calculated
using 5-fold cross validation. The hyperparameters are
re-optimized when the amount of samples increases by
50%. The optimization uses the NSGA-III algorithm and
its termination is based upon the objective space
tolerance, which terminates, if there is no change by more
than 0.001 in the tracked metrics: changes of the objective
functions, inverted generational distance (IGD) [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ] and the
nadir point, over the last 5 generations. This convergence
check is performed every 5 generations. Additionally, a
maximum of 150 generations is allowed during a single
optimization run. The verification mean error limit is set
as 5%.
the algorithm is initiated with default settings7. The
reference directions are obtained using an energy method from
Blank et al. [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] and their number, which also defines the
population size, is set at 30x the number of objectives, that
is 90 in this case. The amount of generated offsprings is
equal to the population size.
3.3
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>Results</title>
        <p>In Figure 5, the comparison of training data with the
prediction of the final trained ANN is shown.</p>
        <p>The sample set converged after 3 surrogate training
iterations, that is with 101 samples evaluated. This is just over
the defined population size in one generation of the
genetic algorithm. The optimization took 70 generations to
converge, which amount to 9000 evaluations of the model.
That as a factor of 90x more than the required surrogate
training sample size, confirming the advantage of using the
surrogate. It is noted however, that thus far the selected
optimization parameters are determined empirically on a
trial-and-error basis.</p>
        <p>As a validation of the result, the numerical values of the
Pareto optimal solutions are unfortunately not presented in
Jain and Deb [30] nor other paper, however, the qualitative
comparison with their solutions in the objective space as
shown in Figure 6 indicates that the ranges of all 3
objectives fall within nearly identical ranges. The final mean
verification error between the Pareto optimal set and the
original model, as discussed in subsection 2.6, was 2.10%.
4</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Conclusion</title>
      <p>In this paper, a working core example of a modular
surrogate-based optimization framework was presented.
A more detailed elaboration on the ANN surrogate was
7pymoo - NSGA-III, http://pymoo.org/algorithms/nsga3.
html [Accessed on 07/08/2020]
drawn, together with a top-level description of each of
the key submodules. An illustrative working example was
shown, that demonstrated the interaction and use of the
framework together with the efficiency of the surrogate
based approach. Compared to commercial solutions, the
modularity makes it suitable for research purposes, while
the open-source nature is beneficiary for small companies
or users with advanced customization needs.</p>
      <p>
        On top of the correct implementation of the methods,
the main challenges related to its use include the
selection of a particular surrogate and optimization algorithm
or the termination, model selection and convergence
criteria. On a fully general scale, the no free lunch (NFL)
theorems [
        <xref ref-type="bibr" rid="ref55">51</xref>
        ] prove that an universal optimal choice is
impossible. However, if the set of problems is limited to an
expected class of practical engineering problems, at least
a good baseline starting point is possible [53]. Thus, the
main aim of the follow-up work is to perform testing on a
wide range of problems with a range of framework setups
to thoroughly establish the influence of the tunable
parameters of the surrogate model as well as of the optimization
algorithm.
      </p>
      <p>Further suggestions include a quantitative comparative
study of the performance of different proposed adaptive
sampling methods, which show the largest potential on the
overall SBO performance, as the expensive simulations
remains main bottleneck overall. Beyond that,
softwarewise, additional suggested features include a graphical
user interface (GUI), inclusion of more surrogate models
from SMT, such as the RBF, extending the amount of
supported evaluators e.g. for Abaqus, and others. Finally, for
practice oriented research, a composite layup optimization
study in terms of the number of plies and their ply angles
is an interesting area of possible application of this
framework within the field of advanced structures.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Martin</given-names>
            <surname>Abadi</surname>
          </string-name>
          et al. “
          <article-title>TensorFlow: A system for large-scale machine learning”</article-title>
          .
          <source>In: 12th USENIX Symposium on Operating Systems Design and Implementation (OSDI 16)</source>
          .
          <year>2016</year>
          , pp.
          <fpage>265</fpage>
          -
          <lpage>283</lpage>
          . URL: https : / / www . usenix . org / system / files / conference/osdi16/osdi16-abadi.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Charles</given-names>
            <surname>Audet</surname>
          </string-name>
          et al. “
          <article-title>Performance indicators in multiobjective optimization”. working paper or preprint</article-title>
          .
          <source>Feb</source>
          .
          <year>2020</year>
          . URL: https : / / hal . archives-ouvertes.fr/hal-02464750.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Bessa</surname>
          </string-name>
          and
          <string-name>
            <given-names>S.</given-names>
            <surname>Pellegrino</surname>
          </string-name>
          . “
          <article-title>Design of ultrathin shell structures in the stochastic post-buckling range using Bayesian machine learning and optimization”</article-title>
          .
          <source>In: International Journal of Solids and Structures</source>
          <volume>139</volume>
          -140 (May
          <year>2018</year>
          ), pp.
          <fpage>174</fpage>
          -
          <lpage>188</lpage>
          . DOI:
          <volume>10</volume>
          .1016/j.ijsolstr.
          <year>2018</year>
          .
          <volume>01</volume>
          .035.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Bessa</surname>
          </string-name>
          et al. “
          <article-title>A framework for data-driven analysis of materials under uncertainty: Countering the curse of dimensionality”</article-title>
          .
          <source>In: Computer Methods in Applied Mechanics and Engineering</source>
          <volume>320</volume>
          (
          <year>June 2017</year>
          ), pp.
          <fpage>633</fpage>
          -
          <lpage>667</lpage>
          . DOI:
          <volume>10</volume>
          .1016/j.cma.
          <year>2017</year>
          .
          <volume>03</volume>
          .037.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Atharv</given-names>
            <surname>Bhosekar</surname>
          </string-name>
          and
          <string-name>
            <given-names>Marianthi</given-names>
            <surname>Ierapetritou</surname>
          </string-name>
          . “
          <article-title>Advances in surrogate based modeling, feasibility analysis, and optimization: A review”</article-title>
          .
          <source>In: Computers &amp; Chemical Engineering</source>
          <volume>108</volume>
          (Jan.
          <year>2018</year>
          ), pp.
          <fpage>250</fpage>
          -
          <lpage>267</lpage>
          . DOI:
          <volume>10</volume>
          . 1016 / j . compchemeng .
          <year>2017</year>
          .
          <volume>09</volume>
          .017.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Christopher</surname>
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Bishop</surname>
          </string-name>
          .
          <source>Pattern Recognition and Machine Learning</source>
          . Springer-Verlag New York Inc.,
          <source>Aug</source>
          .
          <volume>17</volume>
          ,
          <year>2006</year>
          . 738 pp.
          <source>ISBN: 0387310738.</source>
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Julian</given-names>
            <surname>Blank</surname>
          </string-name>
          and
          <string-name>
            <given-names>Kalyanmoy</given-names>
            <surname>Deb</surname>
          </string-name>
          .
          <article-title>“pymoo: Multiobjective Optimization in Python”</article-title>
          .
          <source>In: IEEE Access 8</source>
          (
          <year>2020</year>
          ), pp.
          <fpage>89497</fpage>
          -
          <lpage>89509</lpage>
          . DOI:
          <volume>10</volume>
          . 1109 / access.
          <year>2020</year>
          .
          <volume>2990567</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Julian</given-names>
            <surname>Blank</surname>
          </string-name>
          et al. “
          <article-title>Generating Well-Spaced Points on a Unit Simplex for Evolutionary ManyObjective Optimization”</article-title>
          .
          <source>In: IEEE Transactions on Evolutionary Computation</source>
          (
          <year>2020</year>
          ), pp.
          <fpage>1</fpage>
          -
          <lpage>1</lpage>
          . DOI:
          <volume>10</volume>
          .1109/tevc.
          <year>2020</year>
          .
          <volume>2992387</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Avrim</given-names>
            <surname>Blum</surname>
          </string-name>
          , Adam Kalai,
          <string-name>
            <surname>and John Langford.</surname>
          </string-name>
          “
          <article-title>Beating the Hold-Out: Bounds for K-fold and Progressive Cross-Validation”</article-title>
          .
          <source>In: Proceedings of the twelfth annual conference on Computational learning theory - COLT '99</source>
          . ACM Press,
          <year>1999</year>
          . DOI:
          <volume>10</volume>
          . 1145/307400.307439.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>Mohamed</given-names>
            <surname>Amine</surname>
          </string-name>
          Bouhlel et al. “
          <article-title>A Python surrogate modeling framework with derivatives”</article-title>
          .
          <source>In: Advances in Engineering Software</source>
          <volume>135</volume>
          (
          <issue>Sept</issue>
          .
          <year>2019</year>
          ), p.
          <fpage>102662</fpage>
          . DOI:
          <volume>10</volume>
          .1016/j.advengsoft.
          <year>2019</year>
          .
          <volume>03</volume>
          .005.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Luis</surname>
            <given-names>E.</given-names>
          </string-name>
          <string-name>
            <surname>Garza</surname>
            <given-names>Castañón</given-names>
          </string-name>
          , Saúl Montes de Oca, and
          <string-name>
            <surname>Rubén</surname>
          </string-name>
          Morales-Menéndez.
          <article-title>“An Application of Random and Hammersley Sampling Methods to Iris Recognition”</article-title>
          .
          <source>In: Advances in Applied Artificial Intelligence. Ed</source>
          . by
          <string-name>
            <given-names>M.</given-names>
            <surname>Ali</surname>
          </string-name>
          and
          <string-name>
            <given-names>R.</given-names>
            <surname>Dapoigny</surname>
          </string-name>
          . Vol.
          <volume>4031</volume>
          . LNAI. Springer Berlin Heidelberg,
          <year>2006</year>
          , pp.
          <fpage>520</fpage>
          -
          <lpage>529</lpage>
          . DOI:
          <volume>10</volume>
          .1007/11779568_
          <fpage>56</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12] Kai Cheng et al. “
          <article-title>Surrogate-assisted global sensitivity analysis: an overview”</article-title>
          .
          <source>In: Structural and Multidisciplinary Optimization</source>
          <volume>61</volume>
          .3 (
          <issue>Jan</issue>
          .
          <year>2020</year>
          ), pp.
          <fpage>1187</fpage>
          -
          <lpage>1213</lpage>
          . DOI:
          <volume>10</volume>
          . 1007 / s00158 - 019 - 02413-5.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>H.</given-names>
            <surname>Chi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mascagni</surname>
          </string-name>
          , and
          <string-name>
            <given-names>T.</given-names>
            <surname>Warnock</surname>
          </string-name>
          . “
          <article-title>On the optimal Halton sequence”</article-title>
          .
          <source>In: Mathematics and Computers in Simulation 70.1</source>
          (
          <issue>Sept</issue>
          .
          <year>2005</year>
          ), pp.
          <fpage>9</fpage>
          -
          <lpage>21</lpage>
          . DOI:
          <volume>10</volume>
          .1016/j.matcom.
          <year>2005</year>
          .
          <volume>03</volume>
          .004.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>Keith</given-names>
            <surname>Dalbey</surname>
          </string-name>
          and
          <string-name>
            <given-names>George</given-names>
            <surname>Karystinos</surname>
          </string-name>
          . “
          <article-title>Fast Generation of Space-filling Latin Hypercube Sample Designs”</article-title>
          .
          <source>In: 13th AIAA/ISSMO Multidisciplinary Analysis Optimization Conference. American Institute of Aeronautics and Astronautics</source>
          , Sept.
          <year>2010</year>
          . DOI:
          <volume>10</volume>
          .2514/6.2010-
          <volume>9085</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>K.</given-names>
            <surname>Deb</surname>
          </string-name>
          et al. “
          <article-title>A fast and elitist multiobjective genetic algorithm: NSGA-II”</article-title>
          .
          <source>In: IEEE Transactions on Evolutionary Computation 6</source>
          .2 (
          <issue>Apr</issue>
          .
          <year>2002</year>
          ), pp.
          <fpage>182</fpage>
          -
          <lpage>197</lpage>
          . DOI:
          <volume>10</volume>
          .1109/4235.996017.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>Kalyanmoy</given-names>
            <surname>Deb</surname>
          </string-name>
          and
          <string-name>
            <given-names>Himanshu</given-names>
            <surname>Jain</surname>
          </string-name>
          . “
          <article-title>An Evolutionary Many-Objective Optimization Algorithm Using Reference-Point-Based Nondominated Sorting Approach, Part I: Solving Problems With Box Constraints”</article-title>
          .
          <source>In: IEEE Transactions on Evolutionary Computation</source>
          <volume>18</volume>
          .4 (
          <issue>Aug</issue>
          .
          <year>2014</year>
          ), pp.
          <fpage>577</fpage>
          -
          <lpage>601</lpage>
          . DOI:
          <volume>10</volume>
          .1109/tevc.
          <year>2013</year>
          .
          <volume>2281535</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>John</given-names>
            <surname>Eason</surname>
          </string-name>
          and Selen Cremaschi. “
          <article-title>Adaptive sequential sampling for surrogate model generation with artificial neural networks”</article-title>
          .
          <source>In: Computers &amp; Chemical Engineering</source>
          <volume>68</volume>
          (Sept.
          <year>2014</year>
          ), pp.
          <fpage>220</fpage>
          -
          <lpage>232</lpage>
          . DOI:
          <volume>10</volume>
          .1016/j.compchemeng.
          <year>2014</year>
          .
          <volume>05</volume>
          . 021.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Ke</surname>
            <given-names>Fang</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yuchen Zhou</surname>
          </string-name>
          , and Ping Ma. “
          <article-title>An adaptive sequential experiment design method for model validation”</article-title>
          .
          <source>In: Chinese Journal of Aeronautics 33.6 (June</source>
          <year>2020</year>
          ), pp.
          <fpage>1661</fpage>
          -
          <lpage>1672</lpage>
          . DOI:
          <volume>10</volume>
          . 1016 / j . cja.
          <year>2019</year>
          .
          <volume>12</volume>
          .026.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>Henri</given-names>
            <surname>Faure</surname>
          </string-name>
          and
          <string-name>
            <given-names>Christiane</given-names>
            <surname>Lemieux</surname>
          </string-name>
          . “Generalized Halton sequences in
          <year>2008</year>
          ”. In: vol.
          <volume>19</volume>
          . 4. Association for Computing Machinery (ACM), Oct.
          <year>2009</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>31</lpage>
          . DOI:
          <volume>10</volume>
          .1145/1596519.1596520.
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>Alexander</surname>
            <given-names>I. J.</given-names>
          </string-name>
          <string-name>
            <surname>Forrester</surname>
            and
            <given-names>Andy J.</given-names>
          </string-name>
          <string-name>
            <surname>Keane</surname>
          </string-name>
          . “
          <article-title>Recent advances in surrogate-based optimization”</article-title>
          .
          <source>In: Progress in Aerospace Sciences 45.1-3 (Jan</source>
          .
          <year>2009</year>
          ), pp.
          <fpage>50</fpage>
          -
          <lpage>79</lpage>
          . DOI:
          <volume>10</volume>
          . 1016 / j . paerosci .
          <year>2008</year>
          .
          <volume>11</volume>
          .001.
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          <article-title>“Minimum-weight design for three dimensional woven composite stiffened panels using neural networks and genetic algorithms”</article-title>
          .
          <source>In: Composite Structures</source>
          <volume>134</volume>
          (
          <issue>Dec</issue>
          .
          <year>2015</year>
          ), pp.
          <fpage>708</fpage>
          -
          <lpage>715</lpage>
          . DOI: 10.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          1016/j.compstruct.
          <year>2015</year>
          .
          <volume>08</volume>
          .077.
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [22]
          <string-name>
            <surname>Sushant</surname>
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Garud</surname>
          </string-name>
          , Iftekhar A.
          <string-name>
            <surname>Karimi</surname>
          </string-name>
          , and Markus Kraft. “
          <article-title>Design of computer experiments: A review”</article-title>
          .
          <source>In: Computers &amp; Chemical Engineering</source>
          <volume>106</volume>
          (Nov.
          <year>2017</year>
          ), pp.
          <fpage>71</fpage>
          -
          <lpage>95</lpage>
          . DOI:
          <volume>10</volume>
          . 1016 / j . compchemeng.
          <year>2017</year>
          .
          <volume>05</volume>
          .010.
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          <source>[23] [24] [25] [26] [27] [28]</source>
          [29]
          <string-name>
            <given-names>Xavier</given-names>
            <surname>Glorot</surname>
          </string-name>
          and
          <string-name>
            <given-names>Yoshua</given-names>
            <surname>Bengio</surname>
          </string-name>
          .
          <article-title>“Understanding the difficulty of training deep feedforward neural networks”</article-title>
          .
          <source>In: Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics. Ed. by Yee Whye Teh and Mike Titterington</source>
          . Vol.
          <volume>9</volume>
          .
          <source>Proceedings of Machine Learning Research. Chia Laguna Resort</source>
          , Sardinia, Italy: PMLR, May
          <year>2010</year>
          , pp.
          <fpage>249</fpage>
          -
          <lpage>256</lpage>
          . URL: http : / / proceedings . mlr . press / v9 / glorot10a .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          <string-name>
            <surname>Ed. by M. Daydé</surname>
          </string-name>
          et al. Vol.
          <volume>4395</volume>
          . LNCS. Springer Berlin Heidelberg,
          <year>2007</year>
          , pp.
          <fpage>579</fpage>
          -
          <lpage>588</lpage>
          . DOI: 10 .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          <string-name>
            <given-names>Qi</given-names>
            <surname>Guo</surname>
          </string-name>
          et al. “
          <article-title>Thermo-mechanical optimization of metallic thermal protection system under aerodynamic heating”</article-title>
          .
          <source>In: Structural and Multidisciplinary Optimization (Aug</source>
          .
          <year>2019</year>
          ).
          <source>DOI: 10.1007/ s00158-019-02379-4.</source>
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          <string-name>
            <surname>“</surname>
          </string-name>
          <article-title>An objective reduction algorithm using representative Pareto solution search for many-objective optimization problems”</article-title>
          .
          <source>In: Soft Computing 20.12 (July</source>
          <year>2015</year>
          ), pp.
          <fpage>4881</fpage>
          -
          <lpage>4895</lpage>
          . DOI:
          <volume>10</volume>
          . 1007 / s00500 - 015-1776-4.
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          <string-name>
            <given-names>Warren</given-names>
            <surname>Hare</surname>
          </string-name>
          , Julie Nutini, and Solomon Tesfamariam.
          <article-title>“A survey of non-gradient optimization methods in structural engineering”</article-title>
          .
          <source>In: Advances in Engineering Software</source>
          <volume>59</volume>
          (May
          <year>2013</year>
          ), pp.
          <fpage>19</fpage>
          -
          <lpage>28</lpage>
          . DOI:
          <volume>10</volume>
          .1016/j.advengsoft.
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          <string-name>
            <given-names>Robert</given-names>
            <surname>Hecht-Nielsen</surname>
          </string-name>
          ,
          <article-title>Hecht-Nielsen Ne, and urocomputer Corporation. “Kolmogorov's Mapping Neural Network Existence Theorem”</article-title>
          .
          <source>In: Proceedings of the international conference on Neural Networks</source>
          . Vol.
          <volume>3</volume>
          . IEEE Press New York.
          <year>1987</year>
          , pp.
          <fpage>11</fpage>
          -
          <lpage>14</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          <string-name>
            <surname>Guang-Bin Huang</surname>
          </string-name>
          . “
          <article-title>Learning capability and storage capacity of two-hidden-layer feedforward networks”</article-title>
          .
          <source>In: IEEE Transactions on Neural Networks 14.2</source>
          (
          <issue>Mar</issue>
          .
          <year>2003</year>
          ), pp.
          <fpage>274</fpage>
          -
          <lpage>281</lpage>
          . DOI:
          <volume>10</volume>
          . 1109 / tnn.
          <year>2003</year>
          .
          <volume>809401</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [32] [35] [36] [37]
          <string-name>
            <given-names>Himanshu</given-names>
            <surname>Jain</surname>
          </string-name>
          and
          <string-name>
            <given-names>Kalyanmoy</given-names>
            <surname>Deb</surname>
          </string-name>
          . “
          <article-title>An Evolutionary Many-Objective Optimization Algorithm Using Reference-Point Based Nondominated Sorting Approach</article-title>
          ,
          <string-name>
            <surname>Part</surname>
            <given-names>II</given-names>
          </string-name>
          :
          <article-title>Handling Constraints and Extending to an Adaptive Approach”</article-title>
          .
          <source>In: IEEE Transactions on Evolutionary Computation</source>
          <volume>18</volume>
          .
          <article-title>4 (Aug</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          <year>2014</year>
          ), pp.
          <fpage>602</fpage>
          -
          <lpage>622</lpage>
          . DOI:
          <volume>10</volume>
          .1109/tevc.
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          <string-name>
            <given-names>Ruichen</given-names>
            <surname>Jin</surname>
          </string-name>
          , Wei Chen, and Agus Sudjianto. “
          <article-title>On Sequential Sampling for Global Metamodeling in Engineering Design”</article-title>
          .
          <source>In: Volume</source>
          <volume>2</volume>
          : 28th Design Automation Conference. ASMEDC, Jan.
          <year>2002</year>
          . DOI:
          <volume>10</volume>
          .1115/detc2002/dac-34092.
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          <string-name>
            <given-names>A. N.</given-names>
            <surname>Kolmogorov</surname>
          </string-name>
          . “
          <article-title>On the representation of continuous functions of many variables by superposition of continuous functions of one variable and addition”</article-title>
          .
          <source>In: Doklady Akademii Nauk SSSR 114.5</source>
          (
          <issue>1957</issue>
          ), pp.
          <fpage>953</fpage>
          -
          <lpage>956</lpage>
          . URL: http://mi.mathnet.
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>ru/eng/dan22050.</mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          [33]
          <string-name>
            <given-names>Slawomir</given-names>
            <surname>Koziel</surname>
          </string-name>
          and
          <string-name>
            <given-names>Stanislav</given-names>
            <surname>Ogurtsov</surname>
          </string-name>
          . “
          <article-title>MultiObjective Design of Antennas Using VariableFidelity Simulations and Surrogate Models”</article-title>
          .
          <source>In: IEEE Transactions on Antennas and Propagation</source>
          <volume>61</volume>
          .12 (
          <issue>Dec</issue>
          .
          <year>2013</year>
          ), pp.
          <fpage>5931</fpage>
          -
          <lpage>5939</lpage>
          . DOI:
          <volume>10</volume>
          .1109/ tap.
          <year>2013</year>
          .
          <volume>2283599</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          [34]
          <string-name>
            <given-names>L.</given-names>
            <surname>Lanzi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Bisagni</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S.</given-names>
            <surname>Ricci</surname>
          </string-name>
          . “
          <article-title>Crashworthiness optimization of helicopter subfloor based on decomposition and global approximation”</article-title>
          .
          <source>In: Structural and Multidisciplinary Optimization 27.5 (June</source>
          <year>2004</year>
          ). DOI:
          <volume>10</volume>
          .1007/s00158- 004- 0394- z.
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          <string-name>
            <given-names>Xin</given-names>
            <surname>Liu</surname>
          </string-name>
          and
          <string-name>
            <given-names>Albert C.</given-names>
            <surname>Reynolds</surname>
          </string-name>
          . “
          <article-title>Gradient-based multi-objective optimization with applications to waterflooding optimization”</article-title>
          .
          <source>In: Computational Geosciences</source>
          <volume>20</volume>
          .3 (
          <issue>Sept</issue>
          .
          <year>2015</year>
          ), pp.
          <fpage>677</fpage>
          -
          <lpage>693</lpage>
          . DOI:
          <volume>10</volume>
          .1007/s10596-015-9523-6.
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          <source>In: Structural and Multidisciplinary Optimization</source>
          <volume>26</volume>
          .6 (
          <issue>Apr</issue>
          .
          <year>2004</year>
          ), pp.
          <fpage>369</fpage>
          -
          <lpage>395</lpage>
          . DOI:
          <volume>10</volume>
          . 1007 / s00158-003-0368-6.
        </mixed-citation>
      </ref>
      <ref id="ref40">
        <mixed-citation>
          <string-name>
            <surname>Art</surname>
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Owen</surname>
          </string-name>
          . “
          <article-title>A randomized Halton algorithm in R”</article-title>
          .
          <source>In: June 9</source>
          ,
          <year>2017</year>
          . arXiv:
          <volume>1706</volume>
          .
          <article-title>02808v2 [stat</article-title>
          .CO].
        </mixed-citation>
      </ref>
      <ref id="ref41">
        <mixed-citation>
          [38]
          <string-name>
            <given-names>Pramudita</given-names>
            <surname>Satria</surname>
          </string-name>
          Palar et al. “
          <article-title>On the use of surrogate models in engineering design optimization and exploration”</article-title>
          .
          <source>In: Proceedings of the Genetic and Evolutionary Computation Conference Companion. ACM</source>
          ,
          <year>July 2019</year>
          . DOI:
          <volume>10</volume>
          . 1145 / 3319619 . 3326813.
        </mixed-citation>
      </ref>
      <ref id="ref42">
        <mixed-citation>
          [39]
          <string-name>
            <given-names>Sebastian</given-names>
            <surname>Peitz</surname>
          </string-name>
          and
          <string-name>
            <given-names>Michael</given-names>
            <surname>Dellnitz</surname>
          </string-name>
          . “
          <article-title>GradientBased Multiobjective Optimization with Uncertainties”</article-title>
          .
          <source>In: NEO 2016</source>
          . Springer International Publishing, Sept.
          <year>2017</year>
          , pp.
          <fpage>159</fpage>
          -
          <lpage>182</lpage>
          . DOI:
          <volume>10</volume>
          .1007/978- 3-
          <fpage>319</fpage>
          -64063-
          <issue>1</issue>
          _
          <fpage>7</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref43">
        <mixed-citation>
          [40]
          <string-name>
            <surname>Prajit</surname>
            <given-names>Ramachandran</given-names>
          </string-name>
          , Barret Zoph, and
          <string-name>
            <surname>Quoc</surname>
            <given-names>V.</given-names>
          </string-name>
          <string-name>
            <surname>Le</surname>
          </string-name>
          . “
          <article-title>Searching for Activation Functions”</article-title>
          . In: CoRR (Oct.
          <volume>16</volume>
          ,
          <year>2017</year>
          ). arXiv: http://arxiv.org/abs/ 1710.05941v2 [cs.NE].
        </mixed-citation>
      </ref>
      <ref id="ref44">
        <mixed-citation>
          [41]
          <string-name>
            <given-names>Songqing</given-names>
            <surname>Shan</surname>
          </string-name>
          and
          <string-name>
            <given-names>G. Gary</given-names>
            <surname>Wang</surname>
          </string-name>
          . “
          <article-title>Survey of modeling and optimization strategies to solve high-dimensional design problems with computationally-expensive black-box functions”</article-title>
          .
          <source>In: Structural and Multidisciplinary Optimization</source>
          <volume>41</volume>
          .2 (
          <issue>Aug</issue>
          .
          <year>2009</year>
          ), pp.
          <fpage>219</fpage>
          -
          <lpage>241</lpage>
          . DOI:
          <volume>10</volume>
          .1007/s00158-009-0420-2.
        </mixed-citation>
      </ref>
      <ref id="ref45">
        <mixed-citation>
          [42]
          <string-name>
            <given-names>T. W.</given-names>
            <surname>Simpson</surname>
          </string-name>
          et al. “
          <article-title>Metamodels for Computerbased Engineering Design: Survey and recommendations”</article-title>
          .
          <source>In: Engineering with Computers 17.2 (July</source>
          <year>2001</year>
          ), pp.
          <fpage>129</fpage>
          -
          <lpage>150</lpage>
          . DOI:
          <volume>10</volume>
          . 1007 / pl00007198.
        </mixed-citation>
      </ref>
      <ref id="ref46">
        <mixed-citation>
          [43]
          <string-name>
            <surname>Prashant</surname>
            <given-names>Singh</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Dirk</given-names>
            <surname>Deschrijver</surname>
          </string-name>
          , and Tom Dhaene.
          <article-title>“A balanced sequential design strategy for global surrogate modeling”</article-title>
          .
          <source>In: 2013 Winter Simulations Conference (WSC)</source>
          . IEEE, Dec.
          <year>2013</year>
          . DOI:
          <volume>10</volume>
          . 1109/wsc.
          <year>2013</year>
          .
          <volume>6721594</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref47">
        <mixed-citation>
          [44]
          <string-name>
            <given-names>J.</given-names>
            <surname>Sobieszczanski-Sobieski</surname>
          </string-name>
          .
          <article-title>Structural Optimization: Challenges and Opportunities</article-title>
          .
          <source>NASA (National Aeronautics and Space Administration)</source>
          ,
          <year>1984</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref48">
        <mixed-citation>
          [45] [46] [47]
          <string-name>
            <given-names>Niranjan</given-names>
            <surname>Srinivas</surname>
          </string-name>
          et al. “
          <article-title>Gaussian Process Optimization in the Bandit Setting: No Regret and Experimental Design”</article-title>
          .
          <source>In: Proceedings of the 27th International Conference on International Conference on Machine Learning. ICML'10</source>
          .
          <string-name>
            <surname>Haifa</surname>
          </string-name>
          , Israel: Omnipress,
          <year>2010</year>
          , pp.
          <fpage>1015</fpage>
          -
          <lpage>1022</lpage>
          . ISBN:
          <volume>9781605589077</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref49">
        <mixed-citation>
          <string-name>
            <given-names>Ryoji</given-names>
            <surname>Tanabe</surname>
          </string-name>
          and
          <string-name>
            <given-names>Hisao</given-names>
            <surname>Ishibuchi</surname>
          </string-name>
          . “
          <article-title>An easy-touse real-world multi-objective optimization problem suite”</article-title>
          . In: vol.
          <volume>89</volume>
          .
          <string-name>
            <surname>Elsevier</surname>
            <given-names>BV</given-names>
          </string-name>
          ,
          <year>Apr</year>
          .
          <year>2020</year>
          , p.
          <fpage>106078</fpage>
          . DOI:
          <volume>10</volume>
          .1016/j.asoc.
          <year>2020</year>
          .
          <volume>106078</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref50">
        <mixed-citation>
          <string-name>
            <given-names>G. N.</given-names>
            <surname>Vanderplaats</surname>
          </string-name>
          . “
          <article-title>Thirty years of modern structural optimization”</article-title>
          .
          <source>In: Advances in Engineering Software 16.2</source>
          (
          <issue>Jan</issue>
          .
          <year>1993</year>
          ), pp.
          <fpage>81</fpage>
          -
          <lpage>88</lpage>
          . DOI: 10 .
        </mixed-citation>
      </ref>
      <ref id="ref51">
        <mixed-citation>
          <volume>1016</volume>
          /
          <fpage>0965</fpage>
          -
          <lpage>9978</lpage>
          (
          <issue>93</issue>
          )
          <fpage>90052</fpage>
          -
          <lpage>u</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref52">
        <mixed-citation>
          [48]
          <string-name>
            <given-names>S.</given-names>
            <surname>Venkataraman</surname>
          </string-name>
          and
          <string-name>
            <given-names>R. T.</given-names>
            <surname>Haftka</surname>
          </string-name>
          . “
          <article-title>Structural optimization complexity: what has Moore?s law done for us?</article-title>
          ”
          <source>In: Structural and Multidisciplinary Optimization</source>
          <volume>28</volume>
          .6 (
          <issue>Sept</issue>
          .
          <year>2004</year>
          ), pp.
          <fpage>375</fpage>
          -
          <lpage>387</lpage>
          . DOI:
          <volume>10</volume>
          . 1007/s00158-004-0415-y.
        </mixed-citation>
      </ref>
      <ref id="ref53">
        <mixed-citation>
          [49]
          <string-name>
            <surname>Felipe</surname>
            <given-names>A. C.</given-names>
          </string-name>
          <string-name>
            <surname>Viana</surname>
          </string-name>
          et al. “
          <article-title>Special Section on Multidisciplinary Design Optimization: Metamodeling in Multidisciplinary Design Optimization: How Far Have We Really Come?”</article-title>
          <source>In: AIAA Journal 52.4</source>
          (
          <issue>Apr</issue>
          .
          <year>2014</year>
          ), pp.
          <fpage>670</fpage>
          -
          <lpage>690</lpage>
          . DOI:
          <volume>10</volume>
          . 2514 / 1 . j052375.
        </mixed-citation>
      </ref>
      <ref id="ref54">
        <mixed-citation>
          [50]
          <string-name>
            <given-names>Ivan</given-names>
            <surname>Voutchkov</surname>
          </string-name>
          and
          <string-name>
            <given-names>Andy</given-names>
            <surname>Keane</surname>
          </string-name>
          .
          <article-title>“Multi-Objective Optimization Using Surrogates”</article-title>
          . In: Computational Intelligence in Optimization. Ed. by
          <string-name>
            <given-names>Y.</given-names>
            <surname>Tenne</surname>
          </string-name>
          and C.
          <article-title>-</article-title>
          K. Goh. Springer Berlin Heidelberg,
          <year>2010</year>
          , pp.
          <fpage>155</fpage>
          -
          <lpage>175</lpage>
          . DOI:
          <volume>10</volume>
          .1007/978-3-
          <fpage>642</fpage>
          -12775-
          <issue>5</issue>
          _
          <fpage>7</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref55">
        <mixed-citation>
          <source>[51] [52] [53] [54] [55]</source>
          [56]
          <string-name>
            <given-names>D. H.</given-names>
            <surname>Wolpert</surname>
          </string-name>
          and
          <string-name>
            <surname>W. G. Macready.</surname>
          </string-name>
          “
          <article-title>No free lunch theorems for optimization”</article-title>
          .
          <source>In: IEEE Transactions on Evolutionary Computation 1</source>
          .1 (
          <issue>Apr</issue>
          .
        </mixed-citation>
      </ref>
      <ref id="ref56">
        <mixed-citation>
          <year>1997</year>
          ), pp.
          <fpage>67</fpage>
          -
          <lpage>82</lpage>
          . DOI:
          <volume>10</volume>
          .1109/4235.585893.
        </mixed-citation>
      </ref>
      <ref id="ref57">
        <mixed-citation>
          <string-name>
            <surname>Xin-She Yang</surname>
          </string-name>
          .
          <article-title>Nature-Inspired Optimization Algorithms</article-title>
          . Elsevier,
          <year>2014</year>
          . ISBN:
          <fpage>978</fpage>
          -0-
          <fpage>12</fpage>
          -416743-8.
        </mixed-citation>
      </ref>
      <ref id="ref58">
        <mixed-citation>
          <string-name>
            <surname>Xin-She Yang</surname>
          </string-name>
          . “
          <article-title>Nature-inspired optimization algorithms: Challenges and open problems”</article-title>
          .
          <source>In: Journal of Computational Science (Mar</source>
          .
          <year>2020</year>
          ), p.
          <fpage>101104</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref59">
        <mixed-citation>
          <source>DOI: 10</source>
          . 1016 / j . jocs .
          <year>2020</year>
          .
          <volume>101104</volume>
          . eprint:
          <year>2003</year>
          .03776v1.
        </mixed-citation>
      </ref>
      <ref id="ref60">
        <mixed-citation>
          <string-name>
            <surname>B. D.</surname>
          </string-name>
          Youn et al. “
          <article-title>Reliability-based design optimization for crashworthiness of vehicle side impact”</article-title>
          .
          <source>In: Structural and Multidisciplinary Optimization</source>
          <volume>26</volume>
          .
          <fpage>3</fpage>
          -
          <lpage>4</lpage>
          (
          <issue>Feb</issue>
          .
          <year>2004</year>
          ), pp.
          <fpage>272</fpage>
          -
          <lpage>283</lpage>
          . DOI: 10.
        </mixed-citation>
      </ref>
      <ref id="ref61">
        <mixed-citation>
          <volume>1007</volume>
          /s00158-003-0345-0.
        </mixed-citation>
      </ref>
      <ref id="ref62">
        <mixed-citation>
          <string-name>
            <given-names>Wei</given-names>
            <surname>Zeng</surname>
          </string-name>
          et al. “
          <article-title>Design Optimization of a VX Gasket Structure for a Subsea Connector Based on the Kriging Surrogate Model-NSGA-II Algorithm Considering the Load Randomness”</article-title>
          .
          <source>In: Algorithms</source>
          <volume>12</volume>
          .2 (
          <issue>Feb</issue>
          .
          <year>2019</year>
          ), p.
          <fpage>42</fpage>
          . DOI:
          <volume>10</volume>
          .3390 / a12020042.
        </mixed-citation>
      </ref>
      <ref id="ref63">
        <mixed-citation>
          <string-name>
            <given-names>Michael</given-names>
            <surname>Zhu</surname>
          </string-name>
          and
          <string-name>
            <given-names>Suyog</given-names>
            <surname>Gupta</surname>
          </string-name>
          . “To Prune, or Not to Prune:
          <article-title>Exploring the Efficacy of Pruning for Model Compression”</article-title>
          .
          <source>In: 6th International Conference on Learning Representations, ICLR</source>
          <year>2018</year>
          , Vancouver, BC, Canada, April 30 - May 3,
          <year>2018</year>
          , Workshop Track Proceedings.
          <year>2018</year>
          . arXiv: https://arxiv.
        </mixed-citation>
      </ref>
      <ref id="ref64">
        <mixed-citation>
          <source>org/abs/1710.01878v2 [cs.ML].</source>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>