<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Assessing the suitability of surrogate models in evolutionary optimization?</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Martin Holena</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Radim Demut</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Institute of Computer Science, Academy of Sciences of the Czech Republic Pod Vodarenskou vez 2</institution>
          ,
          <addr-line>18207 Prague</addr-line>
        </aff>
      </contrib-group>
      <fpage>31</fpage>
      <lpage>38</lpage>
      <abstract>
        <p>The paper deals with the application of evolu- eral days to several weeks of time and costs several to tionary algorithms to black-box optimization, frequently en- many thousands of euros. countered in biology, chemistry and engineering. In those The usual way of dealing with a time-consuming areas, however, the evaluation of the black-box tness is of- and costly evaluation of an objective function is to ten costly and time-consuming. Such a situation is usually evaluate such a function only sometimes, and tackled by evaluating the original tness only sometimes, evaluate its suitable response-surface model othearnwdiseev,alcuaallteidngs uitrsroagpaptreopmroiadteel roefstphoenset-nseusrsf.acSeevmeroadlelkiontdhs- erwise [19, 25]. In the context of evolutionary optiof models have been successful in surrogate modelling, and mization, such a model is commonly called surrogate a variety of models of each kind can be obtained through model of the tness, and the approach is called surparametrization. Therefore, real-world applications of sur- rogate modelling [10, 29, 32, 37] (occasionally also outrogate modelling entail the problem of assessing the suit- side that context [3, 23]). Because tness is typically ability of di erent models for the optimization task being assumed to be highly nonlinear, nonlinear models are solved. The present paper attempts to systematically inves- used as surrogate models. So far most frequently entigate this problem. It surveys available methods to assess countered have been Gaussian processes [7, 28, 37] model suitability and reports the incorporation of several (inspired by their success in response surface modsuch methods in our recently proposed approach to surro- elling [20, 21, 23, 35]), radial basis function (RBF) netgate modelling based on radial basis function networks. In works [2, 37] and other kinds of feedforward neural netiatdpdaityisonmtuochthaettceonmtimononallysoutsoeditgslolobcaallssuuiittaabbiilliittyyofofraamgoidveenl, works [16, 18]. input. Finally, it shows some results of testing several of Due to the applicability of di erent kinds of models the surveyed methods in two real-world applications. to surrogate modelling, as well as due to the possibility to construct a variety of models of each kind through an appropriate parametrization, a large number of var1 Introduction ious surrogate models can always be employed. Therefore, real-world applications of surrogate modelling enAn important application area of evolutionary opti- tail the problem of assessing the suitability of di erent mization algorithms [8, 31] is black-bock optimization, models for the optimization task being solved. Unfori.e., optimization of an objective function (in evolu- tunately, no systematic attention seems to have been tionary terms called tness) that cannot be described paid to that problem so far in the area of surrogate explicitly, but is known only from its evaluations in modelling, the research in this area being focused on a nite number of points in the input space. the integration of surrogate models with evolutionary Frequently, the tness is evaluated in some empirical optimization algorithms, their adaptation to the optiway, through measurements or testing. This is typical mization tasks, and on increasing the accuracy of the for applications in biology, chemistry, or materials sci- constructed models [11, 14, 17, 26]. The present paper ence [1]. In those domains, however, the fact that evo- is an attempt to change the situation. We survey availlutionary algorithms rely solely on tness evaluations able methods to assess model suitability, concentratcan be quite disadvantageous because the evaluation of ing in particular on local suitability of the model for empirical functions encountered there is usually time- a given input. Moreover, we give some results of testconsuming and costly. For example in the evolution- ing several of the surveyed methods on two real-world ary optimization of catalytic materials [1, 12], where applications of surrogate modelling. a tness describes the suitability of the material for In the following section, the principles of surrogate a particular chemical reaction, its evaluation in one modelling are recalled and their usefulness for evolugeneration of the evolutionary algorithm needs sev- tionary optimization is documented. The key section of the paper is Section 3, in which the most important ? The research reported in this paper has been sup- methods for assessing model suitability are explained ported by the Czech Science Foundation (GA CR) grant and results of their testing are presented. P202/11/1368.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Surrogate modelling in evolutionary
optimization
(iv) The EA is run with the original tness for
a prescribed number ge of generations with
populations Pgm+1; : : : ; Pgm+ge (frequently,
ge = 1).
(v) The set E is replaced by E [ Pgm+1 [ [</p>
      <p>Pgm+ge and the algorithm returns to (ii).</p>
      <p>In evolutionary optimization, surrogate modelling is
an approach in which the evaluation of the original
black-box tness is restricted to points considered to be
most important in the search for its global maximum, The fact that surrogate modeling is employed in
and its appropriate response-surface model is evalu- the context of costly or time-consuming objective
funcated otherwise. Important for searching the global ma- tions e ectively excludes the possibility to use those
ximum of a tness function are on the one hand the functions for tuning surrogate modeling methods, and
highest values found so far, on the other hand the di- for comparing di erent models and di erent ways of
versity of current population. Therefore, the selection their combining with evolutionary optimization. To
of points in which the original tness is evaluated is get around this di culty, arti cial benchmark
funcalways based on some combination of those two crite- tions can be used, computed analytically but expected
ria. to behave in evolutionary optimization similarly to the</p>
      <p>The di erent ways of interaction between the sys- original tness. As an example, Fig. 1 shows the
applitem input/output interface, the generated evolution- cation of surrogate modelling to a benchmark function
ary algorithm (EA) and the surrogate model can basi- proposed in [34] for the application area of
optimizacally be assigned to one of the following two strategies: tion of catalytic materials (cf. [1]). The benchmark
function was optimized using the system GENACAT
A. The individual-based strategy consists in choosing [13, 15], one of several evolutionary optimization
sysbetween the evaluation of the original tness and tems developed speci cally for that application area.
the evaluation of its surrogate model individual- The evolutionary algorithm employed by GENACAT
wise, for example, in the following steps: is a genetic algorithm (GA) taking into account the
(i) An initial set E of individuals is collected in composition and properties of catalytic materials. As
which the original tness was evaluated surrogate model, a RBF-network trained with data
(e.g., individuals forming several rst genera- from all previous generations was used, combined with
tions of the EA). the GA according to the individual-based strategy.
(ii) The model is trained using pairs f(x; (x)) : The results shown in Fig. 1 clearly document that
surx 2 E g. rogate modelling substantially accelerates the search
(iii) The EA is run with the tness replaced for the maximum of a tness function.
by the model for one generation with a
poppuloaptuiolantiQonosfizseizfeorq Pth,ewohpetrimeiPzatisiotnhoefde,siarnedd 3 Assessing the suitability of di erent
q is a prescribed ratio (e.g., q = 10 or q = 100). models
(iv) A subset P Q of size P is selected so as to
contain those individuals from Q that are most Instead of a single surrogate model, a whole set of
modimportant according to the considered criteria els F an be used. Then it is necessary to decide how
for the progress of optimization. suitable each of them is to be evaluated instead of the
(v) For x 2 P, the original tness is evaluated. original tness . Typically, the suitability of a model</p>
      <p>
        F 2 F is assumed to be indirectly proportional to
(vi)riTthhme sreettuErniss rteop(laiic)e.d by E [ P and the algo- saogmiveenersreoqru"e(nFce),odfedantaednootnuFsedanfodr cthalecucolantsetdruuctsiionng
B. The generation-based strategy consists in choosing of F . Consequently, the most suitable surrogate model
between both kinds of evaluation generation-wise, is the one ful lling
for example, in the following steps:
(i) An initial set E of individuals in which the F^ = arg min "(F ): (
        <xref ref-type="bibr" rid="ref1">1</xref>
        )
original tness was evaluated is collected like F 2F
in the individual-based strategy. There are various ways how "(F ) takes into account
(ii) The model is trained using pairs f(x; (x)) : the evaluation (x) by the original tness and the
evalx 2 E g. uation F (x) by the model for given inputs x, e.g.,
(iii) The EA is run with the tness replaced mean absolute error, mean squared error, root mean
by the model for a number gm of generations, square error, relative entropy, Kullback-Leibler
diverinteractively obtained from the user, with pop- gence, . . . . There are also two basic ways how to assure
ulations P1; : : : ; Pgm of size P . that the given sequence of data was not used for the
i=1
k s
= 1 X
k
      </p>
      <p>
        X (F (x)
jDij x2Di
construction of F : single split and cross-validation, the Observe that the most suitable model F^ in (
        <xref ref-type="bibr" rid="ref1">1</xref>
        )
delatter having the important advantage that all avail- pends on the considered set of surrogate models F ,
able data are used both for model construction and but does not depend on the inputs in which it has to
for the estimation of model error. be evaluated. Therefore, it can be called globally most
      </p>
      <p>Let us exemplify the calculation of "(F ) by recall- suitable with respect to the given sequence of data. Its
ing the de nition of the root mean squared error on obvious advantage is that it needs to be found only
an input data sequence D: once, and then it can be used for all evaluations, as
long as the set F does not change.</p>
      <p>
        RMSE(F ) = RMSED(F ) = (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) Global suitability of surrogate models based on
s 1 cross-validation was tested in more than a dozen
evo= X (F (x) (x))2; lutionary optimization tasks. Here, we show results of
jDj x2D testing it in a task where F was a set of multilayer
perawlhlerroeojtDmjdeaennotseqsuathreedcaerrdrionraloiftyFofbDas.eTdhuons taheko-vfoelrd- caerpchtritoencstu(rMesL.PTs)hewyitwhetrweorehsitdrdicetne dlayteorshaavned ndIi =ere1n4t
crossvalidation with folds D1; : : : ; Dk is: ibneprustonfehuirdodnesn,
nnoeu=ro3nsountHpu1tinnetuhreonsr,satnadndthenHn2umink the second layer ful lling the heuristic pyramidal
conRMSE(F ) = 1 X RMSEDi (F ) = (
        <xref ref-type="bibr" rid="ref3">3</xref>
        ) dition: the number of neurons in a subsequent layer
k must not exceed the number of neurons in a previous
layer. Consequently,
(x))2;
which yields 78 di erent MLP architectures. They were
tested as follows:
1. The employed GA was run for 6 generations using
      </p>
      <p>
        the original tness.
2. For each of the considered 78 architectures, one
surrogate model was trained using all the available
data from the 1st{6th generation.
3. The RMSE of each surrogate model on the data
from the 1st{6th generation was estimated using (i) Kendall's rank correlation coe cient between
leave-one-out cross-validation, according to (
        <xref ref-type="bibr" rid="ref3">3</xref>
        ). the ranks of models according to the suitability
4. The 7th generation G7 of the genetic algorithm based on RMSE and estimated using
leave-onewas produced. out cross-validation, and according to the RMSE
5. The models obtained in step 2 were used to predict on test data from the 7th generation,
the tness of x 2 G7. (ii) achieved signi cance level p of the test of rank
6. For x 2 G7, also the original tness was evaluated. independence based on the correlation coe cient
7. From the results of steps 5{6, the RMSE of each obtained in (i).
      </p>
      <p>
        surrogate model on the data from the 7th
generation was calculated according to (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ), with D = G7. The results were
8. For each surrogate model, the RMSE calculated in
step 7 was compared to the RMSE estimate from = 0:77 and p = 1:7 10 8; (
        <xref ref-type="bibr" rid="ref5">5</xref>
        )
step 2.
      </p>
      <p>
        namely for the 21 MLP architectures that, in addition
to (
        <xref ref-type="bibr" rid="ref4">4</xref>
        ), ful l 6 nH1; nH2 11. The visualized
results indicate that the rank of models according to the
RMSE-based suitability estimated by means of
leaveone-out cross-validation on the data from the 1st{6th
generation correlates with their rank according to the
RMSE on the data from the 7th generation. We also
quanti ed the extent of that correlation, using:
      </p>
      <p>Figure 2 visualizes the results of comparisons in
step 8 for a subset of the considered surrogate models,
which clearly con rm a strong correlation between the
rank of the model suitability and the rank of model
RMSE on test data.
Needless to say, the fact that a globally most suitable
model achieves the least value of an error " calculated
using a given sequence of data does not at all mean
that it yields the most suitable prediction for every x
in which tness can be evaluated. Therefore, we
introduce the model F^x locally most suitable for x as</p>
      <p>F^x = arg min (F; x):</p>
      <p>F 2F</p>
      <p>
        (
        <xref ref-type="bibr" rid="ref6">6</xref>
        )
      </p>
      <p>
        Like " in (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ), in (
        <xref ref-type="bibr" rid="ref6">6</xref>
        ) denotes some error.
Differently to ", however, is de ned on the cartesian
product F X , where X denotes the set of points in
which can be evaluated.
      </p>
      <p>Recall that a surrogate model is evaluated in points
x 2 X in which has not been evaluated yet. Hence,
the calculation of the error (F; x) must not depend on
the value of (x). Though in the context of surrogate
modelling, using such errors has not been reported yet,
a number of error measures exist that could be used
to this end, most importantly:
{ widths of con dence intervals [33];
{ transductive con dence [9, 27, 36];
{ estimation of prediction error relying on density</p>
      <p>estimation [4];
{ sensitivity analysis [5];
{ several heuristics based on the nearest neighbours</p>
      <p>of the point of evaluation [4, 30];
{ heuristic based on the variance of bagged
models [6];</p>
      <p>
        We are currently extending the surrogate model
presented in [2], which is based on RBF-networks, with
three of the above error measures:
(i) Width of prediction intervals for x 2 X , i.e., of
con dence intervals for F^x(x) in (
        <xref ref-type="bibr" rid="ref6">6</xref>
        ) based on the
linearization of the surrogate models and on the
assumption of independent normally distributed
residuals. First, each F 2 F is replaced with its
Taylor expansion of the 1st order, which is a
linear regression model F LIN with d + 1
parameters, where d is the dimensionality of points in X .
      </p>
      <p>Hence, F is replaced with</p>
      <p>F LIN = fF LIN : F 2 F g:
Then for each considered x 2 X , the element F^MLE</p>
      <p>
        x
of F LIN corresponding to the maximum-likelihood
estimate of the d + 1 parameters is found using
a given training sequence of input-output pairs
(x1; y1); : : : ; (xp; yp). That allows to calculate for
each F 2 F the error in (
        <xref ref-type="bibr" rid="ref6">6</xref>
        ) as
(i)(F; x) =
1 +
      </p>
      <p>F1</p>
      <p>p
X(yj
j=1
[1; p
p d</p>
      <p>F^MLE(xj ))2</p>
      <p>x
d]</p>
      <p>p
X(yj
j=1</p>
      <p>
        F (xj ))2 ; (
        <xref ref-type="bibr" rid="ref8">8</xref>
        )
where F1 [1; p d] denotes the 1 quantile of
the Fisher-Snedecor distribution with the degrees
of freedom 1 and p d.
(ii) Di erence between the predicted value and the
nearest-neighbours average is for k nearest neighbours
xn1 ; : : : ; xnk calculated according to
      </p>
      <p>Pk
j=1 ynj</p>
      <p>k
(ii)(F; x) =</p>
      <p>F (x) :</p>
      <p>
        (
        <xref ref-type="bibr" rid="ref9">9</xref>
        )
(iii) Bagged variance requires to have some set B of
basic models and to nd, in B, a given number m of
models bagged with respect to (x1; y1);: : : ;(xp; yp),
i.e., globally most suitable with respect to
bootstrap samples from (x1; y1); : : : ; (xp; yp). Recall
from (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) that to nd such a bagged model, an
error "(B) is needed, calculated using the respective
bootstrap sample. In our implementation, we
always use RMSE to this end. In its calculation
according to (
        <xref ref-type="bibr" rid="ref3">3</xref>
        ), hence, D1; : : : ; Dk are folds of the
bootstrap sample. Using the bagged models, the
      </p>
      <p>nal set of considered surrogate models is de ned
as</p>
      <p>F =
8
&lt;F : F =
:
m
X BjF &amp; B1F ; : : : ; BmF 2 B
j=1
and the bagged variance is calculated according to
(iii)(F; x) =</p>
      <p>m 0
1 X @BjF (x)
m
j=1
1 m</p>
      <p>
        X BjF (x)A :
m j=1
12
9
= ;
;
(
        <xref ref-type="bibr" rid="ref10">10</xref>
        )
(
        <xref ref-type="bibr" rid="ref11">11</xref>
        )
      </p>
      <p>
        Because the extension of the surrogate model
from [2] with those three error measures has been
implemented very recently, it is now in the course of
test(
        <xref ref-type="bibr" rid="ref7">7</xref>
        ) ing in a second evolutionary optimization task. Here,
a part of the results from the rst task will be shown,
concerning the optimization of catalytic materials for
high-temperature synthesis of HCN [24]. In that task,
F was a set of ve RBF networks, each with a di erent
number of hidden neurons in the range 1{5. Those ve
networks were tested in a similar way as was employed
in the above MLP-case. In particular:
1. The employed GA was again run for 6 generations
      </p>
      <p>
        using the original tness.
2. For each of the 5 possible numbers of hidden
neurons, one surrogate model was trained using all the
available data from the 1st{6th generation.
3. The RMSE of each surrogate model on the data
from the 1st{6th generation was estimated using
10-fold cross-validation, according to (
        <xref ref-type="bibr" rid="ref3">3</xref>
        ).
      </p>
      <p>4. Based on the result of step 3, the globally most</p>
      <p>suitable model was determined.
5. The 7th generation G7 of the genetic algorithm</p>
      <p>was produced.
6. The models obtained in step 2 were used to predict</p>
      <p>the tness of x 2 G7.
7. For F 2 F and x 2 G7, the errors (i)(F; x);
(ii)(F; x); (iii)(F; x) were calculated.
8. The surrogate models locally most suitable for
x 2 G7 according to (i); (ii) and (iii) were
determined.
9. For x 2 G7, the original tness was evaluated.
10. For x 2 G7, the absolute errors of the considered
surrogate models were calculated from the results
of steps 6 and 9.</p>
      <p>Figure 3 visualizes the locally most suitable models
determined in step 8 for a subset of 30 randomly
selected catalytic materials from the 7th generation. The
fact that for those catalysts the absolute errors of all</p>
      <p>ve RBF networks were calculated in step 10 allows
to juxtapose the choices of the locally most suitable
models and the globally most suitable model (model
with 2 hidden neurons) with the achieved absolute
errors. The juxtaposition shows that the model with the
lowest absolute error was nearly always assessed as the
locally most suitable by some of the three implemented
error measures. Unfortunately, no one of them could
be relied on in a majority of all cases.
4</p>
      <p>Conclusion
This paper is, to our knowledge, a rst attempt to
systematically investigate available methods for assessing
the suitability of surrogate models in evolutionary
optimization. In addition to the commonly used global
suitability of a model, it paid much attention also to
its local suitability for a given input. We have
incorporated three methods for assessing local suitability into
our recently proposed approach to surrogate modelling
based on RBF networks. The paper not only surveyed
available methods for assessing suitability, but also
described their testing and presented some of the testing
results.</p>
      <p>The presented results clearly con rm the
usefulness of methods for assessing the suitability of
surrogate models: there is a strong correlation between the
ranks of models according to the RMSE-based global
suitability estimated by means of leave-one-out
crossvalidation on data from the 1st{6th generation and
according to the RMSE on the data from the 7th
generation, and for nearly every unseen input, the model
with the lowest absolute error is indeed assessed as
the locally most suitable by some of the implemented
methods. Unfortunately, none of the methods is able
to correctly assess the locally most suitable model for
a majority of the 7th generation, which shows that
further research in this area is needed. We want to take
active part in such research, pursuing the following
three directions:
(i) Implement and test further methods for assessing
local suitability, listed above in Subsection 3.1. We
consider particularly interesting the transductive
con dence machine [9, 27, 36] because it is a novel
method and has solid theoretical fundamentals.
(ii) Investigate whether the success of some of the
tested methods for assessing local suitability
depends on the kind of dataset (in terms of the
number of nominal attributes, number of continuous
attributes, etc.) on which the surrogate models
have been trained, or on their descriptive
statistics. To this end, we want to make use of the
GAME system [22], which collects a large amount
of meta-data about the kind of the processed
dataset and about its descriptive statistics.
(iii) Modify and combine the tested methods, using
the results obtained in (ii), to increase their
success in assessing the local suitability of surrogate
models.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <given-names>M.</given-names>
            <surname>Baerns</surname>
          </string-name>
          and
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Holena: Combinatorial development of solid catalytic materials. Design of high-throughput experiments, data analysis, data mining</article-title>
          . Imperial College Press, London,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <given-names>L.</given-names>
            <surname>Bajer</surname>
          </string-name>
          and
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Holena: Surrogate model for continuous and discrete genetic optimization based on RBF networks</article-title>
          . In C. Fyfe,
          <string-name>
            <given-names>P.</given-names>
            <surname>Tino</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Garcia-Osorio</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>H.</given-names>
            <surname>Yin</surname>
          </string-name>
          , (eds),
          <source>Intelligent Data Engineering and Automated Learning. Lecture Notes in Computer Science 6283</source>
          , pp.
          <volume>251</volume>
          {
          <fpage>258</fpage>
          . Springer Verlag, Berlin,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <given-names>A.J.</given-names>
            <surname>Booker</surname>
          </string-name>
          , J. Dennis,
          <string-name>
            <given-names>P.D.</given-names>
            <surname>Frank</surname>
          </string-name>
          , D.B. Sera ni, Torczon V., and
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Trosset: A rigorous framework for optimization by surrogates</article-title>
          .
          <source>Structural and Multidisciplinary Optimization</source>
          ,
          <volume>17</volume>
          ,
          <year>1999</year>
          ,
          <volume>1</volume>
          {
          <fpage>13</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <given-names>Z.</given-names>
            <surname>Bosnic</surname>
          </string-name>
          and I. Kononenko:
          <article-title>Comparison of approaches for estimating reliability of individual regression predictions</article-title>
          .
          <source>Data &amp; Knowledge Engineering</source>
          ,
          <volume>67</volume>
          ,
          <year>2008</year>
          ,
          <volume>504</volume>
          {
          <fpage>516</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <given-names>Z.</given-names>
            <surname>Bosnic</surname>
          </string-name>
          and I. Kononenko:
          <article-title>Estimation of individual prediction reliability using the local sensitivity analysis</article-title>
          .
          <source>Applied Intelligence</source>
          ,
          <volume>29</volume>
          ,
          <year>2008</year>
          ,
          <volume>187</volume>
          {
          <fpage>203</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6. L. Breiman:
          <article-title>Bagging predictors</article-title>
          .
          <source>Machine Learning</source>
          ,
          <volume>24</volume>
          ,
          <year>1996</year>
          ,
          <volume>123</volume>
          {
          <fpage>140</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7. D. Buche,
          <string-name>
            <given-names>N.N.</given-names>
            <surname>Schraudolph</surname>
          </string-name>
          , and P. Koumoutsakos:
          <article-title>Accelerating evolutionary algorithms with gaussian process tness function models</article-title>
          .
          <source>IEEE Transactions on Systems, Man, and Cybernetics</source>
          , Part C:
          <article-title>Applications</article-title>
          and Reviews,
          <volume>35</volume>
          ,
          <year>2005</year>
          ,
          <volume>183</volume>
          {
          <fpage>194</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <given-names>C.A.C.</given-names>
            <surname>Coello</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.B.</given-names>
            <surname>Lamont</surname>
          </string-name>
          , and
          <string-name>
            <surname>D.A. van Veldhuizen</surname>
          </string-name>
          :
          <article-title>Evolutionary algorithms for solving multiobjective problems</article-title>
          ,
          <source>2nd Edition</source>
          . Springer Verlag, Berlin,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <given-names>A.</given-names>
            <surname>Gammerman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Vovk</surname>
          </string-name>
          , and
          <string-name>
            <given-names>V.</given-names>
            <surname>Vapnik</surname>
          </string-name>
          :
          <article-title>Learning by transduction</article-title>
          .
          <source>In Uncertainty in Arti cial Intelligence</source>
          , pp.
          <volume>148</volume>
          {
          <fpage>155</fpage>
          . Morgan Kaufmann Publishers, San Francisco,
          <year>1998</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <given-names>D.</given-names>
            <surname>Gorissen</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Couckuyt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Demeester</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Dhaene</surname>
          </string-name>
          , and
          <string-name>
            <given-names>K.</given-names>
            <surname>Crombecq</surname>
          </string-name>
          :
          <article-title>A surrogate modeling and adaptive sampling toolbox for computer based design</article-title>
          .
          <source>Journal of Machine Learning Research</source>
          ,
          <volume>11</volume>
          ,
          <year>2010</year>
          ,
          <year>2051</year>
          {
          <year>2055</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <given-names>D.</given-names>
            <surname>Gorissen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Dhaene</surname>
          </string-name>
          , and F. DeTurck: Evolution-
          <fpage>25</fpage>
          . R.H.
          <string-name>
            <surname>Myers</surname>
            ,
            <given-names>D.C.</given-names>
          </string-name>
          <string-name>
            <surname>Montgomery</surname>
            , and
            <given-names>C.M.</given-names>
          </string-name>
          <article-title>Andersonary model type selection for global surrogate modeling</article-title>
          .
          <source>Cook: Response surface methodology: proces and prodJournal of Machine Learning Research</source>
          ,
          <volume>10</volume>
          ,
          <year>2009</year>
          ,
          <year>2039</year>
          {
          <article-title>uct optimization using designed experiment</article-title>
          .
          <source>John Wi2078. ley and Sons</source>
          , Hoboken,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <given-names>M.</given-names>
            <surname>Holena</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Baerns:</surname>
          </string-name>
          Computer-aided strategies 26. Y.S. Ong,
          <string-name>
            <given-names>P.B.</given-names>
            <surname>Nair</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.J.</given-names>
            <surname>Keane</surname>
          </string-name>
          , and
          <string-name>
            <surname>K.W.</surname>
          </string-name>
          <article-title>Wong: for catalyst development</article-title>
          . In G. Ertl,
          <string-name>
            <surname>H.</surname>
          </string-name>
          <article-title>Knozinger, Surrogate-assisted evolutionary optimization frameF. Schuth, and</article-title>
          <string-name>
            <surname>J. Weitkamp</surname>
          </string-name>
          , (eds),
          <article-title>Handbook of Het- works for high- delity engineering design problems</article-title>
          .
          <source>erogeneous Catalysis</source>
          , pp.
          <volume>66</volume>
          {
          <fpage>81</fpage>
          . Wiley-VCH, Wein- In Y. Jin, (ed.),
          <source>Knowledge Incorporation in Evoluheim</source>
          ,
          <year>2008</year>
          . tionary Computation, pp.
          <volume>307</volume>
          {
          <fpage>331</fpage>
          . Springer Verlag,
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>M. Holena</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Cukic</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          <string-name>
            <surname>Rodemerck</surname>
            , and
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Linke</surname>
          </string-name>
          : Berlin,
          <year>2005</year>
          .
          <article-title>Optimization of catalysts using speci c, description 27</article-title>
          . H.
          <string-name>
            <surname>Papadopoulos</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Poredrou</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <string-name>
            <surname>Vovk</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <surname>A.</surname>
          </string-name>
          <article-title>Gambased genetic algorithms</article-title>
          .
          <source>Journal of Chemical</source>
          Infor- merman.
          <article-title>Inductive con dence machies for regression</article-title>
          .
          <source>mation and Modeling</source>
          ,
          <volume>48</volume>
          ,
          <year>2008</year>
          ,
          <volume>274</volume>
          {
          <fpage>282</fpage>
          .
          <source>In Proceedings of the 13th European Conference on</source>
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>M. Holena</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Linke</surname>
          </string-name>
          , and U. Rodemerck: Evolu- Machie Learning, pp.
          <volume>345</volume>
          {
          <issue>356</issue>
          ,
          <year>2002</year>
          .
          <article-title>tionary optimization of catalysts assisted by neural- 28. A. Ratle: Kriging as a surrogate tness landscape in network learning</article-title>
          .
          <source>In Simulated Evolution and Learn- evolutionary optimization. Arti cial Intelligence for ing. Lecture Notes in Computer Science 6457</source>
          , pp.
          <volume>220</volume>
          {
          <string-name>
            <surname>Engineering</surname>
            <given-names>Design</given-names>
          </string-name>
          ,
          <source>Analysis and Manufacturing</source>
          ,
          <volume>15</volume>
          ,
          <fpage>229</fpage>
          . Springer Verlag, Berlin,
          <year>2010</year>
          .
          <year>2001</year>
          ,
          <volume>37</volume>
          {
          <fpage>49</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>M. Holena</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Linke</surname>
            , and U. Rodemerck: Generator 29. T. Ray and
            <given-names>W.</given-names>
          </string-name>
          <string-name>
            <surname>Smith:</surname>
          </string-name>
          <article-title>A surrogate assisted parallel approach to evolutionary optimization of catalysts and multiobjective evolutionary algorithm for robust engiits integration with surrogate modeling</article-title>
          . Catalysis To- neering design.
          <source>Engineering Optimization</source>
          ,
          <volume>38</volume>
          ,
          <year>2006</year>
          , day,
          <volume>159</volume>
          ,
          <year>2011</year>
          ,
          <volume>84</volume>
          {
          <fpage>95</fpage>
          . 997{
          <fpage>1011</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>M. Holena</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Linke</surname>
            , and
            <given-names>N.</given-names>
          </string-name>
          <article-title>Steinfeldt: Boosted neural 30</article-title>
          . S. Schaal and
          <string-name>
            <surname>C.G.</surname>
          </string-name>
          <article-title>Atkeson: Assessing the quality of networks in evolutionary computation</article-title>
          .
          <source>In Neural In- learned local models</source>
          .
          <source>In Advances in Neural Informaformation Processing. Lecture Notes in Computer Sci- tion Processing Systems</source>
          <volume>6</volume>
          , pp.
          <volume>160</volume>
          {
          <fpage>167</fpage>
          . Morgan Kaufence 5864, pp.
          <volume>131</volume>
          {
          <fpage>140</fpage>
          . Springer Verlag, Berlin,
          <year>2009</year>
          . mann Publishers, San Mateo,
          <year>1994</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <given-names>Y.</given-names>
            <surname>Jin</surname>
          </string-name>
          :
          <article-title>A comprehensive survery of tness approxima- 31</article-title>
          . R. Schaefer:
          <article-title>Foundation of global genetic optimization. tion in evolutionary computation</article-title>
          .
          <source>Soft Computing</source>
          ,
          <volume>9</volume>
          , Springer Verlag, Berlin,
          <year>2007</year>
          .
          <year>2005</year>
          ,
          <volume>3</volume>
          {
          <fpage>12</fpage>
          . 32. H.
          <string-name>
            <surname>Ulmer</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Streichert</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Zell</surname>
          </string-name>
          : Model assisted
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <given-names>Y.</given-names>
            <surname>Jin</surname>
          </string-name>
          , M. Husken, M. Olhofer, and
          <string-name>
            <given-names>B.</given-names>
            <surname>Sendho</surname>
          </string-name>
          :
          <article-title>Neu- evolution strategies</article-title>
          . In Y. Jin, (ed.),
          <article-title>Knowledge Incorral networks for tness approximation in evolution-</article-title>
          poration
          <source>in Evolutionary Computation</source>
          , pp.
          <volume>333</volume>
          {
          <fpage>355</fpage>
          .
          <article-title>ary optimization</article-title>
          . In Y. Jin, (ed.), Knowledge Incor- Springer Verlag, Berlin,
          <year>2005</year>
          . poration in Evolutionary Computation, pp.
          <volume>281</volume>
          {
          <fpage>306</fpage>
          . 33. E. Uusipaikka:
          <article-title>Con dence intervals in generalized reSpringer Verlag</article-title>
          , Berlin,
          <year>2005</year>
          .
          <article-title>gression models</article-title>
          . CRC Press, Boca Raton,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <given-names>D.R.</given-names>
            <surname>Jones</surname>
          </string-name>
          :
          <article-title>A taxonomy of global optimization meth- 34.</article-title>
          <string-name>
            <given-names>S.</given-names>
            <surname>Valero</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Argente</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Botti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.M.</given-names>
            <surname>Serra</surname>
          </string-name>
          , P. Serna,
          <article-title>ods based on response surfaces</article-title>
          .
          <source>Journal of Global</source>
          Op- M.
          <article-title>Moliner, and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Corma</surname>
          </string-name>
          <article-title>: DoE framework for cattimization</article-title>
          ,
          <volume>21</volume>
          ,
          <year>2001</year>
          ,
          <volume>345</volume>
          {
          <fpage>383</fpage>
          .
          <article-title>alyst development based on soft computing techniques</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <given-names>D.R.</given-names>
            <surname>Jones</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Schonlau</surname>
          </string-name>
          , and
          <string-name>
            <given-names>W.J.</given-names>
            <surname>Welch</surname>
          </string-name>
          : E cient Computers and Chemical Engineering,
          <volume>33</volume>
          ,
          <year>2009</year>
          ,
          <volume>225</volume>
          {
          <article-title>global optimization of expensive black-box functions</article-title>
          .
          <source>238. Journal of Global Optimization</source>
          ,
          <volume>13</volume>
          ,
          <year>1998</year>
          ,
          <volume>455</volume>
          {
          <fpage>492</fpage>
          . 35.
          <string-name>
            <surname>J. Villemonteix</surname>
            ,
            <given-names>E.</given-names>
            Vazquez, and E.
          </string-name>
          <string-name>
            <surname>Walter</surname>
          </string-name>
          : An
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>J.P.C. Kleijnen</surname>
            , W. van Baers,
            <given-names>and I. van</given-names>
          </string-name>
          <article-title>Nieuwen- informational approach to the global optimization of huyse: Constrained optimization in expensive simu- expensive-to-evaluate functions</article-title>
          .
          <source>Journal of Global Oplation: Novel approach. European Journal of Opera- timization, 44</source>
          ,
          <year>2009</year>
          ,
          <volume>509</volume>
          {
          <fpage>534</fpage>
          . tional Research,
          <volume>202</volume>
          ,
          <year>2010</year>
          ,
          <volume>164</volume>
          {
          <fpage>174</fpage>
          . 36. V.
          <string-name>
            <surname>Vovk</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Gammerman</surname>
          </string-name>
          , and G. Shafer: Algorithmic
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22. P.
          <article-title>Kord k, J</article-title>
          . Koutn k, J.
          <string-name>
            <surname>Drchal</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          <article-title>Kovar k, learning in a random world</article-title>
          . Springer Verlag, Berlin,
          <string-name>
            <given-names>M.</given-names>
            <surname>Cepek</surname>
          </string-name>
          , and
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Snorek: Meta-learning approach 2005</article-title>
          .
          <article-title>to neural network optimization</article-title>
          .
          <source>Neural Networks</source>
          ,
          <volume>23</volume>
          , 37.
          <string-name>
            <given-names>Z.Z.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.S.</given-names>
            <surname>Ong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.B.</given-names>
            <surname>Nair</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.J.</given-names>
            <surname>Keane</surname>
          </string-name>
          , and
          <string-name>
            <surname>K.Y.</surname>
          </string-name>
          <year>2010</year>
          ,
          <volume>568</volume>
          {
          <fpage>582</fpage>
          . Lum:
          <article-title>Combining global and local surrogate models to</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <given-names>S.J.</given-names>
            <surname>Leary</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bhaskar</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.J.</given-names>
            <surname>Keane</surname>
          </string-name>
          :
          <article-title>A derivative accellerate evolutionary optimization</article-title>
          .
          <source>IEEE Transacbased surrogate model for approximating and optimiz- tions on Systems, Man and Cybernetics</source>
          . Part C:
          <article-title>Aping the output of an expensive computer simulation</article-title>
          .
          <source>plications and Reviews</source>
          ,
          <volume>37</volume>
          ,
          <year>2007</year>
          ,
          <volume>66</volume>
          {
          <fpage>76</fpage>
          .
          <source>Journal of Global Optimization</source>
          ,
          <volume>30</volume>
          ,
          <year>2004</year>
          ,
          <volume>39</volume>
          {
          <fpage>58</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          24. S. Mohmel,
          <string-name>
            <given-names>N.</given-names>
            <surname>Steinfeldt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Endgelschalt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Holena</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kolf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>U.</given-names>
            <surname>Dingerdissen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Wolf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Weber</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.</given-names>
            <surname>Bewersdorf</surname>
          </string-name>
          :
          <article-title>New catalytic materials for the high-temperature synthesis of hydrocyanic acid from methane and ammonia by high-throughput approach</article-title>
          .
          <source>Applied Catalysis A: General</source>
          ,
          <volume>334</volume>
          ,
          <year>2008</year>
          ,
          <volume>73</volume>
          {
          <fpage>83</fpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>