<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>September</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Evolutionary Counterfactual Visual Explanation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Jacqueline Höllig</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Stefen Thoma</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Cedric Kulbach</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>FZI Research Center for Information Technology</institution>
          ,
          <addr-line>Haid-und-Neu-Strasse 10-14, 76131 Karlsruhe</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2022</year>
      </pub-date>
      <volume>20</volume>
      <issue>2022</issue>
      <fpage>0000</fpage>
      <lpage>0003</lpage>
      <abstract>
        <p>The increasing success of deep learning models in recent years comes with the drawback of increasing model complexity. Due to the complexity, model insights are hard to obtain. However, understanding the underlying reasoning for a proposed decision becomes crucial in critical settings. Counterfactual explanations are among the most popular methods to interpret predictions of so-called black-box machine learning models. They provide a form of explanation intuitive to human thinking by building ”what-if” scenarios. Despite their popularity for interpreting tabular data, they find limited adaption in the visual domain. Current approaches to image counterfactuals rely heavily on access to model parameters, additional training data, or surrogate models. However, access to additional information might not always be feasible. We, therefore, propose an evolutionary-based method for counterfactual image generation with a custom mutation operator based on data augmentation to overcome these limitations. We show that generating image counterfactuals solemnly on an input instance and access to the prediction function is possible and performs on par with existing methods.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Interpretability</kwd>
        <kwd>Counterfactuals</kwd>
        <kwd>Evolutionary Computation</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Deep Learning models are at the forefront of artificial development as they allow complex
decision-making and can sometimes even discover complex patterns in data that other algorithms
or humans can hardly find. Due to their complexity, those models are “black-boxes” with no
human-understandable explanations for their predictions. With the adaption of such algorithms
to critical areas like medical diagnosis, autonomous driving, or airport security, a
humaninterpretable explanation becomes crucial to gain trust in these algorithms. However, most
machine learning systems lack ways to make decisions transparent to humans. Currently,
interest in model-agnostic techniques of explainable and interpretable machine learning is
growing [
        <xref ref-type="bibr" rid="ref1 ref2 ref3 ref4">1, 2, 3, 4</xref>
        ]. Most of those approaches determine how much each feature or which
feature combination contributes to a particular decision (e.g., [
        <xref ref-type="bibr" rid="ref2 ref5">2, 5</xref>
        ]). Nevertheless, those methods
fail to show how a diferent prediction could have been achieved. According to Miller [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], an
essential factor for human-understandable explanations, besides selectivity (i.e., only some
causes of the prediction are shown), sociability (i.e., interactiveness), and exclusion of probability,
is contrastiveness. Contrastive explanations should not explain why an event  happened but
rather why an event  happened instead.
      </p>
      <p>
        A specific class of algorithms that can provide contrastive explanations are counterfactuals.
Counterfactuals present a perturbation to the original input that leads to a change in the
prediction of an underlying machine learning model. The roots of counterfactuals lie in causal
reasoning and ofer answers to the question "What if?" and "Why?". They are already in daily
use in scientific and ordinary language. Therefore, they provide an intuitive concept for humans
to understand. Despite many eforts to apply counterfactuals to improve the interpretability of
machine learning models (e.g., [
        <xref ref-type="bibr" rid="ref10 ref11 ref12 ref4 ref7 ref8 ref9">7, 8, 9, 10, 11, 12, 4</xref>
        ]), most approaches are restricted to specific
input data types (e.g., [
        <xref ref-type="bibr" rid="ref13 ref7 ref8">7, 8, 13</xref>
        ]), or the underlying model concepts (e.g., [
        <xref ref-type="bibr" rid="ref7 ref8">7, 8</xref>
        ]). Most work
focuses on tabular data (e.g., [
        <xref ref-type="bibr" rid="ref13 ref14 ref7">7, 14, 13</xref>
        ]). The small amount of work on images uses additional
information like surrogate models [
        <xref ref-type="bibr" rid="ref12 ref15">15, 12</xref>
        ], access to training data [
        <xref ref-type="bibr" rid="ref12 ref15">15, 12</xref>
        ], or model parameters
[
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. However, in the real world, this additional information is seldom available. In particular, in
industrial, medical, or privacy-sensitive applications, the user is often not the model developer
and, thus, has no access to model parameters or the expertise to evaluate those. Furthermore,
training data is often not available due to privacy-related issues. Nevertheless, validating and
explaining decisions is crucial for the user to understand the model’s quality, trustworthiness,
and decisions.
      </p>
      <p>
        This work develops an approach to generating model-agnostic image counterfactuals in a
multi-class prediction problem. Our approach, based on NSGA-II [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ], takes on an input image
and the prediction function of some black-box classifier to be explained. To summarise the main
contributions of this work, we show that:
1. the counterfactual optimization problem is applicable on images.
2. data augmentation mutation enables a better search space coverage compared
to uniform mutation.
3. our approach achieves state of the art results on par with the approaches of
      </p>
      <p>
        Wachter et al. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] and Van Looveren &amp; Klaise [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ].
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        To obtain an in-depth understanding of black-box models and their predictions, the current
research focus shifts from classic explainable AI tools (e.g., LIME [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], GradCam [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], SHAP [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ],
or Saliency Maps [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ]) that visualize why a particular decision was taken, to counterfactuals.
Counterfactuals show why a diferent decision was taken via alternatives, thereby providing
contrastiveness.
      </p>
      <p>
        The first steps to adapt counterfactuals from their roots in causal reasoning to a tool for
understanding black-box models were taken by Wachter et al. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. They built on the fundamentals
of Pearl [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] to develop a basic stochastic counterfactual generation approach. They proposed
the following formulation:
′
      </p>
      <p>=  min max  ( (′) − ′)2 + (, ′)
(1)</p>
      <p>
        The first part pushes the models’ prediction  (′) on the counterfactual ′ to a new target
class ′ ̸=  other than the original class . In the second part, the distance measure 
keeps the counterfactual ′ close to the original instance ,  balances the contributions of the
competing terms. Extending their work towards more realistic and interpretable counterfactuals,
multiple authors provide mechanisms like feature extractors [
        <xref ref-type="bibr" rid="ref15 ref8">8, 15</xref>
        ], constraints [
        <xref ref-type="bibr" rid="ref10 ref14 ref20">14, 20, 10</xref>
        ],
or prototypes [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. Sharma et al. [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ] built the first framework for counterfactuals applicable
to various black-box algorithms and data types without the need for extensive additional
information. They were able to show that their approach works for multiple data types but was
unable to produce human-interpretable counterfactuals on MNIST. Dandl et al. [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] created a
general framework for tabular data by formulating a multi-objective problem for counterfactuals
solved with the genetic algorithm NSGA-II.
      </p>
      <p>
        While counterfactuals have already been widely explored for tabular data [
        <xref ref-type="bibr" rid="ref10 ref12 ref14 ref21 ref22 ref4 ref7">7, 14, 22, 10, 21,
12, 4</xref>
        ], less work can be found on images. Some of the model-agnostic approaches for table data
have been applied to images (e.g.,[
        <xref ref-type="bibr" rid="ref21 ref4">21, 4</xref>
        ]), resulting in more adversarial samples than
counterfactuals.1 Approaches to image-specific counterfactuals focus primarily on counterfactuals for
convolutional neural networks [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] and learning of surrogate models [
        <xref ref-type="bibr" rid="ref12 ref15">15, 12</xref>
        ].
      </p>
      <p>In contrast, our approach directly operates on the input image and the classifier prediction,
eliminating the need for parameter access and training surrogate models.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology and Model</title>
      <p>
        Throughout, we consider a black-box machine learning classifier  :  →  where  ∈  is a
set of input features ( = {1, 2, . . . , }) from the feature domain  and  ∈  is a vector
of probability distribution ( = {0, 1, . . . , ||} where ∑︀|=|0 = 1) over the number of classes
||. In this context, black-box denotes that only the model’s output  is observable. The model’s
inner workings are unknown. The goal of counterfactual approaches is, given an input  and a
classifier  , to provide an explanation via counter-examples allowing a human to understand
why classifier  chose class  for data point  and not a counterfactual class ′ [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>
        Adapted to the image domain, it results in: Given a query image  for which a classifier 
predicts the class , a counterfactual image ′ identifies how  could be changed in a proximate
(R1) [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], sparse (R2) [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] and plausible way (R3) [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] so that the classifier maximizes the
change in the predicted class (R4). Proximity refers to the distance between the original instance
 and the counterfactual instance ′, calculated as a distance. Sparsity is the number of feature
changes between  and ′. A plausible adaption indicates that the resulting ′ is in distribution
with the data.
      </p>
      <sec id="sec-3-1">
        <title>3.1. Objectives</title>
        <p>
          Following the definition of a counterfactual and the resulting requirements ( R1-R4), the
optimization problem minimizes the distance (R1) (, ′) between the original data point 
1Adversarial samples are closely related to counterfactuals. However, in contrast to counterfactuals that aim for
small perceptible changes to provide useful explanations, adversarial samples aim to make the changes as small
and imperceptible as possible to detect flaws in the model [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ].
and the newly generated counterfactual data point ′ to obtain a counterfactual that is close to
the original (1). Furthermore, to ensure sparse changes (R2) the optimization problem uses
the 0-norm to minimize the number of pixels subjected to change (2), referred to as sparsity.
The third optimization objective is the output distance (R4) which maximizes the classification
probability of the counterfactual into a target class . Equation (2) shows the optimization
problem to be minimized.
        </p>
        <p>min () := (1(, ′), 2(, ′), 3(′))
(2)
..</p>
        <p>() ̸=  (′)
1(, ′) = (, ′)</p>
        <p>2(, ′) = ∑︁ 1|− ′|</p>
        <p>
          =1
3(′) = 1 −  (′)
As distance measure , most approaches to counterfactuals adapt the 1- or 2-norm [
          <xref ref-type="bibr" rid="ref14 ref4">4, 14</xref>
          ].
However, on images, traditional distance functions do not suficiently account for image
similarity as it disregards the spatial relationships of images [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ]. Therefore we compare the mean
absolute error (using 1-norm) and the root mean squared error (using 2-norm) with diferent
image-based similarity indices see Section 4.1 and appendix A2. R3 is addressed during the
algorithm design in Section 3.2.
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Algorithm</title>
        <p>
          Our algorithm combines a modified version of NSGA-II with Island Populations and an adaption
of the auto-tuning approach of Castelli et al. [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ]. Deb et al. [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ] developed NSGA-II already
in 2002. However, it is still a heavily used algorithm for Multi-Objective Optimization today,
as other algorithms like indicator-based methods (e.g., SMS-EMA [
          <xref ref-type="bibr" rid="ref26">26</xref>
          ], IBEA [
          <xref ref-type="bibr" rid="ref27">27</xref>
          ]) rely on the
additional computation of the indicator, and the results of decomposition-based methods (e.g.,
MOEA/D [
          <xref ref-type="bibr" rid="ref28">28</xref>
          ], NSGA-III [
          <xref ref-type="bibr" rid="ref29">29</xref>
          ]) highly depend on the shape of the Pareto front [
          <xref ref-type="bibr" rid="ref30">30</xref>
          ].3
        </p>
        <p>As Equation 2 indicates, the only mandatory inputs for the algorithm are a black-box classifier
 and an input instance . Our algorithm generates an island  with a sub-population  for
each class  ∈  ∖{ ()} that a classifier  can classify. For each island  the algorithm stated
in Algorithm 1 runs in parallel, allowing the creation of counterfactuals in multiple boundry
directions at once. In every generation  each island  generates new candidates   by selecting,
crossing, and mutating high-performing individuals from the population .
2 https://github.com/JHoelli/Evolutionary_Counterfactual_Visual_Explanations/blob/master/Supplementary_
Material.pdf.
3For full reasoning we refer to the supplementary material A2.</p>
        <p>
          Algorithm 1 Algorithm on island 
1: Input: Population Size  , Generation , Original Image ,
2: Output: Non-Dominated Set 
The initial  individuals of an island  are randomly initialized with the length of the flattened
input image || along with an individual crossover rate  and mutation rate . The
generated individuals are evaluated on each objective stated in Equation (2). After evaluating
the individual’s fitness, non-dominated sorting is applied, and the crowding distance is
calculated according to NSGA-II [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]. The assigned ranks are used as the primary criterion in
the tournament selection. Thereby, two individuals are compared according to their rank. If
they have the same rank, the crowding distance is used as a secondary criterion to retain the
individual lying in the less crowded region to maintain the population’s diversity. The selected
individuals   are crossed by performing a uniform crossover [
          <xref ref-type="bibr" rid="ref31">31</xref>
          ]. The unified crossover
modiifes two individuals  [] ∈   and  [ − 1] ∈   in place by swapping attributes according to
the averaged crossover probability  of the individual. Based on the fitness of the resulting
ofsprings  [ − 1] and  [], a new crossover probability [ − 1] and []is assigned
to the corresponding ofspring. The selected individuals  [] are mutated with a mutation
probability [] by a random change of attribute. Based on the performance []
is adapted. The algorithm stops if it meets the desired number of generations or exceeds a
hypervolume [
          <xref ref-type="bibr" rid="ref32 ref33">32, 33</xref>
          ] threshold of  on all islands (i.e., on all islands, the generated solutions
dominate a portion of  of the objective space). The stopping criterion is applied to all islands
independently as the goal is to achieve a high-quality, non-dominated set for each of them.
        </p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Custom Operators</title>
        <p>
          Some of the operators used by default in evolutionary programming are unsuitable for the
stated problem, as they do not account for spatial dependencies in images or enable images to
be out of distribution. In this section, we depict the adapted operators of NSGA-II.
Initialization By default, NSGA-II initializes the parent population  randomly [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]. However,
initializing images with traditional stochastic techniques like Random Number Generators
leads to a vast search space (number candidate solutions for an image: (ℎ · ℎℎ ·
ℎ)!· 255!), which slows down convergence and the probability of finding a suitable
solution.
        </p>
        <p>To warmstart the algorithm by introducing relevant information and enable plausible
results (R3), we lean on the concepts of superpixels. The original image  of size  ×  ×
, where  is the height,  the width, and  the channels, is divided into  patches of
size  ×  ×  by slicing. Therefore an image  contains  patches  = [1, 2, ...,  ],
where a patch  is of size  ×  × . Each individual in a population is generated by
random shufling the patch positions .</p>
        <p>
          Mutation Traditionally, individuals are mutated to produce new ofsprings that are diferent
from their parents, thereby encouraging diversity. Using the crossover operator alone
leads to decreasing diversity and often results in local optima, as only the good parts of
the parents survive in each generation (premature convergence).[
          <xref ref-type="bibr" rid="ref34">34</xref>
          ]
The proposed mutation operator aims to prevent premature convergence and include
new relevant information in the population by applying data augmentation [35]. The
idea behind using data augmentation is to make sure that the changes are still
plausible (3) by manipulating the image with basic augmentation techniques. Only basic
techniques are used, as we do not use additional data or model parameters. The data
augmentation pipeline consists of functions for Random Flip (horizontally or vertically),
Random Rotation (by factor 0.2, resulting in a counterclockwise rotation by 1.25 ),
Random Contrast (by factors between 0.1 and 1.3, resulting in each pixel being adjusted by
  × ( −     ℎ) ), and Zoom (with height factors between
-0.7 and -0.2, resulting in a zoom-in between [20%, 70%]).
        </p>
        <p>
          Parameter Optimization According to Hassanat et al. [36], parameters of evolutionary
algorithms, especially the mutation and crossover rates, impact the obtainable results and
convergence speed. Tuning these parameters beforehand can result in several preliminary
experiments resulting in good values before the run. However, diferent values of
parameters might be optimal at diferent stages of the evolutionary process. Mutation can be
good in the initial generations to quickly explore the search space, while crossover is more
useful once the search process is close to the optimal solution. The proposed algorithm
implements a self-adaptive parameter control on the individual level, according to Castelli
et al. [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ]. Each individual [] in a population has its own crossover probability []
and its own mutation probability []. Both are initialized with random values
between 0 and 1. During crossover, two selected individuals,  [] and  [], generate an
ofspring with the probability  = 12 (( []) + ( [])), where the resulting
ofspring has the crossover probability ( []) =  + .  is a small positive
number if the fitness of the generated ofspring improved due to crossover and a small negative
number in any other case. During mutation, an individual mutates with its mutation
rate []. The resulting individual has a mutation rate of ( []) =  + ,
where  is a small positive number if the fitness of the generated ofspring improves due
to mutation and a small negative number in any other case.
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Evaluation</title>
      <p>In this section, we evaluate the performance of our counterfactual approach on the two broadly
research image datasets MNIST [37] and Fashion MNIST [38], to answer the following research
questions that aim to contribute to this work:</p>
      <p>Q1 How does the proposed image similarity measure influence the performance of
our algorithm? → Section 4.1
Q2 How does the proposed mutation mechanism influence the performance of our
algorithm? → Section 4.2
Q3 How does the image counterfactual approach perform compared to other
stateof-the-art methods for image counterfactuals? → Section 4.3</p>
      <p>Both datasets include 60.000 training images and 10.000 test images divided into 10 classes.
An image is of size 28 × 28 pixels. Both datasets were split into an 80/20 train/test split. The
train set was only used for training the classification model, while the following experiments
were run on the test set.</p>
      <p>The classification model consists of two convolutional layers for both datasets, followed by
max-pooling. The output layer is flattened and fed into a two-layer feed-forward network with
ReLu activation and a softmax output layer. This model is trained for 30 epochs with a batch
size of 100 on the training set. For MNIST, the model achieves a test set accuracy of 0.9921; for
Fashion MNIST, an accuracy of 0.831. We run all experiments on an Intel(R) Xeon(R) Platinum
8180M CPU with 2.50GHz with 1.5 TB of RAM. The code to our evaluation is made publicly
available on github4.</p>
      <sec id="sec-4-1">
        <title>4.1. Q1: Distance Function</title>
        <p>
          A counterfactual optimization problem usually includes minimizing the distance to the original
data. However, on images, traditional distance measures like the root mean squared error or
4https://github.com/JHoelli/Evolutionary_Counterfactual_Visual_Explanations
mean absolute error do not suficiently account for image similarity as they disregard images’
spatial relationships [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ]. To validate our choice of distance function, we compare the mean
absolute error ( ) to other popular image similarity indexes: Information Based Statistic
Similarity Measure ( ) [39], Feature-Based similarity Index (  ) [40], Root Mean
Squared Error ( ), and the Structural Similarity Index ( ) [41]. All functions were
inversed and mapped to the range [
          <xref ref-type="bibr" rid="ref1">0, 1</xref>
          ]. Appendix B2 defines the distance measures and
transformations.
        </p>
        <p>For each dataset, we randomly sample 15 instances. We run the algorithm without a target
direction  on every distance  ∈ {, ,  ,  ,  } for the selected
images and set the number of epochs to 100, as we do not want the stopping criteria to interfere.
The population size was set to 1000. The evaluation criterion is the hypervolume (i.e. the search
space coverage). The goal is to cover a high fraction of the search space in a small number of
generations.</p>
        <p>
          Figure 1 shows the development of the hypervolume averaged over all samples from both
datasets. Overall, ME has the highest search space coverage, indicating the highest likelihood
of achieving good results. After 100 epochs, the hypervolume of the algorithm optimizing ME
as distance reaches an average of 0.7023, the highest result for any tested distance. Further,
the superiority of   over   confirms Wachter et al. [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]. The sparsity introducing
property of the 1-norm used in ME as distance measure is desirable for human-understandable
counterfactuals, as only a small number of variables are changed. For image examples, we refer
the reader to section C in the appendix2.
        </p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Q2: Mutation Operator</title>
        <p>This section evaluates the mutation operator described in Section 3.3. As a baseline, we use an
implementation of random mutation, replacing a pixel with a random number  ∈ [0, 255] with
a probability of 0.1.</p>
        <p>For both mutation types, the algorithm runs on 15 randomly chosen images per dataset. With
ME as distance, we run the algorithm for 100 epochs with a population size of 1000 and no
target direction. Again we evaluate the hypervolume to evaluate which mutation leads better
through the search space.</p>
        <p>Figure 2 shows the distribution of the hypervolume for our mutation and the random mutation.
On average, our mutation leads to better search space coverage. It covers an, on average, over
10 % larger fraction of the search space than the random mutation baseline while having minor
performance fluctuations. For image examples, we refer the reader to section C in the appendix 2.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Q3: Benchmarking</title>
        <p>
          This section compares our approach to two widely used counterfactual benchmarks: the
approach of Wachter et al. [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] and Van Looveren &amp; Klaise [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ]. The approach of Wachter et al.[
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]
is a simple stochastic optimization between the distance of the original image and the
counterfactual image. Like our approach only the input image and the classification are necessary as
inputs. A more sophisticated approach regarding the data distribution was developed by Van
Looveren &amp; Klaise [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] by training a surrogate model for counterfactual search. Therefore, Van
Looveren &amp; Klaise [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] approach is a slightly harder benchmark for our algorithm to meet as
we do not use additional information regarding the data distribution.
        </p>
        <p>For both datasets, a representative of each class is chosen, resulting in 10 images per dataset.
Our algorithm runs on each image in every possible target direction  ̸∈ { ()} for 500 epochs
with a population size of 1000. We ran the benchmarks in two settings:
1. without a specific target class , to get the overall best counterfactual image.
2. with every possible target direction  ̸∈ { ()} to calculate the benchmark
metrics.</p>
        <p>
          The metrics were adapted and fitted to this context from [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ].
        </p>
        <p>• Distances: We measure the distance between a counterfactual ′ and the original image
 with the 0- and the 1- norm. The 0 norm calculates the number of pixels changed
between original and counterfactual instance and is identical to the sparsity from the
optimization problem (R2). The 1 norm calculates the average change and is consistent
with ME (R1).
• Redundancy: Redundancy measures the unnecessary proposed feature changes in a
counterfactual, by successively flipping one value of ′ after another back to  with the
goal of flipping the label back from  (′) to the original predicted outcome  (). If the
predicted outcome does not change, we increase the redundancy counter.
• yNN: yNN (Equation (3)) evaluates the data support (R3) of a counterfactual based on
instances from the trainings set. Ideally, a counterfactual should be close to a factual image
from the same target class . yNN is calculated by measuring how diferent neighborhood
points around the counterfactual ′ are classified. knn are the k-nearest neighbors of the
original image . We use a value of  = 5.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>
        In this work, we introduced an approach to generate image counterfactuals in a multiclass
classification problem by perturbing the original image with evolutionary computation and data
augmentation. Based on NSGA-II, we presented a promising direction in building counterfactuals
close to the original input with high data support, without the need to access additional
information or model parameters. Further, we show that the counterfactual optimization
problem is applicable in high-dimensional feature spaces such as images and that the mutation
and augmentation of the image data enables a better search space coverage. Finally, our approach
achieves state-of-the-art results on par with the approaches of Wachter et al. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] and Van Loveren
&amp; Klaise [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. Based on the provided approach and the general applicability, we aim to optimize
the runtime of the underlying algorithm further, investigate the mutation step and apply our
method to real-world applications.
      </p>
      <p>(a) MNIST
(b) Fashion MNIST</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>This work was carried out with the support of the German Federal Ministry of Education and
Research (BMBF) within the project "MetaLearn" (Grant 02P20A013).
[35] C. Shorten, T. M. Khoshgoftaar, A survey on Image Data Augmentation for Deep Learning,</p>
      <p>J. Big Data 6 (2019).
[36] A. Hassanat, K. Almohammadi, E. Alkafaween, E. Abunawas, A. Hammouri, V. B. Prasath,
Choosing mutation and crossover ratios for genetic algorithms-a review with a new
dynamic approach, Inf. 10 (2019).
[37] Y. LeCun, C. Cortes, C. Burges, MNIST handwritten digit database, ATT Labs [Online].</p>
      <p>Available http//yann.lecun.com/exdb/mnist 2 (2010).
[38] X. Han, K. Rasul, R. Vollgraf, Fashion-MNIST: a Novel Image Dataset for
Benchmarking Machine Learning Algorithms, arXiv Prepr. arXiv1708.07747 (2017).
arXiv:/arxiv.org/abs/1708.07747.
[39] M. A. Aljanabi, Z. M. Hussain, N. A. A. Shnain, S. F. Lu, Design of a hybrid measure
for image similarity: a statistical, algebraic, and information-theoretic approach, Eur. J.</p>
      <p>Remote Sens. 52 (2019) 2–15.
[40] Y. Zhang, D. W. Gong, Z. H. Ding, Handling multi-objective optimization problems
with a multi-swarm cooperative particle swarm optimizer, Expert Syst. Appl. 38 (2011)
13933–13941.
[41] U. Sara, M. Akter, M. S. Uddin, Image Quality Assessment through FSIM, SSIM, MSE and
PSNR—A Comparative Study, J. Comput. Commun. 07 (2019) 8–18.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>S. M.</given-names>
            <surname>Lundberg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. I.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <article-title>A unified approach to interpreting model predictions</article-title>
          ,
          <source>Adv. Neural Inf. Process. Syst</source>
          . 2017
          <string-name>
            <surname>-Decem</surname>
          </string-name>
          (
          <year>2017</year>
          )
          <fpage>4766</fpage>
          -
          <lpage>4775</lpage>
          . arXiv:
          <volume>1705</volume>
          .
          <fpage>07874</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>M. T.</given-names>
            <surname>Ribeiro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Guestrin</surname>
          </string-name>
          ,
          <article-title>"Why should i trust you?" Explaining the predictions of any classifier</article-title>
          ,
          <source>Proc. ACM SIGKDD Int. Conf. Knowl. Discov. Data Min</source>
          .
          <fpage>13</fpage>
          -17-Augu (
          <year>2016</year>
          )
          <fpage>1135</fpage>
          -
          <lpage>1144</lpage>
          . arXiv:
          <volume>1602</volume>
          .
          <fpage>04938</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>M. T.</given-names>
            <surname>Ribeiro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Guestrin</surname>
          </string-name>
          ,
          <article-title>Anchors: High-precision model-agnostic explanations</article-title>
          ,
          <source>in: Proceedings of the AAAI conference on artificial intelligence</source>
          , volume
          <volume>32</volume>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>S.</given-names>
            <surname>Wachter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Mittelstadt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Russell</surname>
          </string-name>
          ,
          <article-title>Counterfactual Explanations without Opening the Black Box: Automated Decisions and the GDPR, Harv</article-title>
          .
          <source>J. Law Technol</source>
          .
          <volume>31</volume>
          (
          <year>2017</year>
          )
          <fpage>1</fpage>
          -
          <lpage>52</lpage>
          . arXiv:
          <volume>1711</volume>
          .
          <fpage>00399</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>B.</given-names>
            <surname>Carter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Mueller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Jain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Giford</surname>
          </string-name>
          ,
          <article-title>What made you do this? Understanding black-box decisions with suficient input subsets</article-title>
          ,
          <source>22nd Int. Conf. Artif. Intell. Stat</source>
          . (
          <year>2018</year>
          )
          <fpage>567</fpage>
          --
          <lpage>576</lpage>
          . arXiv:
          <year>1810</year>
          .03805.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>T.</given-names>
            <surname>Miller</surname>
          </string-name>
          ,
          <article-title>Explanation in artificial intelligence: Insights from the social sciences</article-title>
          ,
          <source>Artif. Intell</source>
          .
          <volume>267</volume>
          (
          <year>2019</year>
          )
          <fpage>1</fpage>
          -
          <lpage>38</lpage>
          . arXiv:
          <volume>1706</volume>
          .
          <fpage>07269</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>S.</given-names>
            <surname>Dandl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Molnar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Binder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Bischl</surname>
          </string-name>
          ,
          <article-title>Multi-objective counterfactual explanations</article-title>
          ,
          <source>in: International Conference on Parallel Problem Solving from Nature</source>
          , Springer,
          <year>2020</year>
          , pp.
          <fpage>448</fpage>
          -
          <lpage>469</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Ernst</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Batra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Parikh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <article-title>Counterfactual visual explanations</article-title>
          ,
          <source>in: International Conference on Machine Learning, PMLR</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>2376</fpage>
          -
          <lpage>2384</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>T.</given-names>
            <surname>Laugel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. J.</given-names>
            <surname>Lesot</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Marsala</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Renard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Detyniecki</surname>
          </string-name>
          ,
          <article-title>The dangers of post-hoc interpretability: Unjustified counterfactual explanations</article-title>
          ,
          <source>IJCAI Int. Jt. Conf. Artif. Intell</source>
          . 2019
          <string-name>
            <surname>-Augus</surname>
          </string-name>
          (
          <year>2019</year>
          )
          <fpage>2801</fpage>
          -
          <lpage>2807</lpage>
          . arXiv:
          <year>1907</year>
          .09294.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>D.</given-names>
            <surname>Mahajan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Tan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <article-title>Preserving causal constraints in counterfactual explanations for machine learning classifiers</article-title>
          , arXiv preprint arXiv:
          <year>1912</year>
          .
          <volume>03277</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>M.</given-names>
            <surname>Pawelczyk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Bielawski</surname>
          </string-name>
          , J. v. d. Heuvel,
          <string-name>
            <given-names>T.</given-names>
            <surname>Richter</surname>
          </string-name>
          , G. Kasneci,
          <string-name>
            <surname>Carla:</surname>
          </string-name>
          <article-title>a python library to benchmark algorithmic recourse and counterfactual explanation algorithms</article-title>
          ,
          <source>arXiv preprint arXiv:2108.00783</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>A. V.</given-names>
            <surname>Looveren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Klaise</surname>
          </string-name>
          ,
          <article-title>Interpretable counterfactual explanations guided by prototypes</article-title>
          ,
          <source>in: Joint European Conference on Machine Learning and Knowledge Discovery in Databases</source>
          , Springer,
          <year>2021</year>
          , pp.
          <fpage>650</fpage>
          -
          <lpage>665</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>R. K.</given-names>
            <surname>Mothilal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <string-name>
            <surname>C. Tan,</surname>
          </string-name>
          <article-title>Explaining machine learning classifiers through diverse counterfactual explanations</article-title>
          ,
          <source>in: Proc. 2020 Conf. Fairness</source>
          , Accountability, Transpar.,
          <string-name>
            <surname>ACM</surname>
          </string-name>
          , New York, NY, USA,
          <year>2020</year>
          , pp.
          <fpage>607</fpage>
          -
          <lpage>617</lpage>
          . arXiv:
          <year>1905</year>
          .07697.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>A.</given-names>
            <surname>Dhurandhar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. Y.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Luss</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. C.</given-names>
            <surname>Tu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Ting</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Shanmugam</surname>
          </string-name>
          ,
          <string-name>
            <surname>P. Das</surname>
          </string-name>
          ,
          <article-title>Explanations based on the Missing: Towards contrastive explanations with pertinent negatives</article-title>
          ,
          <source>Adv. Neural Inf. Process. Syst</source>
          . 2018
          <string-name>
            <surname>-Decem</surname>
          </string-name>
          (
          <year>2018</year>
          )
          <fpage>592</fpage>
          -
          <lpage>603</lpage>
          . arXiv:
          <year>1802</year>
          .07623.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>S.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Kailkhura</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Loveland</surname>
          </string-name>
          , Y. Han,
          <article-title>Generative counterfactual introspection for explainable deep learning</article-title>
          ,
          <source>Glob. 2019 - 7th IEEE Glob. Conf. Signal Inf. Process. Proc. (</source>
          <year>2019</year>
          ). arXiv:
          <year>1907</year>
          .03077.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>K.</given-names>
            <surname>Deb</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Pratap</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Agarwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Meyarivan</surname>
          </string-name>
          ,
          <article-title>A fast and elitist multiobjective genetic algorithm: NSGA-II, IEEE Trans</article-title>
          .
          <source>Evol. Comput</source>
          .
          <volume>6</volume>
          (
          <year>2002</year>
          )
          <fpage>182</fpage>
          -
          <lpage>197</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>R. R.</given-names>
            <surname>Selvaraju</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Cogswell</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. Das</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Vedantam</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Parikh</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Batra</surname>
          </string-name>
          , Grad-CAM:
          <article-title>Visual Explanations from Deep Networks via Gradient-based Localization</article-title>
          ,
          <source>Int. J. Comput. Vis</source>
          .
          <volume>128</volume>
          (
          <year>2016</year>
          )
          <fpage>336</fpage>
          -
          <lpage>359</lpage>
          . arXiv:
          <volume>1610</volume>
          .
          <fpage>02391</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>A.</given-names>
            <surname>Atrey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Clary</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Jensen</surname>
          </string-name>
          , Exploratory Not Explanatory:
          <article-title>Counterfactual Analysis of Saliency Maps for Deep Reinforcement Learning, arXiv Prepr</article-title>
          .
          <year>arXiv1912</year>
          .
          <volume>05743</volume>
          (
          <year>2019</year>
          )
          <fpage>1</fpage>
          -
          <lpage>23</lpage>
          . arXiv:
          <year>1912</year>
          .05743.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>J.</given-names>
            <surname>Pearl</surname>
          </string-name>
          , Causality, Cambridge university press,
          <year>2000</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>A</surname>
          </string-name>
          .
          <string-name>
            <surname>-H. Karimi</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>Barthe</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Balle</surname>
            ,
            <given-names>I. Valera</given-names>
          </string-name>
          ,
          <article-title>Model-agnostic counterfactual explanations for consequential decisions</article-title>
          ,
          <source>in: International Conference on Artificial Intelligence and Statistics</source>
          , PMLR,
          <year>2020</year>
          , pp.
          <fpage>895</fpage>
          -
          <lpage>905</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>S.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Henderson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          , Certifai:
          <article-title>Counterfactual explanations for robustness, Transparency, Interpretability, and Fairness of Artificial Intelligence models (</article-title>
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>A.</given-names>
            <surname>Dhurandhar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Pedapati</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Balakrishnan</surname>
          </string-name>
          , P.-Y. Chen,
          <string-name>
            <given-names>K.</given-names>
            <surname>Shanmugam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Puri</surname>
          </string-name>
          ,
          <article-title>Model agnostic contrastive explanations for structured data</article-title>
          , arXiv preprint arXiv:
          <year>1906</year>
          .
          <volume>00117</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>M.</given-names>
            <surname>Pawelczyk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Broelemann</surname>
          </string-name>
          ,
          <string-name>
            <surname>G.</surname>
          </string-name>
          <article-title>Kasneci, Learning Model-Agnostic Counterfactual Explanations for Tabular Data</article-title>
          ,
          <source>in: Proc. Web Conf</source>
          .
          <year>2020</year>
          , c, ACM, New York, NY, USA,
          <year>2020</year>
          , pp.
          <fpage>3126</fpage>
          -
          <lpage>3132</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>Zhou</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bovik</surname>
          </string-name>
          ,
          <article-title>Mean squared error: Love it or leave it? A new look at Signal Fidelity Measures</article-title>
          ,
          <source>IEEE Signal Process. Mag</source>
          .
          <volume>26</volume>
          (
          <year>2009</year>
          )
          <fpage>98</fpage>
          -
          <lpage>117</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>M.</given-names>
            <surname>Castelli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Manzoni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Vanneschi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Silva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Popovič</surname>
          </string-name>
          ,
          <article-title>Self-tuning geometric semantic Genetic Programming, Genet</article-title>
          .
          <source>Program. Evolvable Mach</source>
          .
          <volume>17</volume>
          (
          <year>2016</year>
          )
          <fpage>55</fpage>
          -
          <lpage>74</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>M.</given-names>
            <surname>Emmerich</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Beume</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Naujoks</surname>
          </string-name>
          ,
          <article-title>An EMO algorithm using the hypervolume measure as selection criterion</article-title>
          ,
          <source>Lect. Notes Comput. Sci</source>
          .
          <volume>3410</volume>
          (
          <year>2005</year>
          )
          <fpage>62</fpage>
          -
          <lpage>76</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>E.</given-names>
            <surname>Zitzler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Künzli</surname>
          </string-name>
          ,
          <article-title>Indicator-based selection in multiobjective search</article-title>
          ,
          <source>Lect. Notes Comput. Sci. (including Subser. Lect. Notes Artif. Intell. Lect. Notes Bioinformatics</source>
          )
          <volume>3242</volume>
          (
          <year>2004</year>
          )
          <fpage>832</fpage>
          -
          <lpage>842</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>Q.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>MOEA/D: A multiobjective evolutionary algorithm based on decomposition</article-title>
          ,
          <source>IEEE Trans. Evol. Comput</source>
          .
          <volume>11</volume>
          (
          <year>2007</year>
          )
          <fpage>712</fpage>
          -
          <lpage>731</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>K.</given-names>
            <surname>Deb</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Jain</surname>
          </string-name>
          ,
          <article-title>An evolutionary many-objective optimization algorithm using referencepoint-based nondominated sorting approach, part i: solving problems with box constraints</article-title>
          ,
          <source>IEEE transactions on evolutionary computation 18</source>
          (
          <year>2013</year>
          )
          <fpage>577</fpage>
          -
          <lpage>601</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>H.</given-names>
            <surname>Ishibuchi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Setoguchi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Masuda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Nojima</surname>
          </string-name>
          ,
          <article-title>Performance of Decomposition-Based Many-Objective Algorithms Strongly Depends on Pareto Front Shapes</article-title>
          ,
          <source>IEEE Trans. Evol. Comput</source>
          .
          <volume>21</volume>
          (
          <year>2017</year>
          )
          <fpage>169</fpage>
          -
          <lpage>190</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>W. M.</given-names>
            <surname>Spears</surname>
          </string-name>
          , K. A. De Jong,
          <article-title>On the virtues of parameterized uniform crossover</article-title>
          ,
          <source>Technical Report May</source>
          , Naval Research Lab Washington DC,
          <year>1991</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [32]
          <string-name>
            <surname>C. M. Fonseca</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Paquete</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>López-Ibáñez</surname>
          </string-name>
          ,
          <article-title>An improved dimension-sweep algorithm for the hypervolume indicator</article-title>
          ,
          <source>in: 2006 IEEE Congr. Evol. Comput. CEC</source>
          <year>2006</year>
          , IEEE,
          <year>2006</year>
          , pp.
          <fpage>1157</fpage>
          -
          <lpage>1163</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          [33]
          <string-name>
            <given-names>E.</given-names>
            <surname>Zitzler</surname>
          </string-name>
          , L. Thiele,
          <article-title>Multiobjective evolutionary algorithms: A comparative case study and the strength Pareto approach</article-title>
          ,
          <source>IEEE Trans. Evol. Comput</source>
          .
          <volume>3</volume>
          (
          <year>1999</year>
          )
          <fpage>257</fpage>
          -
          <lpage>271</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          [34]
          <string-name>
            <given-names>K.</given-names>
            <surname>Deb</surname>
          </string-name>
          , Introduction to genetic algorithms,
          <source>Sadhana - Acad. Proc. Eng. Sci</source>
          .
          <volume>24</volume>
          (
          <year>1999</year>
          )
          <fpage>293</fpage>
          -
          <lpage>315</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>