<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Towards Fast Visual Explanations of Local Path Planning with LIME and GAN</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Amar Halilović</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Senka Krivić</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Faculty of Electrical Engineering, University of Sarajevo</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Institute of Artificial Intelligence, Ulm University</institution>
        </aff>
      </contrib-group>
      <abstract>
        <p>As robots become a more significant part of humans' daily lives, bridging the gap between robot actions and human understanding of what robots do and how they make their decisions becomes challenging. We present an approach to local navigation explanation based on Local Interpretable Model-agnostic Explanations (LIME), a popular approach from the Explainable Artificial Intelligence (XAI) community for explaining individual predictions of black-box models. We show how LIME can be applied to a robot's local path planner. Moreover, we show how the General Adversarial Network (GAN) can be trained and used for fast explanation generation. We also analyze the quality and runtime of GAN explanations and present a tool for visualizing these explanations online as the robot navigates.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Robotics</kwd>
        <kwd>Path Planning</kwd>
        <kwd>Explainable Artificial Intelligence</kwd>
        <kwd>Explainability</kwd>
        <kwd>Interpretability</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Robots in social environments raise the requirement for explainability of robot behavior [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. As
the tendency of robots’ presence in society grows, this requirement becomes more pronounced.
The introduction of the “Right to explanation” [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] in the European Union as a part of the General
Data Protection Regulation (GDPR) [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] underlines the human right to explanation in the face of
machines making decisions that afect humans. Current decision-making methods in robotics
largely lack explainability and thus limit the faster adoption of robots in important tasks. A
lack of explainability can also become a safety issue when robots behave unexpectedly, putting
humans in highly sensitive environments at risk.
      </p>
      <p>
        We address explainability in robotics by focusing on explainable robot navigation in social
environments: Imagine a robot navigating in a known environment with the possibility of
encountering humans and obstacles. Local path planners allow robots to follow a global path
plan while dynamically reacting to unexpected occurrences. Some of the robot’s decisions
may require abrupt stops or changes of direction and path deviations, thus surprising people
in the neighborhood or even scaring them. This can lead to trust loss, which needs to be
mitigated [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. One mitigation strategy is explanation. We want to mitigate trust loss by
enabling robots to explain their navigational choices. Using Local Interpretable Model-agnostic
Explanations (LIME) [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], an established method from Explainable Artificial Intelligence (XAI)
[
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], we demonstrate how a robot can generate visual explanations of its local decision-making
in path planning and obstacle avoidance. To approach explanation generation in real-time, we
train a Generative Adversarial Network (GAN) [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] model on a dataset produced by LIME. We
demonstrate how the trained GAN model generates visual explanations of local path plans.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Technical background</title>
      <sec id="sec-2-1">
        <title>2.1. Local Interpretable Model-agnostic Explanations (LIME)</title>
        <p>
          LIME [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] is a model-agnostic local XAI technique that explains predictions of a black-box model
by learning an interpretable model around the instance of interest. The instance of interest can
be anything that is an input to an AI model, be it text, numerical data, or images. We focus
on visual explanations and use an image (viz., the local costmap, see below) as the instance of
interest. LIME for images1 takes the input image and partitions it into segments – superpixels,
thereby creating interpretable features. Then, it perturbs interpretable features, turning them
of to generate perturbed samples (perturbations) in the neighborhood of the instance of interest.
For every perturbation, LIME queries the black-box model and thereby generates a local data
set of (perturbed) neighbor images and the respective black-box model’s predictions. On this
new dataset, LIME trains an interpretable model, viz., a weighted linear regression model.
The explanation is obtained by interpreting the coeficients of the trained linear model: The
importance of each segment in the image for the behavior of the black-box model is represented
by one coeficient in the linear model. Depending on the sign of the coeficient, the interpretable
feature (viz., the segment in the image) positively or negatively afects the black-box model’s
prediction. That said, applying LIME to explain local navigation visually, one needs to provide
a suitable method for computing a segmentation of a local costmap (viz., the interpretable
features). Moreover, the output of the local planner has to be interpreted as the prediction of
some black-box model.
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Generative Adversarial Networks (GANs)</title>
        <p>
          GANs were introduced by Ian Goodfellow et al. [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] as a deep learning framework for the
estimation of generative models. Estimation is done by an adversarial process where a generative
model, Generator (G), and a discriminative model, Discriminator (D), are trained concurrently.
G generates new samples by learning the training data distribution, while D estimates whether
the provided sample is from the training data or is produced by G. Mehdi and Osindero [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ]
introduced conditional GAN (cGAN), where G and D are conditioned on some information. Isola
and colleagues [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] show how cGAN can be used for image-to-image translation by conditioning
on images. G is trained to learn translation between input and output images and fool D, while
D learns to classify output images as real (coming from the training dataset) or fake (generated
by G). In our work, we employ their pix2pix cGAN architecture2 to achieve a fast explanation
generation of local navigation decisions. D is trained by minimizing the negative log-likelihood
of identifying real and fake images conditioned on input images, while G is trained using the
1https://github.com/marcotcr/lime
2shorturl.at/giFUX
adversarial loss of D (whether it fools the discriminator or not) and L1 loss (mean absolute
per pixel diference between real and fake images) which are combined into a composite loss
function. We condition G and D on local costmaps (see Fig. 1b,1f,1j) as inputs and explanation
images (LIME outputs) (see Fig. 1c,1g,1k) as outputs. Both input and output images include
(besides obstacle information) the robot’s location, the local plan, and the global plan.
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Experiment I: Explanations with LIME</title>
      <sec id="sec-3-1">
        <title>3.1. Technical Set-Up</title>
        <p>
          Our set-up is situated in the context of the ROS navigation stack [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ]. A global path planner
has generated a global path plan for the robot to navigate to a specified goal position. For path
following and obstacle avoidance, a local path planner takes the local costmap and the global
path as input and outputs a local path (in terms of a velocity vector) for the robot to execute.
For LIME to be applicable, the black-box behavior needs to be deterministic. Therefore, we
do not employ sampling-based planners, such as DWA or RRT, but instead, employ the TEB
planner [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ]. To use LIME for generating visual explanations of local path plans in terms of
obstacles in the local neighborhood of the robot, we use the local costmap as an instance of
interest and the TEB planner as the black box that takes that costmap as input and outputs some
local path. LIME first segments the local costmap into obstacles as interpretable features. The
SLIC [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] segmentation algorithm is used to get obstacle segments. As the second step, LIME
obtains perturbations of the segmented costmap by turning of segments by replacing them with
free space. The perturbed local costmap, together with the global plan, the robot’s footprint,
and its current velocities, form the input to the TEB, which then outputs a local plan for the
perturbation at hand. The deviation of the so-calculated local plan from the global plan is taken
as a target for the interpretable model and is calculated as a sum of the minimal point-to-point
L2 diferences between the local and the global plan.
        </p>
        <p>We get an explanation image for each local navigation decision by coloring segments based
on their LIME coeficients. The sign of the coeficient dictates the color: Positive-weighted
segments are colored green, and negative-weighted segments are colored red. A green-colored
segment contributed positively to the deviation; that is, green indicates “without that segment, the
local plan would deviate less from the global plan”. Conversely, a red-colored segment indicates
“without that segment, the local plan would deviate more from the global plan”. Color intensity is
set proportional to the coeficient with intensities in the range [0, 255] in RGB color space.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Qualitative results</title>
        <p>Figures 1a, 1e, and 1i show three characteristic local navigation cases (C1, C2, and C3) in our
lab. The robot (a TIAGo from PAL robotics) tries to follow the global plan that leads it through
the doorway. Figure 1b, 1f, and 1j show the local costmaps for three local navigation cases with
black obstacles, white robot’s location, and grey free space. In C1, the local plan (yellow dots)
mostly coincides with the global plan (blue dots), while in C2, the starting and ending points
of the local plan could not be connected into a joint trajectory. In C3, the local plan deviates
(a) C1: robot
(b) C1: costmap
(c) C1: LIME expl.</p>
        <p>(d) C1: GAN expl.
(e) C2: robot
(f) C2: costmap
(g) C2: LIME expl.</p>
        <p>(h) C2: GAN expl.
(i) C3: robot
(j) C3: costmap
(k) C3: LIME expl.</p>
        <p>(l) C3: GAN expl.
from the global plan. The LIME explanation explains how obstacles and/or parts of obstacles
contribute to the deviation. From the explanation images 1c, 1g, and 1k we have:
• C1: “The right (green) wall segment increases deviation, while the left (red) wall segment
decreases it, squeezing it to the doorway.”
• C2: “The (green) obstacle increases the deviation because if it were not there, the robot
would follow the global plan. If the wall (red) were not there, the local path planner could
create the connected local plan and deviate from the global plan.”
• C3: “Both obstacles increase the deviation, but the round one does so more significantly.</p>
        <p>If it were not there, the robot would follow the global plan. If the rectangular obstacle
were not there, the robot would still deviate, but less.”</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Quantitative Results</title>
        <p>We analyze explanation runtime. LIME’s runtime is generally high and increases linearly in the
number of perturbations as shown in Fig. 2a, where the runtimes of the most important parts of
the LIME are plotted. Planner total time takes the biggest part of the total explanation runtime
and includes the preparation of input data for the planner (TEB), the planner calculation time
of all the paths for each perturbation, and the collection of the planner’s outputs. The planner
calculation time takes the biggest part of the planner total time. Both runtimes increase relative
to the increase in the number of perturbations. As segmentation only needs to be done once for
each explanation, its runtime is unafected by the number of perturbations.</p>
        <p>(a) Runtime
(b) Visualization</p>
        <p>LIME has clear limitations in that this method alone cannot be used for real-time explanations.
Fast-changing and socially complex environments like streets or places with people might
require real-time explanations. Even when using a small number of perturbations (which afects
the explanation quality), not every TEB call (every 200ms) can be explained in real time.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experiment II: LIME Explanations with GAN</title>
      <p>The first experiment showed that LIME can be used to generate meaningful visual explanations
but that the generation procedure is too slow for online usage. To approach explanation
generation in real time, we utilize GAN as an explanation method. Our main idea is to use LIME
only ofline to generate a dataset of pairs of local costmaps and respective explanations. With
this dataset, we train GAN for image-to-image translation. This way, explanation generation
becomes independent of the number of samples.</p>
      <p>We have trained an image-to-image GAN for 200 epochs with 240 training image pairs, 60
validation image pairs, and 60 test image pairs. The image-pairs dataset was generated using
LIME with the configuration as outlined in the description of our first experiment; see Section
3.1. Of the important GAN settings, resnet_9blocks is used as Generator architecture. Other
settings are kept as in the pix2pix standard implementation. The trained GAN model generates
an explanation image by taking a local costmap with a plotted robot’s position and local and
global plans as input. In the following, we refer to the trained GAN model simply as GAN.</p>
      <sec id="sec-4-1">
        <title>4.1. Results</title>
        <p>
          We assess the quality of the visual explanations generated by the GAN by human visual
examination. This is a recommended practice [
          <xref ref-type="bibr" rid="ref13 ref14">13, 14</xref>
          ]. Figures 1d, 1h, and 1l show GAN
explanations for the use-cases C1, C2, and C3, respectively. One can see distortions in the GAN
explanations, which, however, do not do any harm to the conveyed meaning. In C1, one can
see that the colors of colored segments are not as sharp as in LIME, but the explanation does
not sufer qualitatively. GAN’s explanation for C2 has similar properties, with even somewhat
diferent coloring of less important segments on the right wall. Still, this does not hamper the
explanation very much, as the main contributors are still clearly distinguished. In the GAN
explanation for C3, the green color is somewhat duller and blurred compared to LIME, but the
contributions of the segments are still visible. We report a mean GAN calculation runtime of
0.25 seconds and a mean GAN model loading runtime of 0.36 seconds. Hence, once the GAN
model is loaded, it can output four explanations per second.
        </p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Demonstrator: Visual Explanations with RViz</title>
        <p>We demonstrate how GAN explanations can be visualized in Rviz in real-time in Fig. 2b. The
GAN output is published as PointCloud2 and overlayed over the map view in RViz as a local
explanation layer. The GAN model is loaded once at the beginning of navigation and called
periodically with every new local plan produced by TEB, allowing for the local explanation
layer refresh frequency of 4Hz. This tool thus enables humans to observe which parts of the
environment the robot considers important for its navigational decisions. We envision the tool
to be used for inspection and debugging, teaching path planning, and demonstrating the robot’s
internal reasoning processes to interested laymen.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Discussion</title>
        <p>GAN achieves huge runtime savings compared to LIME and approaches the upper real-time
performance limit of 200 ms. Most importantly, explanations generated by GAN do not depend
on any image segmentation preprocessing, and the performance-hungry process of replanning
the local path for every input image perturbation is no longer needed. This translates to the
possibility of achieving explanations in near real-time even when many obstacles are considered
potential explanations. This allows for explanation generation in highly dynamic environments.</p>
        <p>One drawback of the GAN model is some distortions in the visual explanation. However,
these are not too harmful as they are local and do not significantly afect the color and shade
of color. A limitation of our work is that we have not systematically analyzed how well the
GAN explanations generalize to very complex environments. The GAN explanation procedure
does not make assumptions about the robot platform and its kino-dynamic constraints. It also
does not assume a specific underlying local planner. It generates the explanation only based on
an image containing the local obstacles along with the local plan and the global plan. Thus, it
may turn out that the GAN has to be retrained for every robotic platform. Another limitation is
that our explanation approach relies on the underlying local planner to be deterministic. This
is necessary because the procedure must be certain that a variation in the local path is due to
the obstacles in the surroundings rather than random fluctuations. In the future, we will also
investigate how non-deterministic path planners could be explained.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>M.</given-names>
            <surname>Lomas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Chevalier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. V.</given-names>
            <surname>Cross</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. C.</given-names>
            <surname>Garrett</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Hoare</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kopack</surname>
          </string-name>
          ,
          <article-title>Explaining robot actions</article-title>
          ,
          <source>in: Proceedings of the seventh annual ACM/IEEE international conference on Human-Robot Interaction</source>
          ,
          <year>2012</year>
          , pp.
          <fpage>187</fpage>
          -
          <lpage>188</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>M. E.</given-names>
            <surname>Kaminski</surname>
          </string-name>
          , The right to explanation, explained, Berkeley Tech.
          <source>LJ 34</source>
          (
          <year>2019</year>
          )
          <fpage>189</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>P.</given-names>
            <surname>Voigt</surname>
          </string-name>
          ,
          <string-name>
            <surname>A.</surname>
          </string-name>
          <article-title>Von dem Bussche, The eu general data protection regulation (gdpr), A Practical Guide</article-title>
          , 1st Ed., Cham: Springer International Publishing
          <volume>10</volume>
          (
          <year>2017</year>
          )
          <fpage>10</fpage>
          -
          <lpage>5555</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>S.</given-names>
            <surname>Tolmeijer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Weiss</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hanheide</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Lindner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Powers</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Dixon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Tielman</surname>
          </string-name>
          ,
          <article-title>Taxonomy of trust-relevant failures and mitigation strategies</article-title>
          ,
          <source>in: Proceedings of HRI</source>
          <year>2020</year>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>M. T.</given-names>
            <surname>Ribeiro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Guestrin</surname>
          </string-name>
          , “
          <article-title>Why should I trust you?” explaining the predictions of any classifier</article-title>
          ,
          <source>in: Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining</source>
          ,
          <year>2016</year>
          , pp.
          <fpage>1135</fpage>
          -
          <lpage>1144</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>T.</given-names>
            <surname>Miller</surname>
          </string-name>
          ,
          <article-title>Explanation in artificial intelligence: Insights from the social sciences</article-title>
          ,
          <source>Artificial intelligence</source>
          (
          <year>2019</year>
          )
          <fpage>1</fpage>
          -
          <lpage>38</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>I.</given-names>
            <surname>Goodfellow</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Pouget-Abadie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mirza</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Warde-Farley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ozair</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Courville</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bengio</surname>
          </string-name>
          , Generative adversarial nets,
          <source>Advances in neural information processing systems</source>
          <volume>27</volume>
          (
          <year>2014</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>M.</given-names>
            <surname>Mirza</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Osindero</surname>
          </string-name>
          ,
          <article-title>Conditional generative adversarial nets</article-title>
          ,
          <source>arXiv preprint arXiv:1411.1784</source>
          (
          <year>2014</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>P.</given-names>
            <surname>Isola</surname>
          </string-name>
          , J.-Y. Zhu,
          <string-name>
            <given-names>T.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. A.</given-names>
            <surname>Efros</surname>
          </string-name>
          ,
          <article-title>Image-to-image translation with conditional adversarial networks</article-title>
          ,
          <source>in: IEEE Conference on Computer Vision and Pattern Recognition</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>R. L.</given-names>
            <surname>Guimarães</surname>
          </string-name>
          , A. S. de Oliveira,
          <string-name>
            <given-names>J. A.</given-names>
            <surname>Fabro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Becker</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. A.</given-names>
            <surname>Brenner</surname>
          </string-name>
          ,
          <article-title>Ros navigation: Concepts and tutorial, Robot Operating System (ROS) The Complete Reference</article-title>
          (Volume
          <volume>1</volume>
          ) (
          <year>2016</year>
          )
          <fpage>121</fpage>
          -
          <lpage>160</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>C.</given-names>
            <surname>Rösmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Feiten</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Wösch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Hofmann</surname>
          </string-name>
          , T. Bertram,
          <article-title>Trajectory modification considering dynamic constraints of autonomous robots</article-title>
          ,
          <source>in: ROBOTIK 2012; 7th German Conference on Robotics, VDE</source>
          ,
          <year>2012</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>6</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>R.</given-names>
            <surname>Achanta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Shaji</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Smith</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Lucchi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Fua</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Süsstrunk</surname>
          </string-name>
          ,
          <article-title>Slic superpixels compared to state-of-the-art superpixel methods</article-title>
          ,
          <source>IEEE transactions on pattern analysis and machine intelligence</source>
          <volume>34</volume>
          (
          <year>2012</year>
          )
          <fpage>2274</fpage>
          -
          <lpage>2282</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>T.</given-names>
            <surname>Salimans</surname>
          </string-name>
          , I. Goodfellow,
          <string-name>
            <given-names>W.</given-names>
            <surname>Zaremba</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Cheung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Radford</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <article-title>Improved techniques for training gans</article-title>
          ,
          <source>Advances in neural information processing systems</source>
          <volume>29</volume>
          (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>A.</given-names>
            <surname>Borji</surname>
          </string-name>
          ,
          <article-title>Pros and cons of gan evaluation measures</article-title>
          ,
          <source>Computer Vision and Image Understanding</source>
          <volume>179</volume>
          (
          <year>2019</year>
          )
          <fpage>41</fpage>
          -
          <lpage>65</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>