<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Neural Criticality: Validation of Convolutional Neural Networks</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>V a´clav Divisˇ</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Marek Hru´ z</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Supported by ARRK Engineering and the University of West Bohemia. Faculty of Applied Sciences, University of West Bohemia Pilsen, CZE</institution>
          ,
          <addr-line>306 14</addr-line>
        </aff>
      </contrib-group>
      <fpage>2</fpage>
      <lpage>10</lpage>
      <abstract>
        <p>The black-box behavior of Convolutional Neural Networks is one of the biggest obstacles to the development of a standardized validation process. Methods for analyzing and validating neural networks currently rely on approaches and metrics provided by the scientific community without considering functional safety requirements. However, automotive norms, such as ISO26262 and ISO/PAS21448, do require a comprehensive knowledge of the system and of the working environment in which the network will be deployed. In order to gain such a knowledge and mitigate the natural uncertainty of probabilistic models, we focused on investigating the influence of filter weights on the classification confidence in Single Point Of Failure fashion. We laid the theoretical foundation of a method called the Neurons' Criticality Analysis. This method, as described in this article, helps evaluate the criticality of the tested network and choose related plausibility mechanism.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        The need to understand and rely on the inference processes
of Convolution Neural Networks (CNNs) grows in
importance since probabilistic models are being integrated in
autonomous vehicles
        <xref ref-type="bibr" rid="ref47">(Tesla 2019)</xref>
        ,
        <xref ref-type="bibr" rid="ref31">(Mobileye 2020)</xref>
        , where the
SW development follows functional safety standards and on
which lives may depend.
      </p>
      <p>
        The transparency, evaluation criteria and types of
explanations of the achieved results face low interpretability
        <xref ref-type="bibr" rid="ref2 ref47">(Belle
and Papantonis 2020)</xref>
        due to the increasing complexity of
the models used.
      </p>
      <p>
        Furthermore, as recently shown by various adversary
attack examples
        <xref ref-type="bibr" rid="ref30">(Liu et al. 2016)</xref>
        ,
        <xref ref-type="bibr" rid="ref5">(Brown et al. 2017)</xref>
        ,
        <xref ref-type="bibr" rid="ref12">(Eykholt
et al. 2018)</xref>
        , even a small perturbation in the input
image can cause a major change to the algorithm’s decision.
In addition, the current leading norms (ISO26262:2018,
ISO/PAS21448:2019) do not define validation process nor
metrics related to the probabilistic model
        <xref ref-type="bibr" rid="ref3">(BMW 2019)</xref>
        .
Finally, to our knowledge. there isn’t any method similar to
the Software Criticality Analysis (SWCA) or MC/DC
module test in the CNN field
        <xref ref-type="bibr" rid="ref11 ref34">(Salay and Czarnecki 2018)</xref>
        which
can analyze potential Single Point Of Failure (SPOF). All of
Copyright © 2021, for this paper by its authors. Use permitted
under Creative Commons License Attribution 4.0 International (CC
BY 4.0).
this motivated us to define a methodology and
understandable metrics according to which State-Of-The-Art (SOTA)
CNNs can be analyzed and the achieved results validated.
      </p>
      <p>In this work we introduced a new metric called criticality,
which can be either assigned to neurons (in CNN’s
terminology, those are referred to as ”filters’ weight”) or to layers.
We verified the importance and plausibility of the proposed
criticality as a possible safety metric by conducting two
experiments. We designed and implemented a method called
Neurons’ Criticality Analysis (NCA) and tested it on four
image classification models.</p>
      <p>The results of our experiment are extensively discussed
at the end of this work, where we also summarized the pros
and cons of the presented metric and methodology and
highlighted the use-cases we intend to investigate in the future.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Previous work</title>
      <p>
        As mentioned in
        <xref ref-type="bibr" rid="ref2 ref47">(Belle and Papantonis 2020)</xref>
        , data-driven
techniques struggle to be robust against domain shift, data
corruption and input space perturbation. The robustness can
be influenced by three aspects: architecture, training dataset
and optimization algorithm. It is not among the goals of this
paper to give a comprehensive overview of all three
topics, but to highlight the most influential milestones which
inspired our experiment.
      </p>
      <sec id="sec-2-1">
        <title>2.1 Classification via CNN</title>
        <p>
          As a first step, we looked at the straightforward VGG16
architecture
          <xref ref-type="bibr" rid="ref41 ref48">(Simonyan and Zisserman 2014)</xref>
          , where
BatchNorm layer, 3 3 CONV with stride and padding equal to
1 and 2 2 MAX POOL with stride 2 are several times
repeated. As a second step, we reviewed the ResNet
          <xref ref-type="bibr" rid="ref18 ref19">(He
et al. 2016a)</xref>
          architecture, which differentiates from the
previous models mainly by having residual blocks
          <xref ref-type="bibr" rid="ref17">(He et al.
2015)</xref>
          and using 1x1 convolutional operation. We tested
ResNet50V2
          <xref ref-type="bibr" rid="ref18 ref19">(He et al. 2016b)</xref>
          , which is the successor of the
original ResNet with enhanced residual connections for a
smoother information propagation. The aim of residual
connections is to create additional information flow and to learn
additive residual function. As explained in
          <xref ref-type="bibr" rid="ref29 ref41 ref48">(Lin, Chen, and
Yan 2014)</xref>
          and further investigated in
          <xref ref-type="bibr" rid="ref44">(Szegedy et al. 2014)</xref>
          ,
the concept known as ”Network In Network” (NIN) allows
reducing the filter’s dimensionality and increasing the
models’ non-linearity. With the combination of computationally
more expensive 3 3 and 5 5 convolution and parallel
branches the extraction of features from different scales can
be achieved simultaneously. This kind of blocks are often
called ”projection layer”
          <xref ref-type="bibr" rid="ref28">(Li et al. 2018)</xref>
          .
        </p>
        <p>
          Since building a model in ResNet fashion comes at
computational cost, a family of mobileNets emerged. The
standard convolution operation which was used up to this point
was, in case of MobileNet v1
          <xref ref-type="bibr" rid="ref22">(Howard et al. 2017)</xref>
          , replaced
by depthwise separable convolution. It was shown that
standard convolution can be split into depthwise and pointwise
convolutions, which decreases the number of operations by
a square of the spatial dimension of the used filter kernel.
Further improvements were carried out by Sandle et al. in
MobileNet v2
          <xref ref-type="bibr" rid="ref35">(Sandler et al. 2018)</xref>
          , where residual
connections between the cells and the expand/projection layers
were added.
        </p>
        <p>
          A research done by Google discovered a new method of
scaling the CNN. The goal of this optimization search was
to find scaling coefficients (network width, depth and
resolution) with respect to the accuracy and amount of operations.
EfficientNet
          <xref ref-type="bibr" rid="ref46 ref47">(Tan and Le 2019)</xref>
          was designed to demonstrate
the effectiveness of this scaling method and achieved SOTA
accuracy on the ImageNet dataset in 2019.
        </p>
        <p>
          The need to understand how the decision process is made
lead us to several papers
          <xref ref-type="bibr" rid="ref40">(Simonyan, Vedaldi, and Zisserman
2013)</xref>
          ,
          <xref ref-type="bibr" rid="ref41 ref48">(Zeiler and Fergus 2014)</xref>
          ,
          <xref ref-type="bibr" rid="ref1">(Bach et al.
2015)</xref>
          ,
          <xref ref-type="bibr" rid="ref26 ref38">(Shrikumar, Greenside, and Kundaje 2017)</xref>
          ,
          <xref ref-type="bibr" rid="ref26 ref43">(Sundararajan, Taly, and
Yan 2017)</xref>
          , where different visualization methods are
described. A comprehensive overview of the methods and their
goals can be found in
          <xref ref-type="bibr" rid="ref33">(Rafegas et al. 2019)</xref>
          , where Rafegas
et al. also presented a novel method of quantifying the
neurons’ selectivity to color and class.
2.2
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>Safety related issues</title>
        <p>
          The aleatoric and epistemic uncertainty
          <xref ref-type="bibr" rid="ref26">(Kendall and Gal
2017)</xref>
          of probabilistic models are currently under in-depth
study. The source of the aleatoric uncertainty is brought by
the randomness contained within the training-set, whereas
the epistemic uncertainty is caused by the lack of the
model’s knowledge. Bayesian machine learning
          <xref ref-type="bibr" rid="ref32">(Neal 2012)</xref>
          approach allows propagating the intermediate covariances
to the final layer and quantifying the hypothesis
uncertainty
          <xref ref-type="bibr" rid="ref16">(Graves 2011)</xref>
          ,
          <xref ref-type="bibr" rid="ref37 ref46 ref47">(Shridhar, Laumann, and Liwicki
2019)</xref>
          . Such an approach requires time-consuming training,
and, for the moment, models do not achieve the expected
accuracy. An additional one-shot approach uses Monte Carlo
dropout during inference in order to sample a subset of
networks, build statistics and calculate the thereof resulting
uncertainty
          <xref ref-type="bibr" rid="ref13">(Gal and Ghahramani 2016)</xref>
          . This improves the
demands on the inference time to reasonable limits and can
therefore be applicable in automotive.
        </p>
        <p>
          Many papers additionally address the problem of
datadriven ML algorithms and how to incorporate safety
mechanism in order to monitor the prediction
uncertainty
          <xref ref-type="bibr" rid="ref7">(Cheng 2020)</xref>
          ,
          <xref ref-type="bibr" rid="ref26 ref27">(Lakshminarayanan, Pritzel, and
Blundell 2017)</xref>
          ,
          <xref ref-type="bibr" rid="ref11 ref34">(DeVries and Taylor 2018)</xref>
          . Several
uncertainty estimation methods were lately evaluated by Henne
et al.
          <xref ref-type="bibr" rid="ref20">(Henne et al. 2020)</xref>
          with respect to functional safety.
        </p>
        <p>
          Our approach is driven by ISO/PAS21448
SOTIF (
          <xref ref-type="bibr" rid="ref25">ISO2019 2019</xref>
          ), ISO2626
          <xref ref-type="bibr" rid="ref24">(ISO 2011)</xref>
          norms, which
lack validation-unambiguity, and by the conclusion that
only 40% of the current automotive verification/validation
methods can be transferred to ML application
          <xref ref-type="bibr" rid="ref11 ref34">(Salay and
Czarnecki 2018)</xref>
          . We mostly considered the SOTIF norm,
which is an extension of the well-known ISO2626 norm
and provides a guidance (recommended activities) on
applicable design, verification and validation measures in
order for the product to be norm-compliant. The goal of the
recommended activities is to maximize the area of known
safe scenarios and minimize the unknown or unsafe areas
by applying technical measures.
2.3
        </p>
      </sec>
      <sec id="sec-2-3">
        <title>Adversary attacks</title>
        <p>The original idea of adversary attack is to introduce a small
perturbation to an input image so that the original class
doesn’t have the highest confidence and the adversary noise
stays unrecognized to the human perception system.</p>
        <p>
          One of the first attacks used the Fast Sign Gradient
Descent (FSGD) method
          <xref ref-type="bibr" rid="ref15">(Goodfellow, Shlens, and Szegedy
2015)</xref>
          , which calculates the gradient of a model’s loss
function with respect to the input image and ground-truth label
and either adds or subtracts a small portion of it,
depending whether the gradient was positive or negative.
Additional papers proposed a general and large perturbation
attack algorithm of physical objects considering spatial
constraints and physical limits on imperceptibility,
          <xref ref-type="bibr" rid="ref5">(Brown et al.
2017)</xref>
          ,
          <xref ref-type="bibr" rid="ref12">(Eykholt et al. 2018)</xref>
          . Ian Goodfellow summarized
additional weaknesses of the classification task in
          <xref ref-type="bibr" rid="ref14">(Goodfellow 2018)</xref>
          .
        </p>
        <p>
          The defense mechanism started to be deeply investigated
in the work of Lie et al.
          <xref ref-type="bibr" rid="ref30">(Liu et al. 2016)</xref>
          , which shows a
comprehensive experiment of different ResNets
Architectures trying to resist non-target adversarial images and states
that ResNet-152 has a 0% resistivity. The explanation to that
phenomenon is still in the open research area
          <xref ref-type="bibr" rid="ref5">(Brown et al.
2017)</xref>
          , but one of the latest works
          <xref ref-type="bibr" rid="ref42">(Song et al. 2018)</xref>
          shows
promising results and mentions defending methods, showing
that the robustness against different attacks can improve.
3
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Analysis methodology</title>
      <p>In this section we defined the metric and methodology
related to manipulating and analyzing the CNN’s decision
process. We took the inspiration for the Neurons’ Criticality
Analysis (NCA) method from the Software Criticality
Analysis (SWCA). The SWCA is a method which divides
modules of any action chain between critical (the SPOF of which
could have fatal consequences) and non-critical. In case of
an automotive SW component, the SWCA is carried out by
analyzing the signal flow from the actuator to the sensor,
while heuristically justifying the signal’s non-criticality. In
order to investigate if the decision of any CNN can be
significantly influenced by the SPOF of the filter connections, the
idea of an approach similar to SWCA was explored. From
now on we will refer to filter connection as a neuron, since
the principle can be generally applied to FC, CONV and
Depth-wise CONV layer,
Firstly, we denote the analyzed convolutional neural
network as N , which consists of a set of layers L and
contains weights W and biases b. Secondly, we introduce the
criticality metric according to Equation 1 and the evaluation
algorithm 1 which calculates the criticality for a given input
image xi, belonging to class i, drawn from a test set X .</p>
      <p>8&gt;y^i
fcr = &lt;&gt;&gt; 1 1y^mj ;
&gt;2;
&gt;
&gt;:0;
y^mi; if fm(xi) : (y^i</p>
      <p>y^mi)
if fm(xi) : y^mi
if fm(xi) : y^mi
otherwise;
y^mj and y^mj &lt; 0:5
y^mj and y^mj
0:5
(1)
where fcr returns a criticality with domain [0; 2] for a given
CNN which is masked, fm(xi) N . The masking of a
CNN is carried out by setting neurons’ weights to zero. In
case of convolution, all the values of a filter are set to zero.
A different kind of error modeling would lead to extensive
permutation and was therefore not further investigated.</p>
      <p>The term y^mi denotes the masked network’s prediction
confidence of ground-truth class i, whereas the predicted
confidence value y^mj belongs to another class j. In the first
case of fcr, the difference between the non-masked
predicted confidence y^i and the masked one y^mi is taken as
metric, by considering the parametrizable ”criticality” with
domain [0; 1].</p>
      <p>In the second case of fcr, the network missed the
groundtruth class and predicted a different one. Since this
misclassification can have severe consequences, we define the
criticality measure as the proportion of 1 and difference
between maximum likelihood and predicted confidence y^mj .
The denominator will always result in a number greater than
1, consequently ensuring the distinguishability of neurons
which have class-changing ability. Experiments in the early
phase showed that criticality can reach multi-digit number
and therefore we decided in the third case of fcr to clip its
maximum to 2, so that the results remain tractable. For all
other cases we set the criticality to 0. This covers the cases
where the network predicted the right class with negligible
deterioration of the confidence (&lt; ).</p>
      <p>It can occur that by masking a neuron the decision
likelihood of the correct class will increase, which will result in
a negative criticality. In this case, we refer to this neuron as
anti-critical to the related class i and calculate its criticality
according to Equation 2. However, it should be noted that
the anti-criticality will be computed only in case = 0 and
it doesn’t exclude the criticality of the neuron for a different
class j.</p>
      <p>fanti cr = y^i
y^mi; if fm(xi) : (y^i
y^mi) &lt; 0
(2)
3.2</p>
      <sec id="sec-3-1">
        <title>Analysis algorithm</title>
        <p>We define the task of NCA as the analysis of the neurons’
contribution to the classification hypothesis which can be
seen as equivalent to the Single Point Of Failure analysis.
If all neurons are active, the resulting hypothesis is strong
hstr, whereas in case a certain amount of neurons have been
excluded from the decision, the hypothesis is considered
weakened hweak. The neuron’s criticality observation of the
weakened hypothesis has to be done for every image and
class within a test set. The algorithm is described in
Algorithm 1.</p>
        <sec id="sec-3-1-1">
          <title>Algorithm 1: NCA algorithm</title>
        </sec>
        <sec id="sec-3-1-2">
          <title>Input: Criticality threshold</title>
          <p>Output: Neural criticality statistics for Xi
Data: Let X be a testing set, i a tested class, N the
analyzed CNN, k the number of filters in a
layer L and fcr is the criticality function
for image xi 2 X do
y^i = calculate conf (N; xi)
clsi = predict(N; xi)
for every L in N do
for every k in layer L do
mask neuron(k)
y^mi = calculate conf (N; xi)
clsmi = predict(N; xi)
criticality = fcr(y^i; y^mi; clsi; clsmi)
4</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Experiments</title>
      <p>
        The motivation behind testing different network
architectures was to see the influence of models’
chronological improvements on the decision stability, such as
residual connections, depthwise convolution and scaling. We
therefore evaluated VGG16, Resnet50V2, MobileNetV2
and EfficientNetB0, all pre-trained Keras models on
ImageNet
        <xref ref-type="bibr" rid="ref8">(Deng et al. 2009)</xref>
        . We chose two classes, ”street sign”
and ”mountain bike”, in order to evaluate the criticality. For
each class, 150 samples were taken. All samples had
groundtruth confidence higher than 0.8 so that we ensured that
kernels’ responses would be highly excitated. Adversary
samples were generated by non-target FSGD method until either
achieving a confidence greater than 0.5 or ending after 20
iterations. For all tests we set the criticality threshold to 0:0,
which allows, as described in Section 3.1, the algorithm to
measure and visualize the criticality of all neurons and
distinguish between critical and anti-critical ones. In practice,
the threshold should be justifiable via hazard and risk
assessment and will be presumably higher than 0:0.
4.1
      </p>
      <sec id="sec-4-1">
        <title>Neural criticality</title>
        <p>
          As a first step, we gathered statistics of every neuron as
described in Algorithm 1 with = 0:0. As a second step,
we normalized the list of hypotheses over the number of
layers’ neurons and highlighted in red the layers for which
masking at least one neuron caused a drop of confidence by
0:5 and more or lead to misclassification of the predicted
class. It is noticeable that in Figures 3 and 4, especially the
first projection layers have very high criticality. This
confirms the sparsity theory of projection layers
          <xref ref-type="bibr" rid="ref44">(Szegedy et al.
2014, 2016)</xref>
          , which states that projection layers are helpful
in terms of higher space feature extraction, whereas our
experiment shows that they cause an increase of criticality.
        </p>
        <p>VGG16 (Figure 1) and ResNet50v2 (Figure 2) have on
the other hand an average criticality spread over all layers.</p>
        <p>The experiments’ results shown in this work are only
related to the ”mountain bike” class. The results for the fairly
simple ”traffic sign” class backed up the intuition about
simple features being predominantly filtered in early
layers, since their criticality raised significantly. We advise the
reader to visit our GitHub, where additional figures and
stored statistics for both classes can be found.
To further evaluate the beneficial effects of neural
criticality, we conducted a stability experiment with n most critical
neurons (derived from NCA results) on original and
adversary datasets. We gradually masked the n most critical
neurons and calculated the mean and standard deviation of the
model’s accuracy on the aforementioned test set. The
intuition behind this test was that the mean accuracy increases
with respect to the criticality of lower neurons, and hence
proves our analysis’ reliability. In our text we refer to this
approach as Network Stability Analysis.</p>
        <p>As can be seen in Figure 5, gradually masking the 20 most
critical neurons has a major influence on the accuracy only
in case of MobileNetv2 and EfficientNetB0. VGG16 and
ResNet50V2, on the other hand, show a high accuracy
stability. Figure 6 shows the raising tendency of MobileNets’
accuracy with respect to lower neurons criticality, reaching
a mean accuracy of 0:8 approximately at the 50th most
critical neuron.</p>
        <p>Results of the accuracy on adversary dataset didn’t
confirm the hypothesis that critical neurons are the only neurons
allowing malicious adversary attacks. As can be seen in
Figure 9, masking the critical neuron generally doesn’t improve
the accuracy. On the other hand, several neurons lifted the
ground-truth class accuracy. The awareness of such neurons
could lead to on-the-fly diagnoses, where masking a
combination of specific neurons (e.g. only for xth frame, which
would be excluded from the classification or detection task)
would uncover irregularities in inference process, e.g.
adversary attack. It has to be mentioned that only critical neurons
from projection layers (MobileNetV2 and EfficientNetB0)
have such an ability, but they have to be chosen with respect
to the mean and standard deviation of the calculated
accuracy. Other models sensitivity to adversary noise are plotted
in Figure 8.</p>
        <p>5</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Conclusion</title>
      <p>At the beginning of this work in Chapter 2 we pointed out
the current functional safety issues and open research area
related to convolutional neural networks. In Section 1 we
described our motivation related to autonomous driving and
(b) ResNet50V2</p>
      <p>(a) VGG16
y
c
ra0.6
u
c
c
A
0.4
(b) ResNet50V2</p>
      <p>(a) VGG16
(c) MobileNetv2
(d) EfficientNetB0
Figure 8: Results of accuracy stability related to the 20 most critical neurons for all models, evaluated on adversary dataset. It
is obvious that different results of generating adversary attacks was achieved since initial models accuracy differs. The VGG16
model shows minimal accuracy fluctuation, whereas more modern models contains neurons with higher sensitivity to adversary
noise.</p>
      <p>0.8
missing validation process. Further, as outlined in Section 3,
we introduced a new metric and auxiliary analysis method
which we implemented and verified on four classification
CNNs in Chapter 4.</p>
      <p>We dedicated a great part of our work to introducing and
testing an innovative method, the Neurons’ Criticality
Analysis. The outcome of this analysis was a comprehensive
report diagram depicting the criticality of each layer and
each neuron of evaluated model. The domain of criticality
is [ 1; 2]. We discussed that masking neurons with negative
criticality can also have a positive influence on the model’s
decision confidence. We called this behavior ”anti-critical”.</p>
      <p>The inter-class anti-critical neurons could hypothetically be
removed from the decision process. This idea led us to the
conclusion that the correlation between the neurons removed
during the pruning process and the anti-critical neurons
discovered via NCA should be further investigated.</p>
      <p>We claimed that using spatial aggregation via projection
layers may on the one hand improve the high dimensional
feature representation(Szegedy et al. 2016), but on the other
hand creates very critical dense connections, especially in
the shallow layers, as we pointed out in Section 4.1. From
functional safety point of view this isn’t necessarily
negative, since the plausibility function could be applied to only a
concentrated area of neurons. In addition, some critical
neurons showed the ability to increase mean accuracy on
adversary dataset, which could be used in order to discover
adversary attacks and irregularities during inference. We
hypothesize that an equilibrium between the position of the first
projection layer, number of critical neurons and models’
accuracy should be further investigated.</p>
      <p>As aforementioned, the purpose of NCA is to identify all
critical neurons. With further measures, the mean and
standard deviation of the criticality should be decreased and the
flawless calculation of the neuron should be ensured.
Concretely this can be achieved by several approaches, such as:
• fine-tuning of the model with deterministic dropout and</p>
      <p>loss which will incorporate the layers criticality
• plausibility check of the critical neurons or layers or
re</p>
      <p>dundant computational branch results
• storage of the neurons’ weights and biases in two places</p>
      <p>in RAM and comparing them
• introduction of deconvolutional layers in order to compute
and evaluate the original inputs over critical connections</p>
      <p>Our method can also be used for Out-of-Distribution
detection, where instead of randomly sampling sub-networks
predictions, as it is done by MC dropout, deterministic
dropout would be based on several highly critical neurons
for every class. Such an approach would decrease the
computational demand and arguably increase the reliability and
transparency of such a network. In order to encourage
additional experiments and deeper explorations, we published
our code and supplement results on GitLab 1.</p>
      <p>6</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgment</title>
      <p>The work has been supported by the grant of the University
of West Bohemia, project No. SGS-2019-027.</p>
      <sec id="sec-6-1">
        <title>1https://gitlab.com/divisvaclav/cnn eval tool/-/tree/</title>
        <p>wo gui branch</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <surname>Bach</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Binder</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Montavon</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Klauschen</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ; Mu¨ller, K.-R.; and
          <string-name>
            <surname>Samek</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          <year>2015</year>
          .
          <article-title>On pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation</article-title>
          .
          <source>PloS one 10</source>
          <volume>(7)</volume>
          :
          <fpage>e0130140</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <surname>Belle</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Papantonis</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          <year>2020</year>
          .
          <article-title>Principles and Practice of Explainable Machine Learning</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <surname>BMW</surname>
          </string-name>
          , D. e. a.
          <year>2019</year>
          .
          <article-title>Safety First for Automated Driving (SaFAD) Safety First for Automated Driving (SaFAD).</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          https://www.daimler.com/innovation/case/autonomous/safetyfirst-for-
          <source>automated-driving-2</source>
          .html. Accessed:
          <fpage>2020</fpage>
          -06-08.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <surname>Brown</surname>
            , T. B.; Mane´,
            <given-names>D.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Roy</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Abadi</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Gilmer</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <article-title>Adversarial patch</article-title>
          .
          <source>arXiv preprint arXiv:1712</source>
          .
          <fpage>09665</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <surname>Cheng</surname>
          </string-name>
          , C.-H.
          <year>2020</year>
          .
          <article-title>Safety-Aware Hardening of 3D Object Detection Neural Network Systems</article-title>
          . arXiv preprint arXiv:
          <year>2003</year>
          .11242 .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <surname>Deng</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Dong</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ; Socher,
          <string-name>
            <given-names>R.</given-names>
            ;
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.-J.</given-names>
            ;
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            ; and
            <surname>Fei-Fei</surname>
          </string-name>
          ,
          <string-name>
            <surname>L.</surname>
          </string-name>
          <year>2009</year>
          .
          <article-title>Imagenet: A large-scale hierarchical image database</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <source>In Computer Vision and Pattern Recognition</source>
          ,
          <year>2009</year>
          .
          <source>CVPR</source>
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          IEEE Conference on,
          <fpage>248</fpage>
          -
          <lpage>255</lpage>
          . IEEE.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <surname>DeVries</surname>
          </string-name>
          , T.; and Taylor, G. W.
          <year>2018</year>
          .
          <article-title>Learning confidence for out-of-distribution detection in neural networks</article-title>
          .
          <source>arXiv preprint arXiv:1802</source>
          .04865 .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <surname>Eykholt</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Evtimov</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Fernandes</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Rahmati</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Xiao</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Prakash</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Kohno</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Song</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <year>2018</year>
          .
          <article-title>Robust physicalworld attacks on deep learning visual classification</article-title>
          .
          <source>In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition</source>
          ,
          <fpage>1625</fpage>
          -
          <lpage>1634</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <string-name>
            <surname>Gal</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Ghahramani</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <year>2016</year>
          .
          <article-title>Dropout as a bayesian approximation: Representing model uncertainty in deep learning</article-title>
          .
          <source>In international conference on machine learning</source>
          ,
          <fpage>1050</fpage>
          -
          <lpage>1059</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <string-name>
            <surname>Goodfellow</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          <year>2018</year>
          .
          <article-title>Defense Against the Dark Arts: An overview of adversarial example security research and future research directions</article-title>
          . arXiv preprint arXiv:
          <year>1806</year>
          .04169 .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <string-name>
            <surname>Goodfellow</surname>
            ,
            <given-names>I. J.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Shlens</surname>
          </string-name>
          , J.; and
          <string-name>
            <surname>Szegedy</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <year>2015</year>
          .
          <article-title>Explaining and Harnessing Adversarial Examples</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <string-name>
            <surname>Graves</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <year>2011</year>
          .
          <article-title>Practical variational inference for neural networks</article-title>
          .
          <source>In Advances in neural information processing systems</source>
          ,
          <volume>2348</volume>
          -
          <fpage>2356</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <string-name>
            <surname>He</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Ren</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Sun</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <year>2015</year>
          .
          <article-title>Deep Residual Learning for Image Recognition</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          <string-name>
            <surname>He</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Ren</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Sun</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <year>2016a</year>
          .
          <article-title>Deep residual learning for image recognition</article-title>
          .
          <source>In Proceedings of the IEEE conference on computer vision and pattern recognition</source>
          ,
          <fpage>770</fpage>
          -
          <lpage>778</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          <string-name>
            <surname>He</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Ren</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Sun</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <year>2016b</year>
          .
          <article-title>Identity mappings in deep residual networks</article-title>
          .
          <source>In European conference on computer vision</source>
          , 630-
          <fpage>645</fpage>
          . Springer.
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          <string-name>
            <surname>Henne</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Schwaiger</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Roscher</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ; and Weiss,
          <string-name>
            <surname>G.</surname>
          </string-name>
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          <article-title>Benchmarking Uncertainty Estimation Methods for Deep Learning With Safety-Related Metrics</article-title>
          . In SafeAI@ AAAI,
          <fpage>83</fpage>
          -
          <lpage>90</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          <string-name>
            <surname>Howard</surname>
            ,
            <given-names>A. G.</given-names>
          </string-name>
          ; Zhu,
          <string-name>
            <given-names>M.</given-names>
            ;
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            ;
            <surname>Kalenichenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            ;
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <surname>W.</surname>
          </string-name>
          ; Weyand,
          <string-name>
            <given-names>T.</given-names>
            ;
            <surname>Andreetto</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          ; and Adam,
          <string-name>
            <surname>H.</surname>
          </string-name>
          <year>2017</year>
          .
          <article-title>Mobilenets: Efficient convolutional neural networks for mobile vision applications</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          <source>arXiv preprint arXiv:1704</source>
          .
          <fpage>04861</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          <string-name>
            <surname>ISO.</surname>
          </string-name>
          <year>2011</year>
          .
          <article-title>Road vehicles - Functional safety</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          <string-name>
            <surname>ISO2019.</surname>
          </string-name>
          <year>2019</year>
          .
          <article-title>Road vehicles - Safety of the intended functionality</article-title>
          .
          <source>RFC 1654</source>
          , RFC Editor. URL http://www.rfc-editor.
          <source>org/rfc/ rfc1654.txt.</source>
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          <string-name>
            <surname>Kendall</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ; and Gal,
          <string-name>
            <surname>Y.</surname>
          </string-name>
          <year>2017</year>
          .
          <article-title>What uncertainties do we need in bayesian deep learning for computer vision</article-title>
          ? In
          <source>Advances in neural information processing systems</source>
          ,
          <volume>5574</volume>
          -
          <fpage>5584</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          <string-name>
            <surname>Lakshminarayanan</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Pritzel</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Blundell</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <year>2017</year>
          .
          <article-title>Simple and scalable predictive uncertainty estimation using deep ensembles</article-title>
          .
          <source>In Advances in neural information processing systems</source>
          ,
          <volume>6402</volume>
          -
          <fpage>6413</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Peng</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Yu</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ; Zhang,
          <string-name>
            <given-names>X.</given-names>
            ;
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            ; and
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <surname>J.</surname>
          </string-name>
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          <string-name>
            <surname>Lin</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Yan</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <year>2014</year>
          . Network In Network.
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ; Liu,
          <string-name>
            <given-names>C.</given-names>
            ; and
            <surname>Song</surname>
          </string-name>
          ,
          <string-name>
            <surname>D.</surname>
          </string-name>
          <year>2016</year>
          .
          <article-title>Delving into transferable adversarial examples and black-box attacks</article-title>
          .
          <source>arXiv preprint arXiv:1611</source>
          .
          <fpage>02770</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          <string-name>
            <given-names>Mobileye. 2020. Advanced</given-names>
            <surname>Technologies - Mobileye Future</surname>
          </string-name>
          of Mobility - Advanced
          <string-name>
            <surname>Technologies</surname>
          </string-name>
          . https://www.mobileye.com/. Accessed:
          <fpage>2020</fpage>
          -11-13.
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          <string-name>
            <surname>Neal</surname>
            ,
            <given-names>R. M.</given-names>
          </string-name>
          <year>2012</year>
          .
          <article-title>Bayesian learning for neural networks</article-title>
          , volume
          <volume>118</volume>
          . Springer Science &amp; Business Media.
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          <string-name>
            <surname>Rafegas</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ; Vanrell,
          <string-name>
            <given-names>M.</given-names>
            ;
            <surname>Alexandre</surname>
          </string-name>
          ,
          <string-name>
            <surname>L. A.</surname>
          </string-name>
          ; and Arias,
          <string-name>
            <surname>G.</surname>
          </string-name>
          <year>2019</year>
          .
          <article-title>Understanding trained CNNs by indexing neuron selectivity</article-title>
          .
          <source>Pattern Recognition Letters .</source>
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          <string-name>
            <surname>Salay</surname>
          </string-name>
          , R.; and
          <string-name>
            <surname>Czarnecki</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <year>2018</year>
          .
          <article-title>Using machine learning safely in automotive software: An assessment and adaption of software process requirements in ISO 26262</article-title>
          . arXiv preprint arXiv:
          <year>1808</year>
          .01614 .
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          <string-name>
            <surname>Sandler</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Howard</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Zhu</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Zhmoginov</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Chen</surname>
          </string-name>
          , L.- C.
          <year>2018</year>
          .
          <article-title>Mobilenetv2: Inverted residuals and linear bottlenecks</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          <source>In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition</source>
          ,
          <fpage>4510</fpage>
          -
          <lpage>4520</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          <string-name>
            <surname>Shridhar</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Laumann</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ; and Liwicki,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <year>2019</year>
          .
          <article-title>A comprehensive guide to bayesian convolutional neural network with variational inference</article-title>
          .
          <source>arXiv preprint arXiv:1901</source>
          .02731 .
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          <string-name>
            <surname>Shrikumar</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Greenside</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Kundaje</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <year>2017</year>
          .
          <article-title>Learning important features through propagating activation differences</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          <source>In Proceedings of the 34th International Conference on Machine Learning-</source>
          Volume
          <volume>70</volume>
          ,
          <fpage>3145</fpage>
          -
          <lpage>3153</lpage>
          . JMLR. org.
        </mixed-citation>
      </ref>
      <ref id="ref40">
        <mixed-citation>
          <string-name>
            <surname>Simonyan</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Vedaldi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Zisserman</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <year>2013</year>
          .
          <article-title>Deep inside convolutional networks: Visualising image classification models and saliency maps</article-title>
          .
          <source>arXiv preprint arXiv:1312</source>
          .
          <fpage>6034</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref41">
        <mixed-citation>
          <string-name>
            <surname>Simonyan</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Zisserman</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <year>2014</year>
          .
          <article-title>Very deep convolutional networks for large-scale image recognition</article-title>
          .
          <source>arXiv preprint arXiv:1409</source>
          .
          <fpage>1556</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref42">
        <mixed-citation>
          <string-name>
            <surname>Song</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>He</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Hopcroft</surname>
            ,
            <given-names>J. E.</given-names>
          </string-name>
          <year>2018</year>
          .
          <article-title>Improving the Generalization of Adversarial Training with Domain Adaptation</article-title>
          . arXiv preprint arXiv:
          <year>1810</year>
          .00740 .
        </mixed-citation>
      </ref>
      <ref id="ref43">
        <mixed-citation>
          <string-name>
            <surname>Sundararajan</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Taly</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Yan</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          <year>2017</year>
          .
          <article-title>Axiomatic attribution for deep networks</article-title>
          .
          <source>In Proceedings of the 34th International Conference on Machine Learning-</source>
          Volume
          <volume>70</volume>
          ,
          <fpage>3319</fpage>
          -
          <lpage>3328</lpage>
          . JMLR.
        </mixed-citation>
      </ref>
      <ref id="ref44">
        <mixed-citation>
          <string-name>
            <surname>Szegedy</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ; Jia,
          <string-name>
            <given-names>Y.</given-names>
            ;
            <surname>Sermanet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            ;
            <surname>Reed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            ;
            <surname>Anguelov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            ;
            <surname>Erhan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            ;
            <surname>Vanhoucke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            ; and
            <surname>Rabinovich</surname>
          </string-name>
          ,
          <string-name>
            <surname>A.</surname>
          </string-name>
          <year>2014</year>
          .
          <article-title>Going Deeper with Convolutions</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref45">
        <mixed-citation>
          2016.
          <article-title>Rethinking the inception architecture for computer vision</article-title>
          .
          <source>In Proceedings of the IEEE conference on computer vision and pattern recognition</source>
          ,
          <fpage>2818</fpage>
          -
          <lpage>2826</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref46">
        <mixed-citation>
          <string-name>
            <surname>Tan</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Le</surname>
            ,
            <given-names>Q. V.</given-names>
          </string-name>
          <year>2019</year>
          .
          <article-title>EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks</article-title>
          . arXiv preprint arXiv:
          <year>1905</year>
          .11946 .
        </mixed-citation>
      </ref>
      <ref id="ref47">
        <mixed-citation>
          <string-name>
            <surname>Tesla</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Autopilot and Full Self-Driving Capability Autopilot and Full Self-Driving Capability</article-title>
          . https://www.tesla.com/support/ autopilot. Accessed:
          <fpage>2020</fpage>
          -06-08.
        </mixed-citation>
      </ref>
      <ref id="ref48">
        <mixed-citation>
          <string-name>
            <surname>Zeiler</surname>
            , M. D.; and Fergus,
            <given-names>R.</given-names>
          </string-name>
          <year>2014</year>
          .
          <article-title>Visualizing and understanding convolutional networks</article-title>
          .
          <source>In European conference on computer vision</source>
          , 818-
          <fpage>833</fpage>
          . Springer.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>