<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Repair of Convolutional Neural Networks using Convex Optimization: Preliminary Experiments</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Dario Guidotti</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Francesco Leofante</string-name>
          <email>francesco.leofanteg@edu.unige.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>RWTH Aachen University</institution>
          ,
          <addr-line>Aachen</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>University of Genoa</institution>
          ,
          <addr-line>Genoa</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>University of Sassari</institution>
          ,
          <addr-line>Sassari</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Recent public calls for the development of explainable and veri able Arti cial Intelligence (AI) led to a growing interest in formal veri cation and repair of machine-learned models. Despite the impressive progress that the learning community has made, models such as deep neural networks remain vulnerable to adversarial attacks, and their sheer size represents a major obstacle to formal analysis and implementation. In this paper, we present our current e orts to tackle repair of deep convolutional neural networks using ideas borrowed from Transfer Learning. Using results obtained on popular MNIST and CIFAR10 datasets, we show that models of deep convolutional neural networks can be transformed into simpler ones preserving their accuracy, and we discuss how formal repair through convex programming techniques could bene t from this process.</p>
      </abstract>
      <kwd-group>
        <kwd>Transfer Learning Network Repair Convex Optimization</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>The need for the development of explainable and veri able AI has been put
forward in a number of public events, e.g., the Workshop on Explainable AI held at
IJCAI 20174, and research programs, e.g., the DARPA program on Explainable
Arti cial Intelligence.5 These \calls to arms" did not go unanswered,
originating several related research streams. Among them, particularly vibrant is the
one concerned with automated veri cation and repair of Deep Neural Networks
(DNNs).</p>
      <p>
        Despite the impressive progress that the learning community has made in
the eld, it is well known | see, e.g., [
        <xref ref-type="bibr" rid="ref16 ref3">16, 3</xref>
        ] | that DNNs can be vulnerable to
adversarial perturbations, i.e., minimal changes to correctly classi ed input data
that cause a network to respond in unexpected and incorrect ways.
Independently from the accuracy of a network, the vulnerability to adversarial attacks
calls for techniques to improve robustness and guarantee desired properties.
Repair [
        <xref ref-type="bibr" rid="ref12 ref4 ref5">12, 4, 5</xref>
        ] is one such technique, whereby we seek to adjust the parameters
4 http://home.earthlink.net/ dwaha/research/meetings/ijcai17-xai/
5 https://www.darpa.mil/program/explainable-arti cial-intelligence
of the network in order to formally guarantee that the network will respond
correctly even in the presence of adversarial perturbations. In practice, the sheer
size of these models still represents a major obstacle to formal analysis of any
kind. Typical state of the art neural networks for tasks like image classi cation
have more than a hundred millions of parameters [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ], which makes o -the-shelf
techniques hardly applicable.
      </p>
      <p>
        In this paper we focus on the repair of Convolutional Neural Networks
(CNNs), a type of DNN mainly used in computer vision | see [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] for a survey
related to DNN architectures and their applications. In particular, we discuss
our current e orts to repair CNNs using convex programming and ideas
borrowed from Transfer Learning (TL) [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ]. As posited in [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], the idea is to keep
the convolutional part of the network as a learned feature extractor, and replace
the nal classi cation layer with one featuring less parameters and/or a smaller
model complexity. Noticeably, the replacement may yield networks whose
accuracy is comparable with the one of the original DNNs, yet more amenable to
formal analysis.
      </p>
      <p>
        More speci cally, we contribute an experimental analysis based on the
popular MNIST6 and CIFAR107 datasets. The rst step is to train CNNs on both
datasets and to replace their nal fully-connected layer with linear support
vector machines. This step reduces by orders of magnitude the number of free
parameters, e.g., from 4:7 million to 65 thousand in the case of CIFAR10, while
preserving accuracy. We discover adversarial attacks on such \hybrid" models
using the Fast Gradient Sign Method described in [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Since we replace
nonlinear layers with linear ones, we are able to de ne a repair procedure as a convex
optimization problem | as done in [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] for kernel-based learning models. The
resulting problem can be solved with o -the-shelf tools | cvxopt8 in our case.
The results we obtain are still preliminary, but show some promise as far as
scaling to networks of larger size is concerned. However, the repair procedure
is not yet general enough to make the networks immune to perturbations other
than those considered by the repair procedure. While this might not be a strong
limitation when considering real-world instead of arti cial adversaries | see [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]
| further investigations are needed to con rm whether our method could be
e ective for practical applications.
      </p>
      <p>
        To the extent of our knowledge, this is the rst time that TL is leveraged
in order to repair a CNN through convex programming techniques. The idea of
replacing parts of a CNN to improve its performances is not new, as it has been
explored in [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] and [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], among others. However, the focus of these contributions
is to improve the accuracy of the network, rather than providing models whose
properties can be certi ed more easily than the original one. Trying to apply
formal veri cation techniques to networks of size smaller than the original could
be done following alternative paths. For instance, in [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] the authors show that it
is possible to nd small-sized subnetworks in CNNs which prove to be remarkably
6 http://yann.lecun.com/exdb/mnist/
7 https://www.cs.toronto.edu/ kriz/cifar.html
8 https://cvxopt.org/index.html
accurate on some datasets including MNIST and CIFAR10. These subnetworks
could be extracted and certi ed considering our approach or other
state-of-theart tools like Marabou [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Finally, the aim of obtaining networks robust to
adversarial examples, but not necessarily smaller than the original ones, can be
pursued using results in robust training: recent results | see, e.g., [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] | seem
to open this possibility also for CNNs of considerable size.
      </p>
    </sec>
    <sec id="sec-2">
      <title>Preliminaries</title>
      <sec id="sec-2-1">
        <title>Convolutional Neural Networks</title>
        <p>
          According to [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ], representation learning \is a set of methods that allows a
machine to be fed with raw data and to automatically discover the representations
needed for detection of classi cation". DNNs are representation learning
models characterized by multiple levels of representation, obtained by composing
several non-linear modules. Each module transforms the representation at one
level (starting with raw input) into the next, more abstract, representation. At
the heart of every DNN lie the \classical" neural network modules as shown in
Figure 1 (left) whose mathematical formulation can be expressed in a recursive
form as:
h(1) =
h(i) =
(1)(W(1) x + b(1))
(i)(W(i) h(i 1) + b(i))
(1)
where (i) is the activation function, W(i) 2 Rdi di 1 is a matrix of weights and
b(i) 2 Rdi is the vector of the biases of the i-th layer. h(i) 2 Rdi corresponds to
the output of the i-th layer and the range of i depends on the number of layers.
A module like this is said to be fully connected, because the weighted sum of the
outputs of each neuron in level i is fed to every neuron in level i + 1, creating the
topology shown in Figure 1 (left). CNNs are a specifc kind of DNNs, typically
adopted in computer vision applications, characterized by one or more
convolutional modules. The distinctive element of such modules is that they feature
connections for small, localized regions of the input vector, i.e., each neuron of
the hidden layer is connected only to a small subset of the input neurons. This
subset of the input neurons is called local receptive eld of the hidden neuron.
A graphical example of a local receptive is depicted in Figure 1 (right). Another
important feature of convolutional modules is that all the local receptive elds
share the same weights and bias reducing the overall number of weights
substantially. In practice, each local receptive eld is trained to detect a speci c feature
in the input image, i.e., distinctive elements of input portions. As a consequence,
in a speci c hidden layer, di erent sets of shared weights are used: each of these
sets is trained to detect speci c feature in the image. Usually each convolutional
module is followed by a pooling layer, which simpli es the information received.
For instance, each unit of a pooling layer could take a subset of neurons from
the previous module and select their maximum activation | an operation called
max-pooling. Since our experiments are about image classi cation, in the
following we consider a CNN arrangement widely adopted for this task, i.e., a series of
convolutional modules and pooling layers followed by fully connected modules.
The rst part of the network can be seen as an application of a (learned) kernel
to the original input whereas the second part can be seen as the actual classi er.
For a more detailed study on Convolutional Neural Networks we refer to [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ].
2.2
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>Transfer Learning</title>
        <p>
          As mentioned in [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ], TL is \the improvement of learning in a new task through
the transfer of knowledge from a related task that has already been learned". TL
has been suggested in the context of deep learning applications | see, e.g., [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ]
| where pre-trained models are used as starting points for computer vision or
natural language tasks. Since the training of DNNs requires substantial
computational resources, it is often the case that reusing (parts of) pre-trained models
enables applications which would not be feasible otherwise. For instance, in [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ],
a pre-trained convolutional module is extracted from a CNN and then applied
as a feature extractor in the context of an object recognition task where the
paucity of training samples would make training of the full CNN untenable. On
the other hand, combining the pre-trained convolutional module with a newly
trained classi er, makes for an e ective combination, enabling to solve
classication tasks that were not within the reach of the original CNN. TL and its
applications suggest the possibility of replacing some modules of a DNN which
are hardly analyzable with formal methods, with others that are more amenable
to such analysis. As long as the accuracy of the resulting network, which we call
hybrid network in the following, is close to the original DNN, one may (i) replace
the original network with the hybrid one and (ii) x the hybrid one instead of
the original network, should adversarial examples be found also for the hybrid
network. In particular, we build hybrid networks by collating the convolutional
module of a CNN followed by a linear Support Vector Machine (lSVM), i.e., a
classi er based on separating hyper-planes in which the distance of the
hyperplane from the nearest samples of both classes is maximized. In our experiments
we consider multiclass lSVMs, i.e., in order to discriminate among k classes we
compute k di erent separating hyper-planes each one discriminating among one
class and the remaining k 1. The input-output relation of a multiclass lSVM
f (x) = W x + b
y = argmax(f (x))
where x 2 Rd is the vector of the inputs, b 2 Rk is the vector of the biases,
W 2 Rk d is the matrix of the weights corresponding to k lSVMs, each working
to detect one of the k classes. The function f (x) is the decision function
corresponding to the input x. It contains the signed distances of the input x from
each decision hyper-plane. From the de nition of the decision function we can
derive the correct class y for an input x.
(2)
is de ned as follows:
3
3.1
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Repair of hybrid networks</title>
      <sec id="sec-3-1">
        <title>Hybrid Networks</title>
        <p>For the sake of our experiments, we have developed two CNNs and two
corresponding hybrid networks for each dataset considered. Given the preliminary
nature of this work the datasets considered are CIFAR10 and MNIST, two of
the most famous basic datasets for image classi cation. The MNIST dataset
contains 60000 black-and-white images of handwritten digits whereas the CIFAR10
dataset contains 60000 color images in 10 di erent mutually exclusive classes:
both datasets are divided in a training set of 50000 images and a test set of
10000 images.</p>
        <p>The network considered for the MNIST dataset (MNIST-NN) is a CNN with
2 convolutional layers, 2 max-pooling layers and 2 fully connected layers. The
convolutional layers have kernel size equal to 5 5 and stride length equal to 1,
the max-pooling layers have kernel size equal to 2 2. The two fully connected
layers have 500 and 10 hidden neurons respectively and the inputs of the rst
layer are the value generated from 800 neurons of the second max-pooling layer.
The activation functions are all ReLU. The total number of parameters of the
network is 407330 and 99:5% of them are part of the fully connected layers.</p>
        <p>
          The network developed for the CIFAR10 dataset (CIFAR10-NN) is a CNN
with 6 convolutional layers, 3 max-pooling layer and 3 fully connected layers.
The convolutional layers have kernel size equal to 3 3, stride length equal to
1, padding equal to 1 and present respectively 32, 64, 128, 128, 256 and 256
di erent kernels, the max-pooling layers has kernel size equal to 2 2. There
is a max pooling layer every 2 convolutional layer. The three fully connected
layers have 1024, 512 and 10 hidden neurons respectively and the inputs of the
rst layer are the values generated from 4096 neurons of the third max-pooling
layer. The activation functions are all ReLU. The total number of parameters is
4747904 and 99:5% of them are part of the fully connected layers. The network
considered is similar to the Conv-6 network presented in [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ], but our network
features a rst convolutional layer with 32 kernels whereas the corresponding
layer of the Conv-6 network has 64 kernels.
        </p>
        <p>
          In this work we have used PyTorch [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ] for the implementation and training
of all the networks. The hybrid networks consist of the union of the convolutional
and max-pooling layers of the original networks with lSVM multiclass classi ers:
in this work we have used o -of-the-shelf implementations provided by
scikitlearn [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ]. In particular, both for MNIST-NN and for CIFAR10-NN, we have
designed corresponding linear and non-linear hybrids: the non-linear hybrids
use as kernel the standard radial basis function. We identify the linear hybrids
as MNIST-LH and CIFAR10-LH and the non-linear hybrids as MNIST-KH and
CIFAR10-KH. All networks are trained using standard training parameters
recommended respectively from PyTorch and scikit-learn documentation.
        </p>
        <p>As a preliminary experiment we have analyzed the accuracy gap between
the hybrid models and the corresponding neural networks: for CIFAR10 models
our results are 85:4% (NN), 85:51% (KH) and 85:6% (LH). The accuracies of
the MNIST models are 97:12% (NN), 98:86% (KH) and 98:72% (LH). All the
accuracies were computed as the number of correctly classi ed images against
all the images of the test sets provided by the MNIST and CIFAR10
repositories. These gures tell us that, for the MNIST dataset, hybrid models can be
more accurate than the corresponding CNN, with the kernel-based hybrid
being slightly more accurate than the linear one. CIFAR10 is more complex than
MNIST, nevertheless the results still hold.
3.2</p>
      </sec>
      <sec id="sec-3-2">
        <title>Repair</title>
        <p>
          The main idea behind our repair approach is to circumvent the repair of CNNs
and attempt to repair the corresponding hybrid networks instead. To repair
hybrid networks, we generate adversarial examples for them, and then we solve
an optimization problem in the space of the network's parameter, with the
objective to reduce as much as possible the impact of the adversaries. In order to
make the optimization problem computationally feasible, we consider the
convolutional modules of our hybrid models as a xed feature map and we do not
include their parameters into the optimization problem, but we concentrate on
the nal layers instead. In practice, this corresponds to analyzing the network
in feature space, instead of input space | as done in [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]. Owing to this, and to
the fact that fully connected layers of the CNN are substituted by (l)SVMs in
our hybrid networks, the number of free variables for the convex optimization
problem is drastically reduced. For example, in the case of the MNIST models,
we managed to reduce the number of variables from 405510 to approximately
8000.
        </p>
        <p>In this rst stage of our work, we decided to further simplify the problem
considered by excluding KH models: in this way it is possible to limit the convex
optimization problem to piece-wise linearity | because of absolute values |
eliminating the need of a non-linear solver or abstraction techniques to manage
the radial basis function kernels. In equation (3) we present the mathematical
de nition of our optimization problem: parameters c and d are the number of
possible classes and the number of features of the adversarial sample in the
feature space, respectively; parameters i;j are the modi cation on the weights
wi;j of the lSVM model; the variables i are slack variable necessary to keep
the problem solvable at the price of some error on the prediction of the decision
function for the adversarial sample of interest; nally, yi are the correct values
of the decision function of the lSVM classi er for the adversarial sample and xj
are the features of the adversarial sample of interest in the feature space. All the
variable considered take real values.</p>
        <p>c d
min X X
j i;j j +</p>
        <p>c
X</p>
        <p>i
yi
i
i=1 j=1 i=1</p>
        <p>d
i X(wi;j + i;j )xj</p>
        <p>j=1
0
yi + i
8i = 1; :::; c
8i = 1; :::; c
(3)
In this case we consider only one adversarial, but the extension of the problem
to the case in which more than one adversarial sample is considered is trivial.
The cost function seeks to minimize the (absolute) variation of the weights of the
lSVM, while satisfying the constraint of bringing the prediction of the decision
function of the adversarial example as close as possible to the correct decision
function. In the case of the CIFAR10 model we need a further simpli cation:
even with the replacement of the fully connected layers the number of variables
in the convex optimization problem is 40960 and the optimization procedure is
not able to solve the problem. Therefore we decided to apply a feature-selection
procedure on the output of the convolutional layers of our model. For each
feature we consider two set of samples: the rst one taken from the original
inputs and the second one taken from the adversarial inputs. We compare the
sets of samples using the Wilcoxon Signed Rank test against the null hypothesis
that the two sets come from the same distribution. The procedure computes the
p-values of the test for each feature and selects the ones which present a p-value
below a given threshold: in our experiments we choose a threshold value of 0:1.
The rationale of our procedure is to retain only the features which are a ected
signi cantly by the adversarial inputs and change only the corresponding weights
in the SVM. After feature selection we manage to reduce the number of variables
of the convex optimization problem below 6000, therefore making the problem
manageable for the solver.
4</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Experimental Results</title>
      <p>
        We test our repair procedure on both the MNIST and CIFAR10 datasets, using
the Fast Sign Gradient Method [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] as adversarial attack of choice. In order
to generate adversarial samples for our models we utilize FoolBox [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] which
provides a number of ready-made adversarial attacks. Another advantage of
FoolBox is that it accepts as model to be attacked every valid PyTorch model,
which allows us to attack also our hybrid models without complex workarounds.
In our tests, rstly we analyze the loss of accuracy of our models corresponding
to increasing magnitudes of the adversarial attack. We call the parameter which
control this magnitude ", and we show our results in gure 2.
      </p>
      <p>1.0
0.9
0.8
0.7</p>
      <p>As it can be expected the accuracies of both the MNIST and CIFAR10 models
drop for increasing values of " and in general the LH models seem to be more
vulnerable to this kind of adversarial attack. Given that our aim is to repair
the LH models, this is not a limitation for our approach. As it can be observed
in Figure 2 for " = 0:15 the adversarial perturbation for the MNIST images is
clearly recognizable even if it would not fool a human observer. For the CIFAR10
dataset we consider smaller adversarial perturbations: as can be seen in Figure 2
for " = 0:025 the models accuracy is already below the baseline.</p>
      <p>In our main experiment we analyzed the behavior of MNIST-LH and of
its repaired version (MNIST-RLH) for " = [0:025; 0:05; 0:075; 0:1; 0:125; 0:15]
and the behavior of CIFAR10-LH and of its repaired version (CIFAR10-RLH)
0.001 0.005 0.010</p>
      <p>0.015 0.020 0.025
Epsilon
for " = [:001; :005; :01; :015; 0:02; 0:025]. More speci cally, for each ", we
compute the accuracies of the LH, NN and RLH models on the following test
sets: MNIST/CIFAR10 test set (Data), MNIST/CIFAR10 test sets in which
all the images for which we found a corresponding adversarial example have
been replaced with the adversarial example. Since the adversarial attack is
model-dependent, the latter test set corresponds to three di erent sets
computed on MNIST/CIFAR10-NN (NADV), MNIST/CIFAR10-LH (HADV) and
MNIST/CIFAR10-RLH (RHADV).</p>
      <p>In gure 3 (left) it is possible to see how repair a ects the accuracy of the
models both with respect to adversarial samples only (ADV) and with respect
to the original test set (Data). In the case of MNIST, even considering only
one adversarial in the optimization problem (3), the resulting model
(MNISTRLH) manages to generalize also on other adversarial examples, e.g. it manages
to classify correctly at least 20% of the adversarial examples. In the case of
CIFAR10, even if the repaired model is more accurate than the original one
with respect to the adversarial samples the improvement is not substantial; in
our opinion this is due to the fact that both the model and the dataset are
more complex than the ones in MNIST. In gure 3 (right) it is also possible
to see how the accuracy of the RLH models compares with the accuracies of
the LH and NN ones with respect to the datasets NADV and HADV: from the
images on the right it is clear that, while the repaired model is more robust to
the adversarial sample computed on the non-repaired model, it has not acquired
robustness against adversarial attacks in general. Moreover, it appears clear that
the original model (NN) is still more robust to adversarial attacks. From the same
images it is also possible to see that, as the RLH models are somewhat robust
with respect to the adversarial example computed on the LH ones, so the LH
models are somewhat robust with respect to the adversarial example computed
on the RLH ones. This result suggests that the adversarial examples computed on
the LH and RLH models belong to di erent categories of adversarial examples.
This phenomenon requires further investigation to be con rmed.
5</p>
    </sec>
    <sec id="sec-5">
      <title>Conclusions and Future Work</title>
      <p>
        The main idea presented in this paper is to study the safety of DNNs in a
"modular" fashion using techniques adopted from transfer learning. In particular, we
consider how the properties of CNNs change if we swap the fully connected
module with lSVMs obtaining hybrid networks. Our results con rm that such
swap does not impact on the accuracy in a relevant manner, while making repair
of hybrid networks feasible using a relatively simple encoding in a convex
optimization problem. Adversarial examples can be found on hybrid networks more
easily than on the original network: this result conforms to the hypothesis about
the nature of adversarial examples presented in [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Our experimental results
on MNIST show that, even using very few adversaries, the repair procedures
manage to provide a model which presents an acceptable generalization on all
the adversaries computed using the original hybrid model. On the other hand,
our results on CIFAR10 show a more intricate picture, one in which the repaired
network can be made robust against speci c adversaries but generalization is
still not completely achieved.
      </p>
      <p>Given the results obtained from this work, our future lines of research will
concentrate on understanding the properties of categories of adversarial samples
in hybrid convolutional-lSVM networks and adding veri cation-driven kernels to
our SVMs in order to obtain robust hybrid convolutional-SVM networks.
Moreover, we will try to extend our work in order to repair CNNs without swapping
away the fully connected modules and to explain how adversarial attacks a ect
the convolutional part of the networks and therefore the input in the feature
space.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Frankle</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Carbin</surname>
            ,
            <given-names>M.:</given-names>
          </string-name>
          <article-title>The lottery ticket hypothesis: Finding sparse, trainable neural networks</article-title>
          .
          <source>arXiv preprint arXiv:1803</source>
          .
          <volume>03635</volume>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Gilmer</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Adams</surname>
            ,
            <given-names>R.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goodfellow</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Andersen</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dahl</surname>
            ,
            <given-names>G.E.</given-names>
          </string-name>
          :
          <article-title>Motivating the rules of the game for adversarial example research</article-title>
          . arXiv preprint arXiv:
          <year>1807</year>
          .
          <volume>06732</volume>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Goodfellow</surname>
            ,
            <given-names>I.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shlens</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Szegedy</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Explaining and harnessing adversarial examples</article-title>
          .
          <source>In: ICLR</source>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Guidotti</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Leofante</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Castellini</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tacchella</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Repairing learned controllers with convex optimization: A case study</article-title>
          .
          <source>In: CPAIOR</source>
          . pp.
          <volume>364</volume>
          {
          <issue>373</issue>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Guidotti</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Leofante</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tacchella</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Castellini</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Improving reliability of myocontrol using formal veri cation</article-title>
          .
          <source>IEEE Transactions on Neural Systems and Rehabilitation Engineering</source>
          <volume>27</volume>
          (
          <issue>4</issue>
          ),
          <volume>564</volume>
          {
          <fpage>571</fpage>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Katz</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>D.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ibeling</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Julian</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lazarus</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lim</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shah</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Thakoor</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zeljic</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dill</surname>
            ,
            <given-names>D.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kochenderfer</surname>
            ,
            <given-names>M.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Barrett</surname>
            ,
            <given-names>C.W.:</given-names>
          </string-name>
          <article-title>The marabou framework for veri cation and analysis of deep neural networks</article-title>
          .
          <source>In: CAV</source>
          . pp.
          <volume>443</volume>
          {
          <issue>452</issue>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>LeCun</surname>
          </string-name>
          , Y.,
          <string-name>
            <surname>Bengio</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hinton</surname>
          </string-name>
          , G.E.:
          <article-title>Deep learning</article-title>
          .
          <source>Nature</source>
          <volume>521</volume>
          (
          <issue>7553</issue>
          ),
          <volume>436</volume>
          {
          <fpage>444</fpage>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zeng</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Alsaadi</surname>
            ,
            <given-names>F.E.</given-names>
          </string-name>
          :
          <article-title>A survey of deep neural network architectures and their applications</article-title>
          .
          <source>Neurocomputing</source>
          <volume>234</volume>
          ,
          <issue>11</issue>
          {
          <fpage>26</fpage>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Nielsen</surname>
            ,
            <given-names>M.A.</given-names>
          </string-name>
          :
          <article-title>Neural networks and deep learning</article-title>
          , vol.
          <volume>25</volume>
          . Determination press San Francisco, CA, USA: (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Paszke</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gross</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chintala</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chanan</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>DeVito</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lin</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Desmaison</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Antiga</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lerer</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Automatic di erentiation in pytorch (</article-title>
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Pedregosa</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Varoquaux</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gramfort</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Michel</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Thirion</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grisel</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Blondel</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Prettenhofer</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Weiss</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dubourg</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vanderplas</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Passos</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cournapeau</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brucher</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Perrot</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Duchesnay</surname>
          </string-name>
          , E.:
          <article-title>Scikit-learn: Machine learning in Python</article-title>
          .
          <source>Journal of Machine Learning Research</source>
          <volume>12</volume>
          ,
          <volume>2825</volume>
          {
          <fpage>2830</fpage>
          (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Pulina</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tacchella</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>An abstraction-re nement approach to veri cation of arti cial neural networks</article-title>
          .
          <source>In: CAV</source>
          . pp.
          <volume>243</volume>
          {
          <issue>257</issue>
          (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Rauber</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brendel</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bethge</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Foolbox: A python toolbox to benchmark the robustness of machine learning models</article-title>
          .
          <source>arXiv preprint arXiv:1707.04131</source>
          (
          <year>2017</year>
          ), http://arxiv.org/abs/1707.04131
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Schwarz</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schulz</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Behnke</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Rgb-d object recognition and pose estimation based on pre-trained convolutional neural network features</article-title>
          .
          <source>In: ICRA</source>
          . pp.
          <volume>1329</volume>
          {
          <issue>1335</issue>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Simonyan</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zisserman</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Very deep convolutional networks for large-scale image recognition</article-title>
          .
          <source>arXiv preprint arXiv:1409.1556</source>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Szegedy</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zaremba</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sutskever</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bruna</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Erhan</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goodfellow</surname>
            ,
            <given-names>I.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fergus</surname>
          </string-name>
          , R.:
          <article-title>Intriguing properties of neural networks</article-title>
          .
          <source>In: ICLR</source>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Tang</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          :
          <article-title>Deep learning using linear support vector machines</article-title>
          .
          <source>arXiv preprint arXiv:1306.0239</source>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Torrey</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shavlik</surname>
          </string-name>
          , J.:
          <article-title>Transfer learning</article-title>
          .
          <source>In: Handbook of research on machine learning applications and trends: algorithms</source>
          , methods, and techniques, pp.
          <volume>242</volume>
          {
          <fpage>264</fpage>
          .
          <string-name>
            <given-names>IGI</given-names>
            <surname>Global</surname>
          </string-name>
          (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Wong</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schmidt</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Metzen</surname>
            ,
            <given-names>J.H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kolter</surname>
            ,
            <given-names>J.Z.</given-names>
          </string-name>
          :
          <article-title>Scaling provable adversarial defenses</article-title>
          .
          <source>In: NeurIPS</source>
          . pp.
          <volume>8400</volume>
          {
          <issue>8409</issue>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>