<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>An Adversarial Attacker for Neural Networks in Regression Problems</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Kavya Gupta</string-name>
          <email>kavya.gupta100@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Beatrice Pesquet-Popescu</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Fateh Kaakai</string-name>
          <email>fateh.kaakai.e@thalesdigital.io</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jean-Christophe Pesquet</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Fragkiskos D. Malliaros</string-name>
          <email>fragkiskos.malliarosg@centralesupelec.fr</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Air Mobility Solutions BL</institution>
          ,
          <addr-line>Thales LAS</addr-line>
          <country country="FR">France</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Universite ́ Paris-Saclay, CentraleSupe ́lec, Inria Centre de Vision Nume ́rique</institution>
          ,
          <addr-line>Gif-sur-Yvette</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Adversarial attacks against neural networks and their defenses have been mostly investigated in classification scenarios. However, adversarial attacks in a regression setting remain understudied, although they play a critical role in a large portion of safety-critical applications. In this work, we present an adversarial attacker for regression tasks, derived from the algebraic properties of the Jacobian of the network. We show that our attacker successfully fools the neural network, and we measure its effectiveness in reducing the estimation performance. We present a white-box adversarial attacker to support engineers in designing safety-critical regression machine learning models. We present our results on various open-source and real industrial tabular datasets. In particular, the proposed adversarial attacker outperforms attackers based on random perturbations of the inputs. Our analysis relies on the quantification of the fooling error as well as various error metrics. A noteworthy feature of our attacker is that it allows us to optimally attack a subset of inputs, which may be helpful to analyse the sensitivity of some specific inputs.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Adversarial machine learning has received an increased
attention in the past decade. For all machine learning models,
defense against adversarial attacks is important in terms of
safety. Adversarial attacks in classification constitute
malicious attempts to trick a model classifier. They play a critical
role in real-world application domains such as spam/malware
detection, autonomous systems [Huang and Wang, 2018],
[Eykholt et al., 2018], [Ren et al., 2019], medical systems
[Finlayson et al., 2018] etc. Adversarial attacks cause
vulnerability in model deployment and specially needs to be taken
into account in deployment of security-critical AI
applications. Despite the newfound interest of the research
community in trustworthy and explainable AI, there are only few
works investigating the adversaries in the case of regression
tasks.</p>
    </sec>
    <sec id="sec-2">
      <title>Input</title>
      <p>Output</p>
      <p>Current advances in the adversarial machine learning field
evolve around the issue of designing attacks and defenses
with focus on the use of neural networks in image analysis
and computer vision [Goodfellow et al., 2014], [Kurakin et
al., 2016]. Much less works concern tabular data. However,
most machine learning tasks in the industry rely on tabular
data, e.g., fraud detection, product failure prediction,
antimoney laundering, recommendation systems, click-through
rate prediction, or flight arrival time prediction.</p>
      <p>In this paper, we focus on generating adversarial attacks
for neural networks in the specific scenario when i) a
regression task is performed and ii) tabular data are employed. Our
contributions are the following:
• We propose a simple, novel and flexible method for
generating adversarial attacks for regression tasks (a white
box attack).
• We show that the proposed attacker allows us to
optimally attack on any given subset of input features.
• We explore various error metrics which are useful for
analysing these adversarial attacks.
• Our proposed adversarial attacker is generalised for any
`p norm on input and output perturbations.
• We evaluate our results on open-source regression
datasets and an industrial dataset (output and input
features described in the Table 1) which lies in the domain
of safety critical applications.</p>
      <p>In Section 2, we give a brief overview of existing works. In
Section 3, we formulate the problem and present our method
for generating adversarial examples in regression tasks. In
Section 4, we perform numerical experiments on four datasets
to demonstrate the effectiveness of the proposed attacker.
Some concluding remarks are given in Section 5.
2</p>
      <sec id="sec-2-1">
        <title>Related Work</title>
        <p>In [Szegedy et al., 2013] the concept of adversarial attacks
was first proposed to fool DNNs. Adding a subtle
perturbation to the input of the neural network produces an
incorrect output, while human eyes cannot recognize the difference
in the modification of the input data. Even though different
models have different architectures and might use different
training data, the same kind of adversarial attack strategies
can be used to attack related models. These attacks pose a
huge threat to the performance of DNNs. [Szegedy et al.,
2013] paper proposed L-BFGS to construct adversarial
attacks and since then there has been plethora of works
introducing various adversarial attacks and their defenses for
DNNs.</p>
        <p>[Goodfellow et al., 2014] proposed a simpler and faster
method to construct adversarial attacks (FGSM). The
generated images are misclassified by adding perturbations and
linearizing the cost function in the gradient direction. This is
a non-iterative attack, hence it has a lower computation cost
than the previous method. The Fast Gradient Sign Method
(FGSM) is an `1 bounded attack and is often prone to label
leaking.</p>
        <p>It may be difficult for FGSM to control the perturbation
level in constructing attacks. [Kurakin et al., 2016] proposed
an optimized FGSM, termed Iterative Gradient Sign Method
(IGSM), which adds perturbations in multiple smaller steps
and clips the results after each iteration ensuring that the
perturbations are restricted to the neighborhood of the
example. [Dong et al., 2018] added momentum to IGSM attacks.
[Papernot et al., 2016] proposed the Jacobian-based Saliency
Map Attack (JSMA), which is based on the `0 sparsity
measure. The basic idea is to construct a saliency map with the
gradients and model the gradients based on the impact of each
pixel of the image.</p>
        <p>[Moosavi-Dezfooli et al., 2016] proposed a non-targeted
attack method based on the `2-norm, called DeepFool. It tries
to find the decision boundary that is the closest to the sample
in the image space, and then use the classification boundary to
fool the classifier. FGSM, JSMA, and DeepFool are designed
to generate adversarial attacks corresponding to single image
to fool the trained classifier model. [Moosavi-Dezfooli et al.,
2017] proposed a universal image-agnostic perturbation
attack method which fools classifier by adding a single
perturbation to all images in the dataset. [Carlini and Wagner, 2017]
proposed a powerful attack based on L-BFGS. The attack can
be generated according to `1 , `2, and `1 norm which can
be targeted or non-targeted. [Liu et al., 2016] proposed an
ensemble attack method combining multiple models to
construct adversarial attacks. [Rony et al., 2020] proposed a
method to generate minimally perturbed adversarial examples
based on Augmented Lagrangian for various distance metrics.
In [Balda et al., 2018], authors propose a general framework
for generation of adversarial examples in both classification
and regression tasks for applications in image domain.
Similar to our proposed approach, the technique is based on the
Jacobian of the neural network. Most of the methods in the
literature about adversarial example generation belong to the
class of white box attackers, i.e., the attacker has access to
the information related to the trained neural network model
including the model architecture and its parameters. A black
box attacker is introduced in [Su et al., 2019]. Such attackers
do not know the model but can interact with it. A byproduct
of black-box attack is grey-box attack where attackers might
have limited information regarding the model. To the best of
our knowledge the only work dealing with adversarial attacks
in white box settings for tabular data has been proposed in
[Ballet et al., 2019] and this work handles only classification
tasks.</p>
        <p>In regression tasks there are no natural margins as in the
case of classification tasks, and adversarial learning in
regression setting is hindered with difficulties to define the
adversarial attacks, its success, and evaluation metrics. Despite the
number of works in adversarial attack generation, there are
few articles dealing with regression tasks.[Tong et al., 2018]
looked at adversarial attacks in the setting of an ensemble of
multiple learners, investigating the interactions between these
linear learners and an attacker in regression setting, modeled
as a Multi-Learner Stackelberg Game (MLSG). However, the
investigated linear case is not able to capture the larger class
of non-linear models. The focus only on specific applications
of regression is a common. [Ghafouri et al., 2018]
examined an important problem: selecting an optimal threshold
for each sensor against an adversary for regression tasks in
cyber-physical systems. [Deng et al., 2020] introduced the
concept of adversarial threshold which is related to a
deviation between the original prediction and the prediction of
adversarial example, i.e., an acceptable error range in
driving models. In a regression context, [Nguyen and Raff, 2018]
introduced a defense that is generically useful to reduce the
effectiveness of adversarial attacks. They consider
adversarial attacks as a potential symptom of numerical instability in
the learned function. In the next section, we propose a
general white-box adversarial attacker based on Jacobian of the
learned function for regression tasks in tabular data domain.
3
3.1</p>
      </sec>
      <sec id="sec-2-2">
        <title>Proposed Method</title>
        <sec id="sec-2-2-1">
          <title>Objective</title>
          <p>The problem of adversarial attacks is closely related to the
robustness issue for a neural network, i.e. its sensitivity to
perturbations. Let T : RN0 ! RNm be the considered neural
network having N0 scalar inputs and Nm scalar outputs. If
x 2 RN0 is a given vector of inputs for some data for which
y is the associated target output, the network has been trained
to produce an output T (x) close to y. If the input is now
perturbed by an additive vector e 2 RN0 , the perturbed output
is T (x + e). Attacking the network then amounts to finding
a perturbation e of preset magnitude which makes the output
of the network to maximally deviate from a reference output.
This reference output may be the model output T (x) or the
ground truth output y. Since our purpose is to develop an
approach which remains efficient even if the accuracy of the
network is not very high, we choose y as the reference
output when available. In this context, the measures of deviation
and of magnitude of the perturbation play an important role in
terms of mathematical formulation of the problem. As a
standard choice, the measure of perturbation magnitude will be
here an `p-norm where p 2 [1; +1]. For measuring the
output deviation, we will similarly consider an `q-norm where
q 2 [1; +1]. It must be emphasized that this choice makes
sense when dealing with regression problems. In this context,
the `2 or the `1 norms are indeed frequently used as loss
functions for training. On the other hand, the `+1 norm is also a
popular measure when dealing with reliability issues.</p>
        </sec>
        <sec id="sec-2-2-2">
          <title>3.2 Optimization formulation</title>
          <p>In the described setting, the design of the attacker can be
formulated as the problem of finding the “worst pertubation” e
such that b
eb 2 Argmax kT (x + e)
e2Cp;
ykq;
where Cp; is the closed and convex set defined as
Cp; = fe 2 RN0 j k
1=2ekp
g:
2 RN0 N0 is symmetric positive definite matrix. is a
parameter which controls the maximum allowed perturbation
and is a weighting matrix typically corresponding to the
covariance matrix of the inputs. For instance, if we assume that
it is a diagonal matrix, it simply introduces a normalization
of the perturbation components with respect to the standard
deviations of the associated inputs.</p>
          <p>For standard choices of activation functions, T is a
continuous function. By virtue of Weierstrass theorem, the existence
of a solution (not necessarily unique) to Problem (1) is then
ensured. Although Cp; is a relatively simple convex set, this
problem appears as a difficult non-convex problem due to the
fact that i) T is a complex nonlinear operator, ii) we
maximize an `q measure which, in addition, leads to a nonsmooth
cost function when q = 1 or q = +1. A further difficulty is
that we usually need to attack a large dataset to evaluate the
robustness of a network and the provided optimization
algorithm should therefore be fast.</p>
        </sec>
        <sec id="sec-2-2-3">
          <title>3.3 Algorithm</title>
          <p>We propose to implement a two-step approach.</p>
          <p>• Step 1. We first perform a linearization based on the
following first-order Taylor expansion:</p>
          <p>T (x + e) ' T (x) + J (x)e;
(3)
where J (x) 2 RNm N0 is the Jacobian of the network
at x1. Note that J (x) can be computed by classical
back1We assume that J (x) is defined at x, see [Bolte and Pauwels,
2020] for a justification of this assumption in the nonsmooth case.
(1)
(2)
propagation techniques. We will make a second
approximation, that is y ' T (x). Based on these two
approximations and after the variable change e0 = 1 1=2e,
Problem (1) simplifies to
minimize kJ (x) 1=2e0kq;</p>
          <p>e02Bp
where Bp is the closed `p ball centered at 0 and with
unit radius. Note that the optimal cost value in (4) is the
subordinate norm of matrix J (x) 1=2 when the input
space is equipped with the `p norm and the output space
with the `q one. We recall that this subordinate norm is
defined, for every matrix M 2 RNm N0 , as
kM kp;q =</p>
          <p>sup
z2RN0 nf0g kzkp
kM zkq :
(4)
(5)
Problem (4) is thus equivalent to find a vector e0
b
for which the value of the cost function is equal to
kJ (x) 1=2kp;q. For values of (p; q) listed below the
expression of such vector has an explicit form.</p>
          <p>– If p = q = 2, e0 is any unit `2 norm eigenvector
b
of 1=2J (x)&gt;J (x) 1=2 associated with the
maximum eigenvalue of this matrix. This vector can be
computed by perfoming a singular value
decomposition of J (x) 1=2.
– If p = 2 and q = +1, e0 is any unit `2 norm vector
b
colinear with a row of J (x) 1=2 having maximum
`2 norm.
– If p = +1 and q = +1, e0 is a unit norm vector
b
whose elements are equal to ( (i))1 i N0 where,
for every i 2 f1; : : : ; N0g, i 2 f 1; 0; 1g is the
sign of the i-th element of a row of J (x) 1=2 with
maximum `1 norm.
– If p = 1 and q = 1, e0 is a vector which has
only one nonzero compobnent equal to 1, the
index of this component corresponds to the column
of J (x) 1=2 with maximum `1 norm.
– If p = 1 and q = 2, e0 is a vector with only one
nonzero component eqbual to 1. The index of this
component corresponds to a column of J (x) 1=2
with maximum `2 norm.
– If p = 1 and q = +1, e0 is again a vector with
only one nonzero componbent equal to 1. The
index of this component corresponds to a column of
J (x) 1=2 where is located an element of maximum
absolute value.
• Step 2. In the previous optimization step, the optimal
soltuotiPornobislenmot(u4n),iqthueen. Indebeiesdailfsoeb =asolu1t=io2neb.0 iIsnaasdodliuttiioonn,
there may exist other reasons for the multiplicity of the
solutions. For example, there may be several maximum
norm rows for matrix J (x) 1=2. Among all the possible
choices, we propose to choose the solution e leading to
b
the maximum deviation w.r.t. the ground truth, that is
such that kT (x + eb) ykq is maximum. This requires
to perform a search on a small number of possible
candidates. Note that no approximation error is involved in
this step. If the ground truth for the output is not
available, it can be replaced by the model output.
• Post-optimization. If 1 &lt; q &lt; +1 and T is assumed
to be differentiable, e 7! kT (x + e) ykqq is a
differentiable function. A further refinement consists of
minimizing this function over Cp; by using a projected
gradient algorithm with Armijo search for the stepsize.
The previous estimates of e can then be used to initialize
b
the algorithm. According to our numerical tests,
implementing this strategy when q = 2 only brings a marginal
improvement. Moreover, this approach cannot be used
when q = 1 or q = +1.
3.4</p>
        </sec>
        <sec id="sec-2-2-4">
          <title>Attacking a group of inputs</title>
          <p>It can also be interesting to attack only a selected subset of
inputs. It may help in identifying the more sensitive inputs
of the network. Also, for some inputs like unsorted
categorical ones, attacks are often meaningless since they
introduce a main change in the informative contents of the dataset,
which can be easily detected. Our proposed approach can be
adapted to generate such partial attacks. In Problem (4), it is
indeed sufficient to replace matrix 1=2 by D 1=2D, where
D a masking diagonal matrix whose diagonal elements are
equal to 1 when the input is attacked and 0 otherwise. The
optimal solutions eb0 and eb = D 1=2Deb0 = D 1=2eb0 have
then their components equal to 0 for the non-attacked inputs.
Note that the naive approach which would consist in solving
(4) and setting to zero the resulting perturbation components
for non-attacked inputs would be suboptimal.
4
4.1</p>
        </sec>
      </sec>
      <sec id="sec-2-3">
        <title>Numerical Results</title>
        <sec id="sec-2-3-1">
          <title>Dataset and architecture description</title>
        </sec>
        <sec id="sec-2-3-2">
          <title>Open Source Datasets</title>
          <p>We run our experiments on three open source regression
datasets. The Combined Cycle Power Plant [T u¨fekci, 2014]
dataset has 4 features with 9,568 instances. The task is to
predict the net hourly electrical energy output using hourly
average ambient variables. The Red Wine Quality dataset [Cortez
et al., 2009] contains 1,599 total samples and each instance
has 11 features. The features are physicochemical and
sensory measurements for wine. The output variable is a quality
score ranging from 0 to 10, where 10 represents for best
quality and 0 for least quality. For the Abalone dataset, the task
is to model an Abalone’s age based purely on its physical
measurements. This would allow Abalone’s age estimation
without cutting its shell. There are in total 4,177 instances
with 8 input variables including one categorical variable. The
datasets are divided with a ratio of 4:1 between training and
testing data. The categorical attributes are dealt with by
using one hot encoding based on the number of categories. The
input attributes are normalised by removing their mean and
scaling to unit variance.</p>
          <p>We train fully connected networks for the estimation of
variables from the datasets. The network architecture for the
dataset are given below. The values represent the number
of hidden neurons in the layers. Activation function at each
layer is ReLU except for the last layer.</p>
          <p>• Combined cycle Power Plant dataset - (10; 6; 1)
• Red Wine Quality dataset - (100; 100; 100; 10; 1)
• Abalone Data set - (256; 256; 256; 256; 1)</p>
        </sec>
        <sec id="sec-2-3-3">
          <title>Industrial Dataset – safety critical Application</title>
          <p>An industrial application dataset is also considered with
2,219,097 training, 739,639 validation, and 739,891 test
samples. The description of the input/output variables of the</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Fooling Error</title>
      <p>E
Symmetric Mean Accuracy Percentage Error
SMAPE
=
=
=
1 XK kT (xk + ek)
K</p>
      <p>k=1
1 XK kT (xk + ek)
K
k=1</p>
      <p>ykkq</p>
      <p>T (xk)kq
2</p>
      <p>K+</p>
      <p>X kT (xk + ek)
K+ k=1 kT (xk + ek)
ykkq kT (xk)
ykkq + kT (xk)
ykkq
ykkq</p>
      <p>Noise MAEstd MAEgauss MAEuni MAEbin MAE adCv ombinedECgyaucslse Power PlEanuntiDataset Ebin Eadv
12 1100 11 66::44 1100 33 66::58 1100 33 66::58 1100 33 66::59 1100 33 1104::32 1100 33 12::35 1100 33 12::35 1100 33 12::47 1100 33 48::00 1100 33
Red Wine Quality Dataset
12 1100 11 00..4477 00..4467 00..4478 00..4478 00..5686 00..0049 00..0059 00..00493 00..1221
Abalone age dataset
51 1100 21 11..6688 11..6688 11..6688 11..6688 22..4004 00..0052 00..0052 00..0053 00..7326
Industrial Dataset
12 1100 11 99::22 1100 33 . 91:06:7 101033 190::67 1100 33 190::67 1100 33 2302::95 1100 33 25::61 1100 33 52::26 1100 33 25::74 1100 33 1214::80 1100 33
dataset is given in Table 1. The variable to be predicted is the
Estimation of Arrival time (ETE) of a flight, given variables
including the distance and speed, and also an initial estimate
of ETE. The dataset is related to flight control, an activity
area where safety is critical. The input attributes are
normalized by removing their mean and scaling to unit variance. For
models, we build fully connected networks with ReLU
activation function on all the hidden layers except the last one.</p>
      <p>The network architecture is shown in the Figure 1.
4.2</p>
      <sec id="sec-3-1">
        <title>Experimental setup</title>
        <p>We first train our networks without any constraints using the
network architecture presented in the previous section with
the aim of reducing the prediction/performance loss on the
train dataset. This will be refered to as a standard training
procedure.</p>
        <p>To understand and analyze the performance of the
proposed adversarial attacker, we calculate the three error
metrics described in Table 2. We compare the proposed
adversarial attacker with random noise attackers generated by i.i.d.
perturbations. We use three additive noise distributions—
Gaussian, uniform and binary, for comparisons. The output
of these attackers have been normalized so as to meet the
desired bound on the norm of the perturbation. The metrics are
computed on the test samples where K is the total number
of samples in the test set. The results on the 4 datasets for
varying noise levels are shown in Table 3. We also show the
histograms of (kT (xk+ek) ykkq kT (xk) ykkq)1 k K in
Figures 2, 3, 4, and 5, where (ek)1 k K have been generated
from various noise distributions and the proposed adversarial
attacker.</p>
        <p>For safety critical tasks, Lipschitz and performance targets
can be specified as engineering requirements, prior to
network training. Such a design approach has proven to make
the network more stable and robust to adversarial attacks.
Imposing a Lipschitz target can be done either by controlling the
Lipschitz constant for each layer or for the whole network
depending on the application at hand. One such method for
controlling the Lipschitz has been presented in [Serrurier et
al., 2020] using Hinge regularization. In the experiments, we
train our networks while using a spectral normalisation
technique [Miyato et al., 2018] which has been proven to be very
effective in controlling Lipschitz properties in GANs.</p>
        <p>Given an m layer fully connected architecture and a
Lipschitz target L, we can constrain the spectral norm of each
p
layer to be less than m L. This ensures that the upper bound
on the global Lipschitz constant is less than L. We keep the
network architectures exactly the same for both training
procedures. The performance of adversarial attacker on standard
and spectrally normalized trained model in terms of Fooling
Error (E) and Symmetric Mean Accuracy Percentage error
(SMAPE) for various datasets and varying perturbation
magnitude is given in Table 4.</p>
        <p>All the previous results have been obtained with attack and
noise addition on all the input features present in the datasets.
As pointed in Section 3.4, the introduced adversarial attacker
is capable of attacking a group of inputs. While generating
an adversarial attack we avoid attacking the categorical input
variables [Ballet et al., 2019], hence in Abalone and
industrial datasets we attack only the continuous variables. For the
Combined Power plant dataset, we attack 3 out of 4
continuous variables since it does not contain any categorical
variables. Similarly, for the Red-wine dataset we attack 8
continuous variables out of 11. The performance of the adversarial
attacker, when attacking only few inputs, is shown in Table 5.</p>
        <p>As emphasized in Section 3.3, our adversarial attacker is
applicable for various measures of perturbation on input and
output deviations. The previous results have been obtained
for the value p = q = 2 termed as `2 attacks here. We further
show results for p = q = 1 termed as `1 attacks and for
p = q = +1 termed as `1 attacks in Table 6.</p>
      </sec>
      <sec id="sec-3-2">
        <title>4.3 Result analysis</title>
        <p>Some general conclusions can be drawn from the
experiments.</p>
        <p>• We observe that the proposed adversarial attacker
performs better than all the three random noise attackers
for the three quantitative measures we have defined. In
addition, the histograms in Figures 2, 3, 4, and 5 show
that the error may be increased or reduced by random
attackers, while this shortcoming does not happen with
our adversarial attacker. This observation is verified on
the norms - `2 , `1 and `1 norms in Table 6.
• Spectral normalisation has been proven to robustify the
trained models. As in Table 4, we see that the Fooling
error (E) and SMAPE are reduced in all the cases when
compared to the standard trained model.
• In the considered examples, we observe that categorical
data have little effect when attacking the trained model
as shown in Table 5. The E and SMAPE measures do
not show major differences.
5</p>
        <sec id="sec-3-2-1">
          <title>Conclusion</title>
          <p>In this article, we have introduced a novel easily
implementable Jacobian-based adversarial attacker for estimation
problems. These regression tasks cover a major portion of
safety critical applications. Yet there is lack of works
studying and analysing adversarial attacks in this context, as
opposed to classification tasks. The present study contributes
to filling this gap. We have presented error metrics which
help in analysing the effectiveness of the attacker. Our
attacker is versatile in the sense that it can handle any measure
(`1, `2, `1) on input or output perturbations according to the
target application. Our attacker is also successful in handling
attacks focused on subsets of inputs. This feature may be
useful when handling specific tabular datasets and may also be
insightful when information is available related to the
sensitivity or ability to control some inputs. Our tests concentrated
on fully connected networks, but it is worth pointing out that
the proposed approach can be applied to any network
architecture.
0.0050.000 0.005 0.010 0.015 0.020 0.025</p>
          <p>Error
Gaussian</p>
          <p>Uniform
Adversarial
140
120
100
tn 80
ou
C
tonu
C
60
40
20
0
70
60
50
40
30
20
10
200
150
toun
C
100
50
0
70000
60000
50000
toun
C
40000
30000
20000
10000
0
toun
C
40000
30000
20000
10000
0
attack.
attack.
0.02 0.01 0.00 0.01 0.02 0.03 0.04 0.05</p>
          <p>Error
0.02 0.01 0.00 0.01 0.02 0.03 0.04 0.05</p>
          <p>Error
0.02 0.01 0.00 0.01 0.02 0.03 0.04 0.05</p>
          <p>Error
0.02 0.01 0.00 0.01 0.02 0.03 0.04 0.05</p>
          <p>Error</p>
        </sec>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [Balda et al.,
          <year>2018</year>
          ]
          <string-name>
            <given-names>Emilio</given-names>
            <surname>Rafael</surname>
          </string-name>
          <string-name>
            <surname>Balda</surname>
          </string-name>
          , Arash Behboodi, and
          <string-name>
            <given-names>Rudolf</given-names>
            <surname>Mathar</surname>
          </string-name>
          .
          <article-title>Perturbation analysis of learning algorithms: A unifying perspective on generation of adversarial examples</article-title>
          .
          <source>arXiv preprint arXiv:1812.07385</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [Ballet et al.,
          <year>2019</year>
          ]
          <string-name>
            <given-names>Vincent</given-names>
            <surname>Ballet</surname>
          </string-name>
          , Xavier Renard, Jonathan Aigrain, Thibault Laugel, Pascal Frossard, and
          <string-name>
            <given-names>Marcin</given-names>
            <surname>Detyniecki</surname>
          </string-name>
          .
          <article-title>Imperceptible adversarial attacks on tabular data</article-title>
          .
          <source>arXiv preprint arXiv:1911.03274</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <source>[Bolte and Pauwels</source>
          , 2020]
          <article-title>Je´roˆme Bolte and Edouard Pauwels. Conservative set valued fields, automatic differentiation, stochastic gradient methods and deep learning</article-title>
          .
          <source>Mathematical Programming</source>
          , pages
          <fpage>1</fpage>
          -
          <lpage>33</lpage>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <source>[Carlini and Wagner</source>
          , 2017]
          <string-name>
            <given-names>Nicholas</given-names>
            <surname>Carlini</surname>
          </string-name>
          and
          <string-name>
            <given-names>David</given-names>
            <surname>Wagner</surname>
          </string-name>
          .
          <article-title>Towards evaluating the robustness of neural networks</article-title>
          .
          <source>In 2017 ieee symposium on security and privacy (sp)</source>
          , pages
          <fpage>39</fpage>
          -
          <lpage>57</lpage>
          . IEEE,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [Cortez et al.,
          <year>2009</year>
          ]
          <string-name>
            <given-names>Paulo</given-names>
            <surname>Cortez</surname>
          </string-name>
          , Anto´nio Cerdeira, Fernando Almeida, Telmo Matos, and Jose´ Reis.
          <article-title>Modeling wine preferences by data mining from physicochemical properties</article-title>
          .
          <source>Decision support systems</source>
          ,
          <volume>47</volume>
          (
          <issue>4</issue>
          ):
          <fpage>547</fpage>
          -
          <lpage>553</lpage>
          ,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [Deng et al.,
          <year>2020</year>
          ]
          <string-name>
            <given-names>Yao</given-names>
            <surname>Deng</surname>
          </string-name>
          , Xi Zheng, Tianyi Zhang, Chen Chen, Guannan Lou, and
          <string-name>
            <given-names>Miryung</given-names>
            <surname>Kim</surname>
          </string-name>
          .
          <article-title>An analysis of adversarial attacks and defenses on autonomous driving models</article-title>
          .
          <source>In 2020 IEEE International Conference on Pervasive Computing and Communications (PerCom)</source>
          , pages
          <fpage>1</fpage>
          -
          <lpage>10</lpage>
          . IEEE,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [Dong et al.,
          <year>2018</year>
          ]
          <string-name>
            <given-names>Yinpeng</given-names>
            <surname>Dong</surname>
          </string-name>
          , Fangzhou Liao, Tianyu Pang, Hang Su, Jun Zhu, Xiaolin Hu, and
          <string-name>
            <given-names>Jianguo</given-names>
            <surname>Li</surname>
          </string-name>
          .
          <article-title>Boosting adversarial attacks with momentum</article-title>
          .
          <source>In Proceedings of the IEEE conference on computer vision and pattern recognition</source>
          , pages
          <fpage>9185</fpage>
          -
          <lpage>9193</lpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [Eykholt et al.,
          <year>2018</year>
          ]
          <string-name>
            <given-names>Kevin</given-names>
            <surname>Eykholt</surname>
          </string-name>
          , Ivan Evtimov, Earlence Fernandes,
          <string-name>
            <given-names>Bo</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Amir</given-names>
            <surname>Rahmati</surname>
          </string-name>
          , Chaowei Xiao, Atul Prakash, Tadayoshi Kohno, and
          <string-name>
            <given-names>Dawn</given-names>
            <surname>Song</surname>
          </string-name>
          .
          <article-title>Robust physical-world attacks on deep learning visual classification</article-title>
          .
          <source>In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition</source>
          , pages
          <fpage>1625</fpage>
          -
          <lpage>1634</lpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [Finlayson et al.,
          <year>2018</year>
          ] Samuel G Finlayson,
          <article-title>Hyung Won Chung, Isaac S Kohane,</article-title>
          and Andrew L Beam.
          <article-title>Adversarial attacks against medical deep learning systems</article-title>
          . arXiv preprint arXiv:
          <year>1804</year>
          .05296,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [Ghafouri et al.,
          <year>2018</year>
          ]
          <string-name>
            <given-names>Amin</given-names>
            <surname>Ghafouri</surname>
          </string-name>
          , Yevgeniy Vorobeychik, and
          <string-name>
            <given-names>Xenofon</given-names>
            <surname>Koutsoukos</surname>
          </string-name>
          .
          <article-title>Adversarial regression for detecting attacks in cyber-physical systems</article-title>
          . arXiv preprint arXiv:
          <year>1804</year>
          .11022,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [Goodfellow et al.,
          <year>2014</year>
          ] Ian J Goodfellow, Jonathon Shlens, and
          <string-name>
            <given-names>Christian</given-names>
            <surname>Szegedy</surname>
          </string-name>
          .
          <article-title>Explaining and harnessing adversarial examples</article-title>
          .
          <source>arXiv preprint arXiv:1412.6572</source>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <source>[Huang and Wang</source>
          , 2018]
          <string-name>
            <given-names>Yonghong</given-names>
            <surname>Huang</surname>
          </string-name>
          and Shih-han
          <string-name>
            <surname>Wang</surname>
          </string-name>
          .
          <article-title>Adversarial manipulation of reinforcement learning policies in autonomous agents</article-title>
          .
          <source>In 2018 International Joint Conference on Neural Networks (IJCNN)</source>
          , pages
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          . IEEE,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [Kurakin et al.,
          <year>2016</year>
          ]
          <string-name>
            <given-names>Alexey</given-names>
            <surname>Kurakin</surname>
          </string-name>
          , Ian Goodfellow,
          <string-name>
            <given-names>Samy</given-names>
            <surname>Bengio</surname>
          </string-name>
          , et al.
          <article-title>Adversarial examples in the physical</article-title>
          world,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [Liu et al.,
          <year>2016</year>
          ] Yanpei Liu,
          <string-name>
            <given-names>Xinyun</given-names>
            <surname>Chen</surname>
          </string-name>
          , Chang Liu, and
          <string-name>
            <given-names>Dawn</given-names>
            <surname>Song</surname>
          </string-name>
          .
          <article-title>Delving into transferable adversarial examples and black-box attacks</article-title>
          .
          <source>arXiv preprint arXiv:1611.02770</source>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [Miyato et al.,
          <year>2018</year>
          ]
          <string-name>
            <given-names>Takeru</given-names>
            <surname>Miyato</surname>
          </string-name>
          , Toshiki Kataoka, Masanori Koyama, and
          <string-name>
            <given-names>Yuichi</given-names>
            <surname>Yoshida</surname>
          </string-name>
          .
          <article-title>Spectral normalization for generative adversarial networks</article-title>
          .
          <source>arXiv preprint arXiv:1802.05957</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [
          <string-name>
            <surname>Moosavi-Dezfooli</surname>
          </string-name>
          et al.,
          <year>2016</year>
          ]
          <string-name>
            <surname>Seyed-Mohsen</surname>
            <given-names>MoosaviDezfooli</given-names>
          </string-name>
          , Alhussein Fawzi, and
          <string-name>
            <given-names>Pascal</given-names>
            <surname>Frossard</surname>
          </string-name>
          .
          <article-title>Deepfool: a simple and accurate method to fool deep neural networks</article-title>
          .
          <source>In Proceedings of the IEEE conference on computer vision and pattern recognition</source>
          , pages
          <fpage>2574</fpage>
          -
          <lpage>2582</lpage>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [
          <string-name>
            <surname>Moosavi-Dezfooli</surname>
          </string-name>
          et al.,
          <year>2017</year>
          ]
          <string-name>
            <surname>Seyed-Mohsen</surname>
            <given-names>MoosaviDezfooli</given-names>
          </string-name>
          , Alhussein Fawzi, Omar Fawzi, and
          <string-name>
            <given-names>Pascal</given-names>
            <surname>Frossard</surname>
          </string-name>
          .
          <article-title>Universal adversarial perturbations</article-title>
          .
          <source>In Proceedings of the IEEE conference on computer vision and pattern recognition</source>
          , pages
          <fpage>1765</fpage>
          -
          <lpage>1773</lpage>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          <source>[Nguyen and Raff</source>
          , 2018] Andre T Nguyen and
          <string-name>
            <given-names>Edward</given-names>
            <surname>Raff</surname>
          </string-name>
          .
          <article-title>Adversarial attacks, regression, and numerical stability regularization</article-title>
          .
          <source>arXiv preprint arXiv:1812.02885</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [Papernot et al.,
          <year>2016</year>
          ]
          <string-name>
            <given-names>Nicolas</given-names>
            <surname>Papernot</surname>
          </string-name>
          ,
          <string-name>
            <surname>Patrick</surname>
            <given-names>McDaniel</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Somesh</given-names>
            <surname>Jha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Matt</given-names>
            <surname>Fredrikson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z Berkay</given-names>
            <surname>Celik</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Ananthram</given-names>
            <surname>Swami</surname>
          </string-name>
          .
          <article-title>The limitations of deep learning in adversarial settings</article-title>
          .
          <source>In 2016 IEEE European symposium on security and privacy (EuroS&amp;P)</source>
          , pages
          <fpage>372</fpage>
          -
          <lpage>387</lpage>
          . IEEE,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [Ren et al.,
          <year>2019</year>
          ]
          <string-name>
            <given-names>Kui</given-names>
            <surname>Ren</surname>
          </string-name>
          , Qian Wang,
          <string-name>
            <surname>Cong</surname>
            <given-names>Wang</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhan Qin</surname>
            , and
            <given-names>Xiaodong</given-names>
          </string-name>
          <string-name>
            <surname>Lin</surname>
          </string-name>
          .
          <article-title>The security of autonomous driving: Threats, defenses, and future directions</article-title>
          .
          <source>Proceedings of the IEEE</source>
          ,
          <volume>108</volume>
          (
          <issue>2</issue>
          ):
          <fpage>357</fpage>
          -
          <lpage>372</lpage>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [Rony et al.,
          <year>2020</year>
          ] Je´roˆme Rony, Eric Granger, Marco Pedersoli, and Ismail Ben Ayed.
          <article-title>Augmented lagrangian adversarial attacks</article-title>
          .
          <source>arXiv preprint arXiv:2011.11857</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [Serrurier et al.,
          <year>2020</year>
          ]
          <string-name>
            <given-names>Mathieu</given-names>
            <surname>Serrurier</surname>
          </string-name>
          , Franck Mamalet, Alberto Gonza´
          <fpage>lez</fpage>
          -Sanz, Thibaut Boissin,
          <string-name>
            <surname>Jean-Michel Loubes</surname>
          </string-name>
          ,
          <source>and Eustasio del Barrio</source>
          .
          <article-title>Achieving robustness in classification using optimal transport with hinge regularization</article-title>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [Su et al.,
          <year>2019</year>
          ]
          <string-name>
            <given-names>Jiawei</given-names>
            <surname>Su</surname>
          </string-name>
          , Danilo Vasconcellos Vargas, and
          <string-name>
            <given-names>Kouichi</given-names>
            <surname>Sakurai</surname>
          </string-name>
          .
          <article-title>One pixel attack for fooling deep neural networks</article-title>
          .
          <source>IEEE Transactions on Evolutionary Computation</source>
          ,
          <volume>23</volume>
          (
          <issue>5</issue>
          ):
          <fpage>828</fpage>
          -
          <lpage>841</lpage>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [Szegedy et al.,
          <year>2013</year>
          ]
          <string-name>
            <given-names>Christian</given-names>
            <surname>Szegedy</surname>
          </string-name>
          , Wojciech Zaremba, Ilya Sutskever, Joan Bruna, Dumitru Erhan, Ian Goodfellow, and
          <string-name>
            <given-names>Rob</given-names>
            <surname>Fergus</surname>
          </string-name>
          .
          <article-title>Intriguing properties of neural networks</article-title>
          .
          <source>arXiv preprint arXiv:1312.6199</source>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [Tong et al.,
          <year>2018</year>
          ]
          <string-name>
            <given-names>Liang</given-names>
            <surname>Tong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Sixie</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Scott</given-names>
            <surname>Alfeld</surname>
          </string-name>
          , et al.
          <article-title>Adversarial regression with multiple learners</article-title>
          .
          <source>In International Conference on Machine Learning</source>
          , pages
          <fpage>4946</fpage>
          -
          <lpage>4954</lpage>
          . PMLR,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [Tu¨fekci, 2014]
          <article-title>Pınar Tu¨fekci. Prediction of full load electrical power output of a base load operated combined cycle power plant using machine learning methods</article-title>
          .
          <source>International Journal of Electrical Power and Energy Systems</source>
          ,
          <volume>60</volume>
          :
          <fpage>126</fpage>
          -
          <lpage>140</lpage>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>