<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Reversible Adversarial Attack Based on Reversible Image Transformation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Zhaoxia Yin</string-name>
          <email>yinzhaoxia@ahu.edu.cn</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Hua Wang</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Li Chen</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jie Wang</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Weiming Zhang</string-name>
          <email>zhangwm@ustc.edu.cn</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Anhui Provincial Key Laboratory of Multimodal Cognitive Computation, Anhui University</institution>
          ,
          <addr-line>Hefei 230601</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>School of Information Science and Technology, University of Science and Technology of China</institution>
          ,
          <addr-line>Hefei 230026</addr-line>
        </aff>
      </contrib-group>
      <abstract>
        <p>In order to prevent illegal or unauthorized access of image data such as human faces and ensure legitimate users can use authorization-protected data, reversible adversarial attack technique is rise. Reversible adversarial examples (RAE) get both attack capability and reversibility at the same time. However, the existing technique can not meet application requirements because of serious distortion and failure of image recovery when adversarial perturbations get strong. In this paper, we take advantage of Reversible Image Transformation technique to generate RAE and achieve reversible adversarial attack. Experimental results show that proposed RAE generation scheme can ensure imperceptible image distortion and the original image can be reconstructed error-free. What's more, both the attack ability and the image quality are not limited by the perturbation amplitude.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;deep neural networks</kwd>
        <kwd>adversarial example</kwd>
        <kwd>data protection</kwd>
        <kwd>reversible image transformation</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>1. Introduction</p>
    </sec>
    <sec id="sec-2">
      <title>In order to make the research significance and technical</title>
      <p>basis of the proposed work clear, we make introduction
the following four aspects. The first is the research
background, leading to the important value of adversarial
examples with both attack capability and reversibility.
Then, the research status of adversarial attack with
adversarial examples come. After the parallels and diferences
between information hiding and adversarial examples,
reversible adversarial attacks based on information
hiding put forward. Finally, the motivation and contribution
of the proposed method is highlighted.</p>
      <sec id="sec-2-1">
        <title>1.1. Background</title>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Deep learning [1] performance is getting more and</title>
      <p>more outstanding, especially in many tasks such as
autonomous driving [2] and face recognition [3]. As an
important technique of Artificial Intelligence (AI), it has
also been challenged by diferent kinds of attacks. In 2013,
Szegedy et al. [4] first discovered that adding
perturbations that are imperceptible to human vision in an image
can mislead the neural network model to get wrong
results with high confidence. As shown in Fig. 1, This kind
of images that have been added with specific noise to
mislead a deep neural network model are called Adversarial
Examples [5], and the added noises are called Adversarial
Perturbations.</p>
      <p>As a lethal attack technology in the AI security field,
if adversarial examples are equipped with both attack
capability and reversibility, it will be undoubtedly having
important application value, i.e., attacking unauthorized
models and harmless to authorized models with lossless
recovery capability [6].</p>
      <p>Reversible adversarial attack aims to add adversarial
perturbations into images in a reversible way to
generate adversarial examples. On one hand, the generated
Reversible Adversarial Examples (RAE) can attack the
unauthorized models and prevent illegal or unauthorized
access of image data; on the other hand, authorized
intelligent system can restore the corresponding original
images from RAE completely and avoid interference safely.</p>
      <p>The emergence of RAE equips adversarial examples with
new capabilities, which is of great significance to further
expand the attack-defense technology and applications
of AI. Quiring et al. [12] analyzed the similarities and
difer</p>
      <p>However, the research has just started, and the perfor- ences between Adversarial Example and Watermarking.
mance are not satisfied. Many problems and questions, Both of them modify the target object to cross the
desuch as how to balance and optimize attack capability, cision boundary at the lowest cost. In watermarking,
reversibility and image visual quality, are still waiting to the watermarking detector is regarded as a two-classifier,
be solved and answered. and the watermarking in the signal could be destroyed
by the watermarking attacks, so that the classification
re1.2. Adversarial Attack and Adversarial sult could be changed from image-with-watermarking to
Examples image-without-watermarking. In machine learning, this
boundary separates diferent categories, and the attacked
Attacks and defenses of adversarial examples have at- signal, i.e. Adversarial Examples, will be misjudged by
tracted more and more attention from researchers in the the model. Schöttle et al. [13] analyzed the similarities
ifeld of machine learning security, and have become a and diferences between steganography and adversarial
hot research topic in recent years. Here we briefly sum- examples. Steganography attempts to modify individual
marize the current research status of adversarial attack pixel values to embed secret information, so that it is
and adversarial examples[7]. dificult for steganography analysts to detect the hidden</p>
      <p>
        Adversarial attack is to design algorithms to turn nor- information. Schöttle et al. believe that the detection of
mal samples into adversarial examples to fool AI system. adversarial examples belongs to the category of
steganalAccording to the diferent degree of attacker’s under- ysis, and develops a heuristic linear predictive
adversarstanding of the target model information, it can be di- ial detection method based on steganalysis technology.
vided into white box and black box attacks. White-box Zhang et al. [14] compared deep steganography and
attack refers to the construction of adversarial examples universal adversarial perturbation, and found that the
based on information such as the structural parameters success of both is attributed to the deep neural network’s
of the target model, Eg. Iterative Fast Gradient Sign exceptional sensitivity to high frequency content.
Method (IFGSM) [8]. Black-box attack is to construct ad- When we know these interesting cross-cutting
studversarial examples without any information of the target ies of Adversarial Example and Information Hiding, we
model and adversarial examples are usually generated would inevitably wonder, what would we get by
comby training alternative models, Eg. single pixel attack [9]. bining Adversarial Example with another Information
Further more, taking image classification as an example, Hiding technique, i.e. Reversible Data Hiding.
non-target attack only needs to make the model result in Liu et al. achieved the first Reversible Adversarial
wrong classification for a given adversarial example and attack by combining Reversible Data Hiding with
Adverusually the perturbation is relatively small. For example, sarial Examples and proposed the concept of Reversible
DeepFool attack [10]. The other kind of attack can make adversarial examples (RAE) [15]. Since RAE get both
atthe model classify a given adversarial example into a spec- tack capability and reversibility at the same time, illegal
ified category rather than any incorrect categories and or unauthorized access of image data can be prevented
the representative algorithm is well known as C&amp;W [11]. and legitimate using can be guaranteed by original image
So we can say, by slightly modifying the input digital recovery. As shown in Fig. 2, Reversible Data Embedding
image signal, adversarial example are generated to show (RDE) technique [16] is adopted to embed the
adversardiferent information to machine or intelligent system. ial perturbations into its adversarial image to get the
But for human vision, the information and content of the reversible adversarial example image, from which, the
image have not been changed. original image can be restored error-free. The
framework consists of three steps: (
        <xref ref-type="bibr" rid="ref2">1</xref>
        ) adversarial examples
1.3. Reversible Adversarial Examples generation; (
        <xref ref-type="bibr" rid="ref1 ref3">2</xref>
        ) reversible adversarial examples
generation by reversible data embedding; (
        <xref ref-type="bibr" rid="ref4">3</xref>
        ) original images
So we can say, by slightly modifying the input digital recovery. This is really a great creative work even though
image signal, adversarial example are generated to show the performance is far from satisfied. Let’s call it
RDEdiferent information to machine or intelligent system. based RAE method and discuss the details in the coming
But for human vision, the information and content of the section.
image have not been changed. Actually, there is another
similar technique that also aims to achieve some special 1.4. Motivation and Contribution
goals by slightly modifying the input digital image signal,
called Information Hiding, which consists of diferent
research topics such as Watermarking, Steganography
and Reversible Data Hiding (RDH) [6].
      </p>
    </sec>
    <sec id="sec-4">
      <title>As mentioned above, to obtain RAE, Liu et al. adopted Reversible Data Embedding technique to embed the adversarial perturbations into its adversarial image, then the original image can be restored without distortion.</title>
      <p>RDH (Reversible Image Transformation)</p>
      <p>Adversarial</p>
      <p>Examples
(e
R
ilrevb sseL
se lo
Im ssR
aeg ec
R vo
eco rey
r)evy</p>
      <p>Original Images</p>
      <p>Reversible
Adversarial
Examples
Protected Resources</p>
      <p>Original Images
3 Original Image Restoration</p>
      <p>Adversarial
Perturbations</p>
      <p>Irreversible
1 Adversarial Example Generation</p>
      <p>Unauthorized</p>
      <p>Models
Authorized
Models
Original Images</p>
      <p>Adversarial
Perturbations
(e
R
irebv sseL
s o
le l
D ss
taaR ecoR
ceeov revy
r)y</p>
      <p>RDH (Reversible Data EmBedding)</p>
      <p>Reversible
Adversarial
Examples
Protected Resources</p>
      <p>Original Images
3 Original Image Restoration</p>
      <p>Unauthorized</p>
      <p>Models
Authorized</p>
      <p>Models</p>
      <p>However, no matter which kind of RDE algorithm is Figure 3: The overall framework of the proposed RIT-based
adopted, the embedding capacity is always limited. That RAE method.
means, the maximum amount of the embedding data that
can be carried by the adversarial image is also limited.</p>
      <p>
        Therefore, when adversarial perturbations are strength- 2. The Proposed Method
ened, the amount of data that needs to be embedded
increases, that would result in the following three prob- In order to achieve reversible adversarial attack, we
prolems: (
        <xref ref-type="bibr" rid="ref2">1</xref>
        ) The generated adversarial perturbations cannot pose a more efective method to generate reversible
adbe embedded completely and then the original image versarial examples. As shown in Fig. 3, we replace
recannot be restored completely , which leads to the fail- versible data hiding with RIT strategy to obtain RAE.
ure of reversibility; (
        <xref ref-type="bibr" rid="ref1 ref3">2</xref>
        ) Since too much data has to be The original image restoration process is the inverse
embedded, the reversible adversarial image is severely process of RIT, i.e., reversible image recovery. In this
distorted, which leads to unsatisfied image quality; (
        <xref ref-type="bibr" rid="ref4">3</xref>
        ) section, We describe the implementation of our method
Due to increased distortion of RAE, the attack ability as three steps: (
        <xref ref-type="bibr" rid="ref2">1</xref>
        ) Adversarial examples generation; (
        <xref ref-type="bibr" rid="ref1 ref3">2</xref>
        )
decreases accordingly. Reversible adversarial examples generation; (
        <xref ref-type="bibr" rid="ref4">3</xref>
        ) Original
      </p>
      <p>
        To solve these problems, here we propose to replace image restoration.
the idea of Reversible Data Embedding with Reversible
Image Transformation (RIT) technique. In order to ver- 2.1. Adversarial Examples Generation
ify the efectiveness of the strategy, we chose one RIT
method [17] as an example to construct RAE and make Firstly, we need to generate adversarial examples for step
performance comparisons with the method from [15]. (
        <xref ref-type="bibr" rid="ref1 ref3">2</xref>
        ). Since adversarial attacks are mainly divided into
Experiments show that the proposed scheme can com- white box and black box. White box attack algorithms
pletely solve the problems that analyzed above. Further- have better performance, and black box attacks usually
more, in the proposed method, realization of reversibility rely on white box attacks indirectly, so this paper
generdoes not depend on embedding the signal diference be- ates adversarial examples based on white-box settings.
tween original images and adversarial examples, i.e., it is Next, we introduce several state-of-the-art white box
not limited to the strength of adversarial perturbations. attack algorithms.
      </p>
      <p>As well-known, the greater the adversarial perturbation,
the stronger the attack ability. Therefore, the proposed
method can achieve better RAE performance in terms
of reversibility, image quality and attack capability. We
name it RIT-based RAE method and describe it step by
step in Section 2. Details of experiments and results are
given in Section 3, following with Conclusion in Section
4.
• IFGSM [8] proposed as an iterative version of</p>
      <p>FGSM [5]. It is a quick way to generate
adversarial examples, applies FGSM multiple times with
small perturbation instead of adding a large
perturbation.
• DeepFool[10] is a untargeted attack algorithm
that generates adversarial examples by
exploring the nearest decision boundary, the image is
slightly modified in each iteration to reach the
boundary, and the algorithm will not stop
until the modified image changes the classification
result.
• C&amp;W [11] is an optimization-based attack that
makes perturbation undetectable by limiting the
0, 2, ∞ norms.</p>
      <sec id="sec-4-1">
        <title>2.2. Reversible Adversarial Examples</title>
      </sec>
      <sec id="sec-4-2">
        <title>Generation</title>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Secondly, we take Reversible Image Transformation (RIT)</title>
      <p>algorithm to generate protected resources with restricted
access capabilities, i.e., reversible adversarial examples.
Specifically, we take the adversarial example as the
target image, and use RIT to disguise original image as the
adversarial example to directly get reversible adversarial
example. Next, we will introduce the RIT algorithm in
detail. In fact, RIT algorithm is also a kind of reversible
data hiding technique to achieve image content
protection. It can reversibly transform an original image into an
arbitrarily-chosen target image with the same size to get
a camouflage image, which looks almost
indistinguishable from the target image. When the diference between
the two images is smaller, the amount of auxiliary
information for restoring original image is greatly reduced,
that makes it perfect for RAE since the diference
between an original image and its Adversarial Example is
usually very small.
2.2.1. Algorithm Implementation
original image and the target image. Next, to
restore the original image from the camouflage
image, the receiver must know the Class Index
Table of the original image. By matching the blocks
in the original image with the blocks in the target
image with similar Standard Deviations into a
pair, the original image and the target image can
be obtained separately Class Index Table.
• Block Transformation Firstly, according to the
block matching method, each pair of blocks has
a close Standard Deviation value. Do not change
the Standard Deviation of the original image,
just change the mean value of the original
image through the average shift. Then, in order
to keep the similarity between the transformed
image and the target image as much as possible,
further rotate the transformed block to one of
four directions: 0∘, 90∘, 180∘, 270∘, and choose
the best direction to minimize Root Mean Square
Error between the rotating block and the target
block.
• Accessorial Information Embeding In order
to obtain the final camouflage image, it is
necessary to embed auxiliary information into the
transformed image, including: compressed Class
Index Table and the average shift and rotation
direction of each block of the original image.
Choose a suitable RDH algorithm embeds these
auxiliary information into the transformed image
to get the final camouflage image.</p>
      <sec id="sec-5-1">
        <title>2.3. Original Image Restoration</title>
        <p>In order to facilitate the understanding of the
implementation of the RIT algorithm, the following takes grayscale Finally, the original image needs to be restored when an
image (one channel) as an example to illustrate the spe- unauthorized model accesses it, the restoration process
cific implementation process of the algorithm [ 18]. For of RIT can be directly used to realize the reverse
transcolor images, the R, G, and B color channels are trans- formation of the reversible adversarial example to the
formed in the same way. RIT achieves the reversible original image. Since our reversible adversarial examples
transformation on two pictures, and there are two stages are based on RIT technology, the process of restoring
of transformation and restoration, in the transformation the reversible adversarial examples to the original image
stage, the original image undergoes a series of pixel value is the restoration process of RIT, while the restoration
transformations to generate a camouflage image [ 18]. In process is the inverse process of the RIT transformation
the recovery stage, the hidden image transformation in- process. Therefore, in the case of only reversible
adversarformation needs to be extracted from the camouflage ial examples, we can extract the hidden transformation
image, and is used for reversible restoration. Since the information, and take the information to reverse the RIT
restoration is the reverse process of the transformation, transformation process to non-destructively restore the
we only need to introduce the transformation process. original images.</p>
        <p>
          The transformation process is divided into three steps:
(
          <xref ref-type="bibr" rid="ref2">1</xref>
          ) Block Paring (
          <xref ref-type="bibr" rid="ref1 ref3">2</xref>
          ) Block Transformation (
          <xref ref-type="bibr" rid="ref4">3</xref>
          ) Accessorial
Information Embeding. 3. Evaluation and Analysis
• Block Paring The original image and the target
image are divided into blocks in the same way
ifrstly. Then, calculate Mean and Standard
Deviation of the pixel values of each block of the
        </p>
        <p>To verify the efectiveness and superiority of the
proposed method, here we introduce the experiment design,
results and comparisons, following with discussion and
analysis.
3.1. Experimental Setup
• Dataset: Since it is meaningless to attack
images that have been mis-classified by the model,
we randomly choose 5000 images from ImageNet
(ILSVRC 2012 verification set) that can be
correctly classified by the model for experiments.
• Deep Network: The pretrained Inception_v3 in
torchvision.models that is evaluated by Top-1
accuracy.
• Attack Methods: IFGSM, C&amp;W, DeepFool. To
ensure the visual quality, we set the learning rate
of C&amp;W_L2 to 0.005, the perturbation amplitude
 of IFGSM no more than 8/225.</p>
      </sec>
      <sec id="sec-5-2">
        <title>3.2. Performance Evaluation</title>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>In order to evaluate the performance of the proposed</title>
      <p>method, we measure attack success rates as well as
image quality of our reversible adversarial examples, and
compare our RIT-based RAE method with RDE-based
RAE method proposed by of Liu et al. [15].</p>
      <p>In order to detect the attack ability of the generated
reversible adversarial examples, firstly, three white-box
attack algorithms are taken to attack the selected
original images to get adversarial examples. Then, we take
reversible image transformation to transform original
images into target adversarial images to generate reversible
adversarial examples. Finally, we utilize the generated
reversible adversarial images to attack the model to get
attack success rates. As shown in Table 1, the second
line shows the attack success rates of the generated
adversarial examples (which are non-reversible). The third
and fourth lines are the attack success rates of Liu et al.’s
and our reversible adversarial examples under diferent
settings, respectively. On IFGSM, when  is 4/225, 8/225,
the attack success rates of our RAEs are: 70.80%, 94.55%
respectively. In the same case, the attack success rates of
Liu et al.’s RAEs are only 35.22%, 81.00%, respectively. On
C&amp;W_L2, when confidence  is 50, 100, the attack
success rates of our RAEs are: 81.02% and 94.84%, while that
Liu et al.’s method are just 52.73%, 55.01%, respectively.
From the results presented in Table 1, we observe that
the attack ability of the RAEs obtained by our method is
superior to that of Liu et al’s method. But on DeepFool,
because the adversarial perturbation generated by this
attack closes to the theoretical minimum, its robustness
is also relatively poor. Therefore, the amount of
information embedded in the adversarial examples generated
by the DeepFool exceeds a certain amount, which will
seriously weaken the attack performance of the
adversarial examples. In this kind of attack algorithm with
minimal disturbance and low robustness, the amount of
auxiliary information embedded in RIT-based RAEs is
greater than the amount of perturbation signal embedded
in RDE-based RAEs of Liu et al., so the success rates of
our RAEs are lower.</p>
      <p>Further more, we found that, when adversarial
perturbations get stronger, the amount of data that needs to
be embedded increases, which leads to the failure of
reversibility for RDE-based RAEs. Take Giant Panda image
from Fig.1 as an example, on C&amp;W, when confidence 
is 100, the amount of data that needs to be embedded by
RDE-based RAE is 316311 bits, that’s far from the
corresponding highest embedding capacity 114986 bits. At the
same time, to achieve reversible attack by using the
proposed RIT-based RAE method, the amount of additional
data that needs to be embedded is only 105966 bits.</p>
      <p>Then, to quantitatively evaluate the image quality of
RAEs, we measure three sets of PSNR: RAEs and
original images,RAEs and adversarial examples as well as
original images and adversarial examples. The general
benchmark for PSNR value is 30dB, and the image
distortion below 30dB can be perceived by human vision.
In order to make a fair comparison with the method of
Liu et al.[15], we keep the original image and the
adversarial example consistent in the experiment, and the
corresponding values of PSNR are shown in the last
column of Table 2. By comparing RAEs on IFGSM and C&amp;W
with the original images, we found that the PSNR values
of our method are higher, that means the generated RAEs
are less distorted than that of Liu et al. The comparison
between the RAEs and the original adversarial examples
shows that our PSNR values are basically greater than
30dB, indicating our RAEs are closer to the original
adversarial examples. This result is also consistent with the
data in Table 1, that means the specific structure of
adversarial perturbation is better preserved in our method,
so that the final RAEs have almost the same attack efect
as the original adversarial example on IFGSM and C&amp;W.
Similar to the experimental data in Table 1 again, for
attack algorithms like DeepFool, the perturbation
embedding amount in Liu et al.’s method is smaller than the
auxiliary information embedding amount in our work,
so the PSNR values of our RAEs are smaller.</p>
      <p>In addition, Fig.4 shows the sample images of RAEs
generated by Liu et al. and our method, respectively.
After partial magnification, we can see that the image
distortion of RDE-based RAEs significantly exceeds that
of RIT-based RAEs. Since the amount of auxiliary
information embedded in RIT-based RAEs is relatively
stable, while the amount of perturbation embedded in
RDEbased RAEs is related to the perturbation signal. The
greater the perturbation, the more the amount of
information embedded, and the more the image distorted.</p>
      <sec id="sec-6-1">
        <title>3.3. Discussion and Analysis</title>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>Both RDE-based RAE and RIT-based RAE use RDH technology to achieve adversarial reversible attacks. In RDE</title>
      <p>the attack efect of our reversible adversarial examples
is afected to a certain extent by the amount of
auxiliary information needed to restore the original image,
and the amount of auxiliary information is usually
relatively stable. Generally speaking, we can reduce the
impact of auxiliary information embedding by
enhancing the adversarial perturbation. That is to say, when
generating an adversarial image, the robustness of the
(A) Original Image (B(C)WAd，vecrosnafriidaelnEcxea=m50pl)e (C)ALdiuveertsaalr.i’alsERxaemveprlseible (D) Our ReEvexrasmibpleleAdversarial adversarial example is improved by increasing the
perturbation amplitude, and finally the attack success rate
of the generated reversible adversarial example is
imFigure 4: Sample figures of reversible adversarial examples proved. However, when faced with an attack algorithm
generated by diferent methods. similar to DeepFool with less perturbation and low
robustness, since RIT auxiliary information embedding has
a greater impact on its performance than perturbation
sigbased RAE framework, Liu et al. take reversible data nal embedding, the attack success rate of our reversible
embedding algorithm to hide the perturbation diference adversarial examples is lower than Liu et al. While the
in the adversarial example to get a reversible adversarial proposed scheme is a special application of RIT, original
image. Constrained by the RDH payload, to achieve re- image and its target adversarial image have a high degree
versibility, the perturbation signal can only be controlled of similarity. Our future work is to improve reversible
within the range of the payload. A slight increase in image transformation algorithm based on the similarity
the perturbation amplitude will cause serious visual dis- between original image and adversarial example so that
tortion of the reversible adversarial example, severely the attack success rate of reversible adversarial examples
weakened attack ability, and even unable to fully embed is further improved.
the perturbation signal so that the original image
cannot be restored reversibly. In the proposed RIT-based 4. Conclusion
RAE framework, since reversible image transformation
is unnecessary to consider the size of the adversarial To solve the problems of RAE technique and improve
perturbation, the problem of dificulty in embedding ad- the performance in terms of reversibility, image
qualversarial perturbation is solved, and it further improves ity and attack ability, we take advantage of reversible
the visual quality of the reversible adversarial example image transformation to construct reversible adversarial
to promote the overall attack success rates. In a sense,
examples, which aims to achieve reversible attack. In this [9] J. Su, D. V. Vargas, K. Sakurai, One pixel attack for
work, we regard a generated adversarial example as the fooling deep neural networks, IEEE Transactions
target image and its original image can be disguised as its on Evolutionary Computation 23 (2019) 828–841.
adversarial example to get RAE. Then the original image [10] S.-M. Moosavi-Dezfooli, A. Fawzi, P. Frossard,
can be recovered from its reversible adversarial example Deepfool: a simple and accurate method to fool
without distortion. Experimental results illustrate that deep neural networks, in: IEEE Conference on
Comour method overcomes the problems of perturbation in- puter Vision and Pattern Recognition (CVPR), 2016,
formation embedding. Moreover, it’s even achieved that pp. 2574–2582. doi:10.1109/CVPR.2016.282.
the larger adversarial perturbation, the better RAE can [11] N. Carlini, D. Wagner, Towards evaluating the
robe generated. RAE can prevent illegal or unauthorized bustness of neural networks, in: IEEE Symposium
access of image data such as human faces and ensure on Security and Privacy (SP), IEEE, 2017, pp. 39–57.
legitimate users can use authorization-protected data. doi:10.1109/SP.2017.49.
Today, when deep learning and other artificial intelli- [12] P. Schöttle, A. Schlögl, C. Pasquini, R. Böhme,
Degence technologies are widely used, this technology is tecting adversarial examples-a lesson from
multiof great significance. In future work, it is worth trying media security, in: European Signal Processing
to further combine more reversible information hiding Conference (EUSIPCO), IEEE, 2018, pp. 947–951.
technologies to study RAE solutions that meet actual [13] E. Quiring, D. Arp, K. Rieck, Forgotten siblings:
needs. Unifying attacks on machine learning and digital
watermarking, in: IEEE European Symposium on
Security and Privacy (EuroS&amp;P), IEEE, 2018, pp.</p>
      <p>Acknowledgments 488–502.
[14] C. Zhang, P. Benz, A. Karjauv, I. S. Kweon, Universal
This research work is partly supported by National Nat- adversarial perturbations through the lens of deep
ural Science Foundation of China (61872003, U1636201). steganography: Towards a fourier perspective, in:
AAAI Conference on Artificial Intelligence, 2021,
References pp. 3296–3304.
[15] J. Liu, D. Hou, W. Zhang, N. Yu, Reversible
adversarial examples., arXiv preprint arXiv: 1811.00189
(2018).
[16] W. Zhang, X. Hu, X. Li, N. Yu, Recursive histogram
modification: establishing equivalency between
reversible data hiding and lossless data compression,
IEEE Transactions on Image Processing 22 (2013)
2775–2785.
[17] D. Hou, C. Qin, N. Yu, W. Zhang, Reversible
visual transformation via exploring the correlations
within color images, Journal of Visual
Communication and Image Representation 53 (2018) 134–145.
[18] W. Zhang, H. Wang, D. Hou, N. Yu, Reversible data
hiding in encrypted images by reversible image
transformation, IEEE Transactions on Multimedia
18 (2016) 1469–1479.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>2 Reversible Adversarial Example Generation</mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [1] LeCun, Yann, Bengio, Yoshua, Hinton, Geofrey, Deep learning,
          <source>Nature</source>
          <volume>521</volume>
          (
          <year>2015</year>
          )
          <fpage>436</fpage>
          -
          <lpage>444</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>S.</given-names>
            <surname>Aradi</surname>
          </string-name>
          ,
          <article-title>Survey of deep reinforcement learning for motion planning of autonomous vehicles</article-title>
          ,
          <source>IEEE Transactions on Intelligent Transportation Systems</source>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>J. Y.</given-names>
            <surname>Choi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <article-title>Ensemble of deep convolutional neural networks with gabor face representations for face recognition</article-title>
          ,
          <source>IEEE Transactions on Image Processing</source>
          <volume>29</volume>
          (
          <year>2019</year>
          )
          <fpage>3270</fpage>
          -
          <lpage>3281</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>C.</given-names>
            <surname>Szegedy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Zaremba</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Sutskever</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bruna</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Erhan</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Goodfellow</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Fergus</surname>
          </string-name>
          ,
          <article-title>Intriguing properties of neural networks</article-title>
          ,
          <source>in: International Conference on Machine Learning (ICML)</source>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>I. J.</given-names>
            <surname>Goodfellow</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Shlens</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Szegedy</surname>
          </string-name>
          ,
          <article-title>Explaining and harnessing adversarial examples</article-title>
          ,
          <source>in: International Conference on Learning Representations (ICLR)</source>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>D.</given-names>
            <surname>Hou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , J. Liu,
          <string-name>
            <given-names>S.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <article-title>Emerging applications of reversible data hiding</article-title>
          ,
          <source>in: International Conference on Image and Graphics Processing (ICIGP)</source>
          , ACM,
          <year>2019</year>
          , pp.
          <fpage>105</fpage>
          -
          <lpage>109</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>Adversarial examples: Opportunities and challenges</article-title>
          ,
          <source>IEEE transactions on neural networks and learning systems 31</source>
          (
          <year>2019</year>
          )
          <fpage>2578</fpage>
          -
          <lpage>2593</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>A.</given-names>
            <surname>Kurakin</surname>
          </string-name>
          , I. Goodfellow,
          <string-name>
            <given-names>S.</given-names>
            <surname>Bengio</surname>
          </string-name>
          ,
          <article-title>Adversarial examples in the physical world</article-title>
          ,
          <source>arXiv preprint arXiv:1607.02533</source>
          (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>