<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>MediaEval 2020: Maintaining Human-Imperceptibility of Image Adversarial Attack by using Human-Aware Sensitivity Map</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Zhiqi Shen</string-name>
          <email>dcsshenz@nus.edu.sg</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Muhammad Furqan Habibi</string-name>
          <email>furqan.habibi@u.nus.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Shaojing Fan</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mohan Kankanhalli</string-name>
          <email>mohan@comp.nus.edu.sg</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>National University of</institution>
          <country country="SG">Singapore</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2020</year>
      </pub-date>
      <fpage>14</fpage>
      <lpage>15</lpage>
      <abstract>
        <p>With the rapid rise of big data with developments in artificial intelligence, privacy has come under the spotlight. Adversarial attacks using image perturbation have recently been introduced to fool machines on pattern recognition tasks. They also have been successfully employed to protect privacy of images. However, only a few works consider the imperceptibility of perturbations for humans. This report presents our submission to the pixel privacy task, where we improve the imperceptibility of image perturbations by using a human-aware sensitivity map, while protecting image privacy via adversarial attack techniques.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>INTRODUCTION</title>
      <p>
        The Pixel Privacy task [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] of MediaEval aims to protect personal
privacy by embedding human-imperceptible noise on images that
fools the BIQA classifiers. The attack models use InceptionResNetV2
structure and are pre-trained on KonIQ-10k dataset. The organizers
evaluated the performance in terms of success attack rate (accuracy)
and imperceptibility of perturbation.
      </p>
      <p>
        Prior work usually applies 2 norm [
        <xref ref-type="bibr" rid="ref1 ref5 ref6">1, 5, 6</xref>
        ] to the loss
function to improve the imperceptibility of perturbed images. However,
2 norm only guarantees the overall noise to be small without
considering the perceptual characteristics of regions. For example,
observers will perceive diferently when we add the same noise to a
lfat background versus a content-rich background. With this insight,
we can apply a sensitivity map to the loss function that indicates
which regions’ changes are least sensitive to observers, so that the
algorithms know where to add the noise. Recent works [
        <xref ref-type="bibr" rid="ref11 ref2 ref4">2, 4, 11</xref>
        ]
published after our earlier work [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] do take human
imperceptibility of perturbations into account. Unlike our deep learning-based
method, most of them compute human imperceptibility based on
texture information.
      </p>
      <p>
        Our method is an optimization-based approach based on the CW
attack [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. We manipulate each input image’s model logits to its
target class. We then optimize the attack to minimize the loss function
by modifying the input image. To improve human imperceptibility,
we improve the loss function by integrating human sensitivity maps
learned from [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. Experimental evaluation indicates our approach
achieves good results in terms of human imperceptibility.
      </p>
    </sec>
    <sec id="sec-2">
      <title>APPROACH</title>
    </sec>
    <sec id="sec-3">
      <title>Preliminaries</title>
      <p>We denote an image by  ∈ ∗ ∗ , where H, W, C is the frame
height, frame width, and the number of channels, respectively. The
BIQA classifier is denoted by  ( ,  ) =  that takes an image
as input and produces the corresponding logits 1 , which  is the
class number. A softmax layer is followed to the network to transfer
the logits to each class’s probability . The whole BIQA classifier is
represented by    (  ( ,  )) = .</p>
      <p>The image adversarial attack approach aims to find an image
perturbation  that maximizes the classification error. We denote
 =  −  the adversarial image perturbation.</p>
      <p>We propose an optimization-based approach. The general idea
of generating perturbation for an image is by using the following
optimization equation.</p>
      <p>arg min  ( ) − ℓ (  ( +  ,  ), ˆ)

(1)
where  (.) is the perception regularization to keep the
perturbation to be small and imperceptible to humans. ˆ is the target logits.
ℓ (., .) is the loss function to measure the diference between the
actual prediction and the target prediction. To obtain a high attack
rate success, we minimize the distance between actual logits and
the target logits.  is a hyper-parameter to balance these two terms.
integrate the human perceptual sensitivity, we extend the L2 norm
by multiplying it with the sensitivity map, as shown below.
2.2</p>
    </sec>
    <sec id="sec-4">
      <title>Loss to fool machines</title>
      <p>
        We follow the loss in [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] to fool machines. For the sake of clarity,
we use  = ℓ (  ( +  ,  ), ˆ), the detailed formulation is as follows:
 =
( | (  ( +  )) −  (ˆ) |, if   ( +  ) ≠  ˆ
0, otherwise
(2)
      </p>
      <p>Where  ( +  ) and ˆ are the one-hot vectors representing the
current logits and desired logits. The losses consist of two parts. The
ifrst part represents the situations when the perturbed image has not
been into our desired class. The loss value is the absolute distance
between the most trusted class in current logits and the desired
class. The second part depicts the situation when the perturbed
image has been classified into our desired class, so we set the loss
value to zero.
2.3</p>
    </sec>
    <sec id="sec-5">
      <title>Loss to fool humans</title>
      <p>We observed that the traditional norms (e.g., 0, 2,  ) consider
all pixels in the images to be equal, while humans have diferent
priorities when viewing diferent image regions. More specifically,
even adding the same perturbation noise to diferent regions will
lead to diferent humans’ perceptibility. For quantifying humans’
perceptibility of each pixel, we integrate a sensitivity map with our
loss function. The value of each pixel in the sensitivity map ranges
from 0 to 1. The larger value indicates more chance to be perceived
when adding noise on such pixels.</p>
      <p>
        Human-aware sensitivity map Human perception is a complex
phenomenon which is not easily captured in a neat mathematical
formulation. Therefore, we train a neural network to generate the
spatially dense prediction of each pixel with human sensitivity
scores. The network is designed based on a fully convolutional
network (FCN) [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. The backbone network is a VGG-16 [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] model
pre-trained ImageNet dataset. A 1*1 convolutional layer is used
to combine all feature maps extracted from VGG-16 to obtain the
ifnal sensitivity map. The architecture of our DNN is illustrated in
Figure 2.
      </p>
      <p>
        Embed sensitivity maps to attack approach For this workshop,
we train the sensitivity map generation model on the EMOd dataset
[
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] and then test it on the given Place365 testing set. In order to
2
 ( ) =   2
(3)
3
      </p>
    </sec>
    <sec id="sec-6">
      <title>RESULTS AND ANALYSIS</title>
      <p>We submitted five runs towards the Pixel Privacy task. The
organizers selected 20 images with the largest BIQA variance for human
evaluation. They then put the same image of all qualified runs in
one folder and let 7 experts select the most appealing (i.e., “Best”)
three runs out of 17 runs. A run can be selected as “Best” for at
most 140 times.</p>
      <p>From Table 1, we can observe that the accuracy of our first run
(with parameter  = 10) has dropped to lower than random guess
(50%), meaning that our perturbed images have fooled machines’
prediction. More importantly, more than half of the images are
selected as the best three images out of 17 runs. From the trend of
parameters, we can see the potential of our algorithm. If we can
try more parameters (e.g,. smaller than 10), the performance might
be even better than the current one. For the other runs, we have
not achieved a good attack rate. This is because the parameter  is
too large that forces the perturbed images to focus more on image
quality during back-propagation.
This report introduces our approach for privacy protection, which
integrates the human-aware sensitivity map to the loss function
to improve the quality of perturbed images’. The results
demonstrate the efectiveness of the sensitivity map in maintaining noise
imperceptibility. However, some aspects can be further improved.
The current sensitivity map prediction network is trained on the
EMOd dataset, which has only 698 images. Another problem is that
the network structure (FCN) is rudimentary. We can foresee that
with a more sophisticated structure, trained on a larger data-set,
can improve the performance.</p>
      <p>Pixel Privacy: Quality Camouflage for Social Images</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Nicholas</given-names>
            <surname>Carlini</surname>
          </string-name>
          and
          <string-name>
            <given-names>David</given-names>
            <surname>Wagner</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Towards evaluating the robustness of neural networks</article-title>
          .
          <source>In 2017 ieee symposium on security and privacy (sp)</source>
          .
          <source>IEEE</source>
          ,
          <fpage>39</fpage>
          -
          <lpage>57</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Francesco</given-names>
            <surname>Croce</surname>
          </string-name>
          and
          <string-name>
            <given-names>Matthias</given-names>
            <surname>Hein</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Sparse and imperceivable adversarial attacks</article-title>
          .
          <source>In Proceedings of the IEEE International Conference on Computer Vision</source>
          . 4724-
          <fpage>4732</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Shaojing</given-names>
            <surname>Fan</surname>
          </string-name>
          , Zhiqi Shen, Ming Jiang, Bryan L Koenig,
          <article-title>Juan Xu, Mohan S Kankanhalli,</article-title>
          and
          <string-name>
            <given-names>Qi</given-names>
            <surname>Zhao</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Emotional attention: A study of image sentiment and visual attention</article-title>
          .
          <source>In Proceedings of the IEEE Conference on computer vision and pattern recognition</source>
          .
          <volume>7521</volume>
          -
          <fpage>7531</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Diego</given-names>
            <surname>Gragnaniello</surname>
          </string-name>
          , Francesco Marra, Giovanni Poggi, and
          <string-name>
            <given-names>Luisa</given-names>
            <surname>Verdoliva</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Perceptual Quality-preserving Black-Box Attack against Deep Learning Image Classifiers</article-title>
          . arXiv preprint arXiv:
          <year>1902</year>
          .
          <volume>07776</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Alexey</given-names>
            <surname>Kurakin</surname>
          </string-name>
          , Ian Goodfellow, and
          <string-name>
            <given-names>Samy</given-names>
            <surname>Bengio</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Adversarial examples in the physical world</article-title>
          .
          <source>arXiv preprint arXiv:1607.02533</source>
          (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>Yanpei</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Xinyun</given-names>
            <surname>Chen</surname>
          </string-name>
          , Chang Liu, and
          <string-name>
            <given-names>Dawn</given-names>
            <surname>Song</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Delving into transferable adversarial examples and black-box attacks</article-title>
          .
          <source>arXiv preprint arXiv:1611.02770</source>
          (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Zhuoran</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Zhengyu</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Martha</given-names>
            <surname>Larson</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Laurent</given-names>
            <surname>Amsaleg</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>Exploring Quality Camouflage for Social Images</article-title>
          .
          <source>In Working Notes Proceedings of the MediaEval Workshop.</source>
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Jonathan</given-names>
            <surname>Long</surname>
          </string-name>
          , Evan Shelhamer, and
          <string-name>
            <given-names>Trevor</given-names>
            <surname>Darrell</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <article-title>Fully convolutional networks for semantic segmentation</article-title>
          .
          <source>In Proceedings of the IEEE conference on computer vision and pattern recognition</source>
          .
          <volume>3431</volume>
          -
          <fpage>3440</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Zhiqi</given-names>
            <surname>Shen</surname>
          </string-name>
          , Shaojing Fan, Yongkang Wong,
          <string-name>
            <surname>Tian-Tsong Ng</surname>
            , and
            <given-names>Mohan</given-names>
          </string-name>
          <string-name>
            <surname>Kankanhalli</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Human-imperceptible privacy protection against machines</article-title>
          .
          <source>In Proceedings of the 27th ACM International Conference on Multimedia. 1119-1128.</source>
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>Karen</given-names>
            <surname>Simonyan</surname>
          </string-name>
          and
          <string-name>
            <given-names>Andrew</given-names>
            <surname>Zisserman</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>Very deep convolutional networks for large-scale image recognition</article-title>
          .
          <source>arXiv preprint arXiv:1409.1556</source>
          (
          <year>2014</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Eric</surname>
            <given-names>Wong</given-names>
          </string-name>
          , Frank R Schmidt, and
          <string-name>
            <given-names>J Zico</given-names>
            <surname>Kolter</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Wasserstein adversarial examples via projected sinkhorn iterations</article-title>
          . arXiv preprint arXiv:
          <year>1902</year>
          .
          <volume>07906</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>