<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Sympo-
sium on Adversary-Aware Learning Techniques and Trends in Cy-
bersecurity, Arlington, VA, USA</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Projecting Trouble: Light Based Adversarial Attacks on Deep Learning Classifiers</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Nicole Nichols 1,2</string-name>
          <email>nicole.nichols@pnnl.gov</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>1Pacific Northwest National Laboratory</institution>
          ,
          <addr-line>Seattle, Washington</addr-line>
          ,
          <institution>2Western Washington University</institution>
          ,
          <addr-line>Bellingham, Washington</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2018</year>
      </pub-date>
      <volume>1</volume>
      <fpage>8</fpage>
      <lpage>19</lpage>
      <abstract>
        <p>This work demonstrates a physical attack on a deep learning image classification system using projected light onto a physical scene. Prior work is dominated by techniques for creating adversarial examples which directly manipulate the digital input of the classifier. Such an attack is limited to scenarios where the adversary can directly update the inputs to the classifier. This could happen by intercepting and modifying the inputs to an online API such as Clarifai or Cloud Vision. Such limitations have led to a vein of research around physical attacks where objects are constructed to be inherently adversarial or adversarial modifications are added to cause misclassification. Our work differs from other physical attacks in that we can cause misclassification dynamically without altering physical objects in a permanent way. We construct an experimental setup which includes a light projection source, an object for classification, and a camera to capture the scene. Experiments are conducted against 2D and 3D objects from CIFAR-10. Initial tests show projected light patterns selected via differential evolution could degrade classification from 98% to 22% and 89% to 43% probability for 2D and 3D targets respectively. Subsequent experiments explore sensitivity to physical setup and compare two additional baseline conditions for all 10 CIFAR classes. Some physical targets are more susceptible to perturbation. Simple attacks show near equivalent success, and 6 of the 10 classes were disrupted by light.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Machine learning models are vulnerable to adversarial
attacks by making small but targeted modifications to inputs
that cause misclassification. The research around
adversarial attacks on deep learning systems has grown significantly
since
        <xref ref-type="bibr" rid="ref17">(Szegedy et al. 2013)</xref>
        demonstrated intriguing
properties. The scope and limitations of such attacks is an active
area of research in the academic community. Most of the
research has focused on the purely digital manipulation.
Recently, researchers have developed techniques that alter or
manipulate physical objects to fool classifiers, which could
pose a much greater real world threat.
      </p>
    </sec>
    <sec id="sec-2">
      <title>Related Research</title>
      <p>
        Researchers have proposed many theories about the cause
of model vulnerabilities. Evidence suggests that adversarial
samples lie close to the decision boundary in the low
dimensional manifold representing high dimensional data.
Adversarial manipulation in the high dimension is often
imperceptible to humans and can shift the low dimensional
representation to cross the decision boundary (Feinman et al. 2017).
Many approaches are available to perform this manipulation
if the attacker has access to the defender’s classifier.
Furthermore, adversarial examples have empirically been shown
to transfer between different classifier types
        <xref ref-type="bibr" rid="ref11 ref13 ref15 ref17">(Papernot,
McDaniel, and Goodfellow 2016; Szegedy et al. 2013)</xref>
        . This
enhances the attacker’s potential capability when there is no
access to the defender’s classifier.
      </p>
      <p>
        It is difficult for defenses to keep pace with attacks, and
the advantage lies with the adversary. This was highlighted
when seven of the eight white box defenses announced at
the prestigious ICLR2018 were defeated within a week of
publication
        <xref ref-type="bibr" rid="ref2">(Athalye, Carlini, and Wagner 2018)</xref>
        .
      </p>
      <p>
        Researchers have successfully demonstrated physical
world attacks against deep learning classifiers. Some of the
first physical attacks were demonstrated by printing an
adversarial example, photographing the printed image, and
verifying the adversarial attack remained
        <xref ref-type="bibr" rid="ref11 ref15">(Kurakin,
Goodfellow, and Bengio 2016)</xref>
        . (Sharif et al. 2016) demonstrated
printed eyeglasses frames that thwart facial recognition
systems and fully avoid face detection by the Viola-Jones object
detection algorithm. It has also been noted that near
infrared light can also be used to evade face detection
        <xref ref-type="bibr" rid="ref18">(Yamada,
Gohshi, and Echizen 2013)</xref>
        . Our work is different because
we leverage dynamic generation methods use real world
feedback when learning the patterns of light to project.
      </p>
      <p>
        Putting aside adversarial attacks, most image classifiers
are not inherently invariant to object scale, translation, or
rotation. Notable exceptions are
        <xref ref-type="bibr" rid="ref6">(Cohen and Welling 2014)</xref>
        ,
which attempts to learn object recognition by construction
of parts, and
        <xref ref-type="bibr" rid="ref14">(Qi et al. 2017)</xref>
        which use 3D point cloud
representation for object classification. To some degree, this
invariance can be learned from training data if it has
intentionally been designed to address this gap. For example the early
work by
        <xref ref-type="bibr" rid="ref12">(LeCun, Huang, and Bottou 2004)</xref>
        was evaluated
with the NORB dataset which was systematically collected
to assess pose, lighting, and rotation of 3D objects.
      </p>
      <p>
        Simulating scale, translation, and rotation of 2D images
is conducive to experiment automation, and many recent
advances in rotational invariance such as Spatial Transformer
Networks
        <xref ref-type="bibr" rid="ref9">(Jaderberg et al. 2015)</xref>
        , use this framework for
evaluation of robustness to these properties. However,
further research is needed to validate the ability of this
simulated rotational invariance to transfer to real world rotation
of 3D figures. We emphasize the need for invariant models
because it is impossible to disambiguate the success of an
attack when it is can only be validated with a weak model.
      </p>
      <p>
        Maintaining adversarial attack under a range of pose or
lighting conditions may prove to be the most difficult
aspect of this task. Some preliminary research suggests this is
possible and demonstrated two toy examples in the
physical world
        <xref ref-type="bibr" rid="ref1">(Athalye and Sutskever 2017)</xref>
        . They introduce an
Expectation over Transformation (EoT) method for
differentiating texture patterns through a 3D renderer to produce
an adversarial object. An additional demonstration of
physical attack is to introduce an adversarial patch to the
physical scene, which is invariant to location, rotation, scale, and
cause specific misclassification (Brown et al. 2017).
      </p>
    </sec>
    <sec id="sec-3">
      <title>Experimental Setup and Results</title>
      <p>We constructed a test environment to perform light based
adversarial attacks and collect data in an office environment
with minimal lighting control. Our attacks were conducted
against 2D and 3D target objects placed in the scene. We
used a projector to project light onto the target and a
common web camera to capture the scene. For the 2D and initial
3D experiments, the projector was a Casio XJ-A257 and the
camera was a Logitech C930e. During the second phase of
3D experiments, we used an Epson VS250 projector,
Logitech C615 HD camera and an Altura HD-ND8, neutral
density filter to control the light intensity of the projector.</p>
      <sec id="sec-3-1">
        <title>2D Presentation</title>
        <p>
          For the 2D scene, we chose a random image (horse) from
the CIFAR-10 dataset to be attacked. The image was printed
and secured to the wall in front of the camera and
projector. Following a similar methodology of earlier work
          <xref ref-type="bibr" rid="ref1 ref14 ref16">(Su,
Vargas, and Kouichi 2017)</xref>
          on single pixel attacks we use
differential evolution (DE) to optimize a light based attack
to cause misclassification. Differential evolution is a
heuristic global optimization strategy similar to genetic algorithms
where the algorithm maintains a population of candidate
solutions, selecting a small number (potentially one) for
further rounds of modification and refinement. We projected
a digital black 32x32 square containing a single pixel at a
variable location and RGB values. Because projectors can’t
project black (the absence of light) the projector adjusted the
black pixels to present the illusion of a black background.
This adjustment is impacted somewhat by RGB value of
the single pixel being projected. Each iteration of the
differential evolution was projected, captured, and input to a
standard ResNet38 for classification of the image captured
by the camera. Though only one pixel was modified in the
digital attack pattern, because of the distance between the
projector and object, a larger area in the captured scene and
many input pixels to the camera are modified. The original
and attacked scenes are shown in Figure 1.
        </p>
        <p>Through this attack, the probability of horse was
decreased from 98% to 22%.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3D Presentation</title>
        <p>To demonstrate the potential for light based attacks, we
extended the 2D methodology to a 3D scene in two
experimental phases. First, we placed a toy car in the field of view of
the web camera to capture the scene. To perform the attack,
the projector iteratively applies the same adversarial noise
procedure to the 3D physical scene and the same ResNet38
model is used for evaluation. The object probabilities for the
original scene were 89% automobile and 11% truck.
The attacked scene probabilities were 43 % automobile
and 57% truck. The second phase of experiments was
designed to improve the repeatability and confidence of the
initial demonstration. Results are expanded to evaluate all 10
CIFAR classes: airplane, automobile, bird, cat,
deer, dog, frog, horse, ship, truck. The figurines
used for each of these classes are shown in Figure 4a. The
yellow car in phase 1 was not available and was replaced
with a red car in phase 2.</p>
        <p>Rotation invariance is important for interpreting the
presented experimental setup. This impacts our data
collection because we observed in a baseline condition, with
no added light, the distance to the camera and object
orientation yielded highly variable classification results. We
tested four experimental conditions: ambient light, white
light from the projector, white light with a randomly located
pixel in the 32x32 grid, and differential evolution process
to control color and location of one pixel in a 32x32 white
grid. We observed classification variability in the physical
scene when no modifications were applied. For this reason
we introduced some lighting controls which
observationally provided a significantly more stable baseline
classification. Three physical modifications were made. The projected
background color was changed from black to white to
provide more uniformity to the scene. We used a foam block to
minimize stray reflections caused by the projector.
Additionally we used a neutral density filter to scale the light
intensity. To verify stability, we collected twenty image captures
of each test condition, and 200 for differential evolution (50
population sample and 4 evolution phases).</p>
        <p>Reproducibility of the physical placement of each object
in the scene is imprecise, thus each test condition was
collected in sequence without any disturbance (besides light).
An unrecorded calibration phase was used to reposition
the object for a maximum baseline classification score
before the recorded baseline and light projected data was
collected. For each class and test condition, we report the mean,
median, standard deviation, variance, minimum, maximum,
mean and median. The mean and median are the
computation of the reduction in probability score for the
given attack type relative to baseline. Larger numbers
represent more powerful decrease in the true class probability.
(a) The 2D scene without adversarial attack.
(b) The 2D scene with adversarial attack.</p>
        <sec id="sec-3-2-1">
          <title>All scores are reported in Table 1.</title>
          <p>Interpreting the table yields one immediate observation:
some examples (Automobile, Bird, Horse, Ship) are
invariant to the light attack, consistently being identified as
the true class at 100% (within rounding error) while other
classes (Airplane, Cat, Deer, Dog, Frog, and Truck)
have varying degrees of susceptibility. It is unclear whether
these differences are inherent in the classes themselves, or
to the particular figurines we chose. As one might expect
with a research classifier, there is a high degree of variability
based on the particular example. We incremented the
complexity of light attack from pure white light, random square,
and differential evolution, to assess if there was something
unique in the more sophisticated attack, or if it was merely
the addition of light, or a pattern, that was causing the
observed decrease in classification. In many cases, the simple
addition of white light is as effective as the other attacks. For
example, the mean airplane class was decreased from 1.000
to 0.151, with only the addition of white light. The
corresponding trials with random and differential evolution light
patterns yielded only slightly stronger attacks, with 0.113
and 0.133 mean scores respectively. However, the decline is
noteworthy, independent of sophistication.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Discussion</title>
      <p>
        Physical attacks on machine learning systems could be
applied in a wide range of security domains. The literature
has primarily discussed the safety of road signs and
autonomous driving
        <xref ref-type="bibr" rid="ref4 ref7">(Eykholt et al. 2017; Chen et al. 2018)</xref>
        ,
however other security applications may also be impacted.
An adversary may be trying to hide themselves or
physical ties to illegal activities to evade law enforcement (e.g.
knives/weapons, contraband, narcotics manufacturing, etc).
Any AI to be deployed for law-enforcement applications
needs to be robust in an adversarial environment where
physical obfuscation could be employed. Light based
attacks:
      </p>
      <sec id="sec-4-1">
        <title>Can perform targeted and non-targeted attacks.</title>
      </sec>
      <sec id="sec-4-2">
        <title>Do not modify physical object in a permanent way.</title>
      </sec>
      <sec id="sec-4-3">
        <title>Can be a transient effect occurring at specified times.</title>
        <p>This work aims to be a first step towards understanding the
abilities and limitations of such physical attacks. We picked
a relatively easy first target to verify the possibility and plan
to extend this to more complex physical scenarios and
classification models.</p>
        <p>
          We chose to attack the CIFAR-10 framework in a manner
similar to what was demonstrated in the original single pixel
attack
          <xref ref-type="bibr" rid="ref1 ref14 ref16">(Su, Vargas, and Kouichi 2017)</xref>
          . This framework is
an easier target because it is a low resolution, low
parameter model. To assess the robustness of stronger models, a
ResNet50 classifier trained on ImageNet was also used to
evaluate all of the collected images. Because of a lack of
corresponding true class identification, scores are not reported,
but it was observed that the top1 class prediction was shifted
with the addition of light based attacks.
        </p>
        <p>There is also a closed world assumption of 10 relatively
dissimilar classes, where the probability of all classes sums
to one. When a misclassification occurs, it tends to be more
outlandish than it could otherwise be. For example, rose
and tulip might be a more forgiving mistake than frog
and airplane but in the CIFAR closed world framework,
the model is limited to the 10 known classes.</p>
        <p>
          In our attack on the 3D presentation, the true class was
correctly identified as car when no attack was present. By
applying the adversarial light attack, we were able to
decrease the confidence of car from 89% to 43%, and instead
predict truck with 57% probability. We would not
identify this as a 3D attack because we had a fixed orientation
between the camera, projector, and object. In this example,
the single square attack is visually perceptible but transient.
However, the notion of human perception is not as simple
as an L1 distance in pixel space. This is highlighted by
the fact that consecutive video frames can be significantly
mis-classified by top performing image classification
systems
          <xref ref-type="bibr" rid="ref19">(Zheng et al. 2016)</xref>
          . Images that are imperceptibly
different can have large distance in pixel or feature space, and
images that are perceptually different can be close.
        </p>
        <p>A key topic that needs further understanding is why the
extreme variability in class identification. One potential
explanation is the degree of self similarity within a class, and
training data bias. For example, the horse images in the
training data, are potentially all self similar and also closely
match the example figurine. The variation between different
types of horses is likely smaller than the visual difference
between different breeds of dogs.</p>
        <p>Another possible explanation is the scale or percentage of
the scene that the object occupies. Most of the classes which
(a) The 3D scene without any adversarial attack.
(b) The 3D scene with adversarial attack.
(a) Downsampled image without any adversarial attack.
(b) Downsampled image with adversarial attack.
(a) The toy figurines used to represent the CIFAR classes.
(b) Physical setup demonstrating relative position of projector,
camera, object, and lighting control.
were susceptible to attack were relatively small. The notable
exception was the truck which was actually the largest figure
used for data, yet was still susceptible to misclassification
errors with the addition of light.</p>
        <p>There are a several important constraints present when
crafting a light based physical attack that are unconstrained
in a digital attack. Specifically, light is always an additive
noise and turning a dark color to white with the addition of
light is impossible. The angle of projection and the texture
of the scene may impact the colors reflected to the camera.
The camera itself will introduce color balance changes as
it adjusts to the adversarial addition of light. Even a fully
manual camera will always have CCD shot noise, which is a
function of shutter speed and temperature, that could
influence the success or failure of a light based attack. The
projected pixel was not constrained to overlap the target object,
and would appear in the background. Empirically, these
single pixel projections onto the background of an image could
significantly change classifier predictions.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Conclusion and Future Work</title>
      <p>
        The presented work is an empirical demonstration of light
based attacks on deep learning based object recognition
systems. Adversarial machine learning research has
emphasized attacks against deep learning architectures, however
it has been observed that other models are equally
susceptible to attack and that adversarial examples often transfer
between model types
        <xref ref-type="bibr" rid="ref11 ref13 ref15">(Papernot, McDaniel, and Goodfellow
2016)</xref>
        . The empirical demonstration of light based attack
was against a deep learning architecture. However, based
on this prior work, it is likely that it could be demonstrated
against other model types.
      </p>
      <p>We plan on conducting experiments with higher
resolution and more robust classifiers and more subtle
manipulations. We believe that more targeted optimization
approaches that initially focus on sensitive image areas will
likely lead to faster identification of successful attacks. We
expect light based attacks could use more complex projected
textures and take advantage of 3D geometry. Presented
results clearly show light has the potential to be another
avenue of adversarial attack in the physical domain.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>The research described in this paper is part of the Analysis in
Motion Initiative at Pacific Northwest National Laboratory;
conducted under the Laboratory Directed Research and
Development Program at PNNL, a multi-program national
laboratory operated by Battelle for the U.S. Department
of Energy. The authors are especially grateful to Mark
Greaves, Artem Yankov, Sean Zabriskie, Michael Henry,
Jeremiah Rounds, Court Corley, Nathan Hodas, Will Koella
and our Quickstarter supporters.</p>
      <sec id="sec-6-1">
        <title>Automobile Bird Cat Deer</title>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <surname>Athalye</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Sutskever</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          <year>2017</year>
          .
          <article-title>Synthesizing robust adversarial examples</article-title>
          .
          <source>arXiv preprint arXiv:1707</source>
          .
          <fpage>07397</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <surname>Athalye</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Carlini</surname>
          </string-name>
          , N.; and
          <string-name>
            <surname>Wagner</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <year>2018</year>
          .
          <article-title>Obfuscated gradients give a false sense of security: Circumventing defenses to adversarial examples</article-title>
          . arXiv preprint arXiv:
          <year>1802</year>
          .00420.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          2017.
          <article-title>Adversarial patch</article-title>
          .
          <source>arXiv preprint arXiv:1712</source>
          .
          <fpage>09665</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <surname>Chen</surname>
          </string-name>
          , S.-T.;
          <string-name>
            <surname>Cornelius</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Martin</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Chau</surname>
            ,
            <given-names>D. H.</given-names>
          </string-name>
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <article-title>Robust physical adversarial attack on faster r-cnn object detector</article-title>
          . arXiv preprint arXiv:
          <year>1804</year>
          .05810.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <surname>Cohen</surname>
            ,
            <given-names>T. S.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Welling</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <year>2014</year>
          .
          <article-title>Transformation properties of learned visual representations</article-title>
          .
          <source>arXiv preprint arXiv:1412</source>
          .
          <fpage>7659</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <surname>Eykholt</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Evtimov</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Fernandes</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Rahmati</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Xiao</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Prakash</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Kohno</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Song</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <year>2017</year>
          .
          <article-title>Robust Physical-World Attacks on Deep Learning Models</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          2017.
          <article-title>Detecting adversarial samples from artifacts</article-title>
          .
          <source>arXiv preprint arXiv:1703</source>
          .
          <fpage>00410</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <surname>Jaderberg</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Simonyan</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Zisserman</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ; et al.
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <article-title>Spatial transformer networks</article-title>
          .
          <source>In Advances in neural information processing systems</source>
          , 2017-
          <fpage>2025</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <surname>Kurakin</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Goodfellow</surname>
            ,
            <given-names>I.;</given-names>
          </string-name>
          and
          <string-name>
            <surname>Bengio</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <year>2016</year>
          .
          <article-title>Adversarial examples in the physical world</article-title>
          .
          <source>Arxiv (c)</source>
          :
          <fpage>1</fpage>
          -
          <lpage>15</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <surname>LeCun</surname>
          </string-name>
          , Y.;
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>F. J.;</given-names>
          </string-name>
          and
          <string-name>
            <surname>Bottou</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <year>2004</year>
          .
          <article-title>Learning methods for generic object recognition with invariance to pose and lighting</article-title>
          .
          <source>In Computer Vision and Pattern Recognition</source>
          ,
          <year>2004</year>
          .
          <article-title>CVPR 2004</article-title>
          .
          <article-title>Proceedings of the 2004</article-title>
          IEEE Computer Society Conference on, volume
          <volume>2</volume>
          , II-104. IEEE.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <string-name>
            <surname>Papernot</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>McDaniel</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Goodfellow</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          <year>2016</year>
          .
          <article-title>Transferability in Machine Learning: from Phenomena to BlackBox Attacks using Adversarial Samples</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <string-name>
            <surname>Qi</surname>
            ,
            <given-names>C. R.</given-names>
          </string-name>
          ; Su,
          <string-name>
            <given-names>H.</given-names>
            ;
            <surname>Mo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            ; and
            <surname>Guibas</surname>
          </string-name>
          ,
          <string-name>
            <surname>L. J.</surname>
          </string-name>
          <year>2017</year>
          .
          <article-title>Pointnet: Deep learning on point sets for 3d classification and segmentation</article-title>
          .
          <source>Proc. Computer Vision</source>
          and
          <article-title>Pattern Recognition (CVPR), IEEE 1(2):4</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          2016.
          <article-title>Accessorize to a crime: Real and stealthy attacks on state-of-the-art face recognition</article-title>
          .
          <source>In Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security</source>
          ,
          <fpage>1528</fpage>
          -
          <lpage>1540</lpage>
          . ACM.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <string-name>
            <surname>Su</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Vargas</surname>
            ,
            <given-names>D. V.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Kouichi</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <year>2017</year>
          .
          <article-title>One pixel attack for fooling deep neural networks</article-title>
          .
          <source>arXiv preprint arXiv:1710</source>
          .
          <fpage>08864</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <string-name>
            <surname>Szegedy</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Zaremba</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ; Sutskever,
          <string-name>
            <surname>I.</surname>
          </string-name>
          ; Bruna,
          <string-name>
            <given-names>J.</given-names>
            ;
            <surname>Erhan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            ;
            <surname>Goodfellow</surname>
          </string-name>
          ,
          <string-name>
            <surname>I.;</surname>
          </string-name>
          and Fergus,
          <string-name>
            <surname>R.</surname>
          </string-name>
          <year>2013</year>
          .
          <article-title>Intriguing properties of neural networks</article-title>
          .
          <source>arXiv preprint arXiv:1312</source>
          .
          <fpage>6199</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          <string-name>
            <surname>Yamada</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Gohshi</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Echizen</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          <year>2013</year>
          .
          <article-title>Privacy visor: Method for preventing face image detection by using differences in human and device sensitivity</article-title>
          .
          <source>In IFIP International Conference on Communications and Multimedia Security</source>
          ,
          <fpage>152</fpage>
          -
          <lpage>161</lpage>
          . Springer.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          <string-name>
            <surname>Zheng</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ; Song,
          <string-name>
            <given-names>Y.</given-names>
            ;
            <surname>Leung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            ; and
            <surname>Goodfellow</surname>
          </string-name>
          ,
          <string-name>
            <surname>I.</surname>
          </string-name>
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>