<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>ML, October</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Alignment and Adversarial Robustness: Are More Human-Like Models More Secure?</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Blaine Hoak</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Kunyang Li</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Patrick McDaniel</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Wisconsin-Madison</institution>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <volume>26</volume>
      <issue>2025</issue>
      <fpage>0000</fpage>
      <lpage>0003</lpage>
      <abstract>
        <p>A small but growing body of work has shown that machine learning models which better align with human vision have also exhibited higher robustness to adversarial examples, raising the question: can human-like perception make models more secure? If true generally, such mechanisms would ofer new avenues toward robustness. In this work, we conduct a large-scale empirical analysis to systematically investigate the relationship between representational alignment and adversarial robustness. We evaluate 144 models spanning diverse architectures and training paradigms, measuring their neural and behavioral alignment and engineering task performance across 105 benchmarks as well as their adversarial robustness via AutoAttack. Our findings reveal that while average alignment and robustness exhibit a weak overall correlation, specific alignment benchmarks serve as strong predictors of adversarial robustness, particularly those that measure selectivity toward texture or shape. These results suggest that diferent forms of alignment play distinct roles in model robustness, motivating further investigation into how alignment-driven approaches can be leveraged to build more secure and perceptuallygrounded vision models.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Representational alignment—how closely a model resembles biological vision—has been studied
extensively with the goal of measuring, bridging, or increasing alignment in machine learning models [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
Recent observations [
        <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
        ] suggest that alignment may have implications beyond neuroscience: models
that are more aligned with human perception have also exhibited increased robustness to adversarial
examples—inputs with near-imperceptible perturbations that induce model misclassification— hinting
at a deeper connection between alignment and security.
      </p>
      <p>
        However, the relationship between representational alignment and adversarial robustness remains
poorly understood. While the former seeks to align models with human cognition, adversarial examples
in security highlight a fundamental misalignment: imperceptible perturbations can drastically degrade
model accuracy while leaving human perception unafected. Prior robustness techniques, such as
adversarial training [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], are computationally expensive and potentially vulnerable to new attack strategies.
Although recent work has suggested links between perceptual alignment and robustness, especially
in models incorporating biological priors [
        <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
        ], a broad and systematic evaluation of this relationship
across diverse models and alignment benchmarks is still lacking. A fundamental question remains: do
these objectives complement each other, leading to better-aligned and more robust models, or do they
introduce conflicting trade-ofs?
      </p>
      <p>
        In this work, we investigate the relationship between human alignment and robustness to adversarial
examples in vision models through a diverse, large-scale empirical analysis. In our analysis, we study
144 models across diferent architectures and training schemes, and measure their alignment across 105
diferent benchmarks on neural, behavioral, and engineering tasks via the BrainScore library [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. We
then evaluate the robustness of these models using AutoAttack [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], a state-of-the-art ensemble attack.
      </p>
      <p>Our analysis reveals that while robustness is only weakly correlated with average vision alignment,
specific alignment benchmarks are strongly predictive of adversarial robustness. We also find that models
with similar alignment profiles exhibit similar robustness. Together, these findings suggest that diferent
forms of alignment contribute diferently to robustness, highlighting the value of alignment-driven
approaches for improving security in vision systems.</p>
      <p>• Average alignment, particularly for behavioral benchmarks, poorly predicts robustness (explaining
6% of variance), demonstrating that not all methods of alignment increase robustness.
• Individual benchmarks are much better indicators of robust accuracy, but the specific benchmark
matters; we found instances of benchmarks that contributed positively and ones that contributed
negatively to robustness in every alignment category.
• The top benchmarks that are most indicative of robust accuracy show clear trends that inform
the factors leading to higher robustness: (1) robust models process texture information specifically
more similarly to humans than non-robust models to and (2) models’ ability to recognize objects
without global structures intact actually hurts their robustness.
• t-SNE visualizations reveal that robustness is structured in alignment space, with similar models
forming robustness-consistent clusters and demonstrating that alignment on certain benchmarks
is a good indicator of robustness.</p>
      <p>In summary, we uncover specific features of alignment that closely tie with model robustness and
show that increasing alignment on these benchmarks ofers a new avenue for building more robust and
human-aligned models.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Background</title>
      <sec id="sec-2-1">
        <title>2.1. Representational Alignment</title>
        <p>
          Representational alignment studies the extent to which internal representations of machine learning
models correspond to human cognitive processes. Early studies found that deep neural networks
(DNNs) trained on large-scale image datasets develop hierarchical feature representations similar to
those observed in the primate ventral stream, particularly in high-level visual areas like the inferior
temporal (IT) cortex [
          <xref ref-type="bibr" rid="ref5 ref7">7, 5</xref>
          ]. This led to eforts to quantify the alignment between artificial and biological
vision, using techniques such as Representational Similarity Analysis (RSA) [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] and Centered Kernel
Alignment (CKA) [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ]. To improve alignment, researchers have proposed strategies that incorporate
cognitive constraints or psychological priors into model architectures [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]. Supervised fine-tuning with
human-annotated datasets [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ] ensures that learned representations align more closely with
humanunderstandable features. Furthermore, novel techniques [
          <xref ref-type="bibr" rid="ref11 ref12 ref3">11, 3, 12</xref>
          ] have been developed to encourage
similarity between model activations and human neural responses as recorded through fMRI and EEG
experiments. In this study, we use a comprehensive set of neural, behavioral, and engineering alignment
metrics to quantify representational alignment.
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Adversarial Examples</title>
        <p>
          Although machine learning models have shown strong capabilities to achieve high accuracy in various
tasks [
          <xref ref-type="bibr" rid="ref10 ref13 ref14 ref15">13, 10, 14, 15</xref>
          ], they remain vulnerable to adversarial examples [
          <xref ref-type="bibr" rid="ref16 ref17 ref18 ref4 ref6">6, 4, 16, 17, 18</xref>
          ]. Adversarial
examples are specially crafted inputs that contain perturbations which are imperceptible to humans, yet
can significantly decrease model accuracy. In computer vision systems, there have been many studies
on developing attack algorithms, such as FGSM [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ], PGD [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ], and AutoAttack [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ]. These methods aim
to maximize model’s loss subject to constraints of perturbations defined by certain ℓ -norms:
 = arg max‖‖ ≤ ( + , )
where  and  represent the original image and its predicted label, respectively,  is the perturbation
to solve for, and  is the model’s loss function. The perturbation constraint  is measured through an
ℓ-norm—most commonly ℓ . While many works have historically evaluated the robustness of their
∞
model through PGD [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ], it has been shown that “robust” models can often sufer from gradient masking,
causing gradient-based attacks like PGD to fail [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ], and leading to a sense of overestimated robustness.
To overcome this, multiple attacks, including both white- and black-box attacks should be used [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ].
Thus, the AutoAttack ensemble [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] has become the de-facto standard for evaluating robustness.
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Methods</title>
      <p>This section details how we measure the alignment and robustness of machine learning models with the
goal of assessing if more human-like vision models are also more resilient to security vulnerabilities.</p>
      <sec id="sec-3-1">
        <title>3.1. Alignment.</title>
        <p>
          To measure alignment and download candidate models, we leverage the BrainScore [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] library.
BrainScore provides a standardized framework for evaluating model similarity to biological vision through
a set of neural, behavioral, and engineering benchmarks, supplying 106 benchmarks in total. These
benchmarks quantify how closely a model’s internal representations and outputs correspond to
neurophysiological recordings, human psychophysical behavior, and performance on engineered vision tasks.
Neural alignment is measured by comparing activations from DNNs to neural recordings from primate
visual cortex regions (e.g., V1, V2, V4, and IT), using similarity metrics like Representational Similarity
Analysis (RSA) [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ]. Behavioral alignment assesses whether models replicate human psycho-physical
responses in object recognition and perturbation tests, while engineering alignment evaluates model
robustness to controlled distortions, such as contrast reductions, or performance on out-of-distribution
data. We use 105 benchmarks by discarding one of them, which has NaN value for all the selected
models.
        </p>
        <p>
          In total, the BrainScore library has documented benchmark scores for 434 models. Out of those,
there are 240 models available in their registry (the remaining models were either submitted privately
or have been deprecated). From the 240 models in the registry, we further removed an additional 89
models that were either incompatible with ImageNet (e.g., do not output 1000 classes or expect video
streams) or could not be run on RobustBench due to gradient masking. After this, we had to discard an
additional 7 models, which represented all the VOne class models [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ] because they were not able to
run on AutoAttack due to gradient alteration or masking, suggesting that previous results finding that
VOne models are more robust to adversarial examples could have been due to overestimated robustness
and highlighting the importance of evaluating robustness under comprehensive attack strategies. After
this filtering process, we were left with 144 models for our evaluation.
        </p>
        <p>
          Model diversity is critical for evaluating the generalization of alignment-robustness relationships.
Thus, the 144 models selected for our evaluation are representative by covering a broad spectrum of
architectures and training recipes. The majority of architectures are convolutional neural networks
(CNNs) [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ] such as various ResNet [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ] and VGG variants [
          <xref ref-type="bibr" rid="ref22">22</xref>
          ]. We also include more recent
architectural designs such as Vision Transformers (ViTs) [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ], which uses self-attention mechanisms to capture
global dependencies in images, its variants [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ], and hybrid models [
          <xref ref-type="bibr" rid="ref13 ref24">13, 24</xref>
          ] (i.e., a combination of CNNs
and ViTs). This architectural diversity is complemented by models trained with diferent strategies such
as standard supervised learning and self-supervised learning (e.g., contrastive learning [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ]). Notably,
the evaluated models also include those specifically designed or trained with properties relevant to the
study’s central hypothesis: some models have undergone adversarial training [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ], which improves
robustness against adversarial examples, while others use mechanisms to emphasize shape bias [
          <xref ref-type="bibr" rid="ref26 ref27 ref28">26, 27, 28</xref>
          ],
shifting model reliance from texture to shape information. This design has been shown to be more
aligned with human perception. The comprehensive model set allows for a thorough exploration of how
architectural design and training methods impact the relationship between representational alignment
and adversarial robustness.
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Robustness.</title>
        <p>
          To evaluate the robustness of our models, we use AutoAttack [
          <xref ref-type="bibr" rid="ref29 ref6">6, 29</xref>
          ], which serves as the standard for
evaluating the robustness of neural networks due to its strong attack performance and fully automated
parameter-free design. AutoAttack contains 4 attacks: APGD-CE, APGD-DLR, FAB, and Square Attack.
By evaluating on AutoAttack, we are not only evaluating on the most performant attacks, but also
integrating in both white-box attacks and black-box attacks which has been recommended in previous
works to combat reporting overestimated robustness due to gradient masking or obfuscation [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ].
        </p>
        <p>
          To better understand how the relationship between adversarial robustness and alignment changes
as attacks change, we evaluate the ℓ∞ robustness of our models at three diferent epsilon levels:
0.25 , 205.55 , 1
 = { 255 255 } to represent adversaries at diferent capability levels and small, medium, and large
image distortion levels. While these values are typically lower than what would be benchmarked on
platforms such as RobustBench [
          <xref ref-type="bibr" rid="ref29">29</xref>
          ], we choose these values with the goal of having a wide distribution
of robust accuracies to identify separability between models, rather than the goal of bringing the model
down to 0% accuracy as is typically done.
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results</title>
      <p>In this work, we hypothesize that there is a relationship between model robustness and alignment, due
to the inherent similarity of the goals in each of these spaces. Here, we focus on answering the question
are more aligned machine learning models more robust to adversarial examples?</p>
      <p>
        To address this, we conduct a series of experiments. We use the BrainScore library v2.2.4 to measure
alignment [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] and load models. Details on models evaluated can be found in section 3. Once these
models have been loaded and their alignment has been measured across the 105 alignment benchmarks,
we evaluate their robustness using AutoAttack [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] from the TorchAttacks [
        <xref ref-type="bibr" rid="ref30">30</xref>
        ] library v3.5.1. The
ImageNet [31] validation set is used for clean inputs to the model and serves as the starting point to
generate adversarial examples. All experiments are run across 12 A100 GPUs with 40 GB of VRAM
and CUDA version 11.1 or greater. Our code is available for download at https://github.com/kyangl/
alignment-security.
      </p>
      <sec id="sec-4-1">
        <title>4.1. Average Alignment</title>
        <p>We first investigate how well diferent classes of alignment—neural alignment, behavioral alignment,
and engineering task performance—predict model robustness. For each class, we take the average score
across all the benchmarks, giving us a single score for each model in the class. While many works
have typically studied average vision alignment overall (i.e., the average of all the benchmarks across
all classes), it has been shown that this can overemphasize behavioral alignment at the cost of neural
alignment [32].</p>
        <p>We then assessed each model’s robust accuracy against AutoAttack at three diferent values of epsilon
0.25 , 205.55 , 1
 = {0.001, 0.00196, 0.00392} , which corresponds to { 255 255 }, respectively. In Figure 1, we
analyze the average score for neural alignment, behavioral alignment, and engineering task performance
0.6
y0.5
c
rau0.4
c
cA0.3
t
sub0.2
o
R0.1
0.0
performance (shown from the third black bar to the end of the figure) tends to have more stable (and
more positive) correlations as  increases.</p>
        <p>Furthermore, we identified specific benchmarks that are strongly correlated with robust accuracy.
Interestingly, many of these benchmarks that exhibit strong positive correlations with robust accuracy
even at high epsilon values measure, to some degree, a model’s selectivity towards texture information,
meaning that models which are more robust to adversarial examples tend to process texture information
more similarly to humans. Below, we highlight and discuss the benchmarks we found to be most strongly
(positive or negative) correlated with robust accuracy.</p>
        <p>In the neural category, we found the correlation between robust accuracy and alignment to be highly
dependent on the area of visual processing. At small to medium values of  , correlations are strongest
in benchmarks that measure alignment to V4 and IT areas of the visual system. At later values of
 , which represents a stronger attack, FreemanZiemba2013.V2-pls which measures responses to
naturalistic texture stimuli in V2 [33], was the metric with the strongest positive correlation. This
suggests that processing texture more similarly to biological vision systems in later visual areas is even
more important than early visual areas to yield robust models. Interestingly, we found that metrics
utilizing fMRI data were typically the most negatively correlated.</p>
        <p>
          In the behavioral category, we see many strong positive correlations at low values of  . The sets of
benchmarks that stand out the most are the BMD and Baker [34] groups, which measure the ability
to recognize objects by silhouettes when the shape has been distorted. In Baker and Elder [34], it was
found that humans are less able to recognize objects when shape information is distorted than neural
networks are, demonstrating that machine learning models have a certain degree of insensitivity to
changes in shape information and thus don’t rely as heavily on it as humans do for object recognition.
This finding further supports the notion that neural networks may rely more on texture information
rather than shape information [
          <xref ref-type="bibr" rid="ref26">26</xref>
          ]. The strong positive relationship between these benchmarks and
performance on adversarial examples suggests that models’ ability to recognize objects without spatial
relationships actually hurts their overall robustness.
        </p>
        <p>
          Finally, on engineering tasks, we see strong positive correlations across nearly all benchmarks with
low  . However, these correlations decrease as  increases for most benchmarks. The benchmarks
that remain most strongly positively correlated with high  are both related to bias towards shape
over texture information. First is the Geirhos2021cueconflict-top1 benchmark [
          <xref ref-type="bibr" rid="ref26">26</xref>
          ], which
measures the probability of a model classifying an object using shape information rather than texture
via texture-shape conflicted images. The other is the set of benchmarks from [
          <xref ref-type="bibr" rid="ref28">28</xref>
          ]: Hermann 2020 cue
conflict shape _bias and Hermann2020cueconflictshape_match, which similarly measures
the probability of a model classifying an object using shape information and the percentage of the times
the model classifies according to the shape class, rather than texture or other classes. In all, these results
show that when models are able to classify inputs according to their shape information, as humans do,
rather than texture information, they will be more robust to adversarial examples.
        </p>
        <p>Across all categories of visual alignment, we find that models which exhibit a stronger reliance on
shape information and robust processing of texture cues (i.e., more aligned to how humans process
information) tend to be more resilient to adversarial examples. These results suggest that increasing
alignment towards preferring and processing high-level visual features, particularly textures, in the way
biological vision systems do, serves as a fruitful direction for creating more robust and more aligned
models.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.3. Robustness in Alignment Space</title>
        <p>Building on our finding that certain alignment benchmarks (especially those measuring texture
sensitivity) are strongly correlated with adversarial robustness, we next explore whether models with
similar alignment profiles also exhibit similar levels of robust accuracy. The goal of this analysis is to
see if model similarity on alignment benchmarks is predictive of model performance against adversarial
examples, which would support the hypothesis that aligning models under certain metrics results in
better robustness.
2
ten 5
n
o
p
oCm 0
E
N
S
t 5
10</p>
        <p>To compare the similarity of model performance across benchmarks, we utilize t-SNE to reduce the
dimensions of the results and project them into a subspace that can be visualized. In Figure 3, each point
represents one of our 144 models, colored by its robust accuracy under ℓ∞ adversarial perturbations
with  = 0.001 . We observe that the models are distributed across the embedding space with clustering
patterns that are unique to their robustness.</p>
        <p>Analyzing the non-robust models (blue dots), we find that while some cluster in the upper left,
many are scattered broadly throughout the space. This dispersion suggests that poor robustness is not
associated with a single alignment profile. Rather, there are many diferent ways in which models can
fail to defend against adversarial examples. That is, vulnerability appears to be distributed across a range
of misaligned configurations.</p>
        <p>In contrast, the robust models (red dots) form a tight, well-separated cluster. These models not only
share a similar alignment profile, but their separation from the rest of the models suggests that robust
behavior is tied to a specific region of alignment space. This strongly supports the view that certain
alignment characteristics are predictive of robustness.</p>
        <p>These findings ofer a new perspective on adversarial robustness; rather than unique quirks of the
model resulting in vulnerability to adversarial examples, there are specific, isolated properties that allow
models to more robustly process inputs. In other words, it’s not that specific misalignment will lead to
model vulnerability, but rather specific alignment will lead to robustness. Furthermore, these properties
can be measured in alignment metrics, and thus optimized for in order to build more robust models.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Related Work</title>
      <p>
        There has been substantial progress in bridging the representational diferences between humans and
machine learning models over the last few years. Geirhos et al. [35] show that many of the
highperformance models match or exceed human feedforward performance on OOD datasets. Several works
suggest that more human-aligned model architectures may also be more robust. Dapello et al. [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] design
the biologically inspired VOne block to simulate V1 processing, improving robustness to adversarial and
common corruptions. Li et al. [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] regularize models based on neural recordings from mice to increase
robustness and alignment.
      </p>
      <p>
        Furthermore, Subramanian et al. [36] show that diferences in spatial frequency processing explain
both shape bias and adversarial vulnerability. Models with higher human alignment show improved
performance on datasets like ImageNet-A [37]. Other works highlight that models tend to rely on
texture over shape [
        <xref ref-type="bibr" rid="ref26 ref28">26, 28</xref>
        ], a bias that increases susceptibility to natural adversarial samples [38, 39].
      </p>
      <p>Our work expands on these findings with a large-scale empirical study across diverse architectures and
benchmarks. Unlike prior work that focuses on specific models or alignment methods, we systematically
analyze how diferent forms of alignment relate to adversarial robustness, discovering meaningful
connections between model perception and security.</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <p>The author(s) have not employed any Generative AI tools.
[31] O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, Z. Huang, A. Karpathy, A. Khosla,
M. Bernstein, A. C. Berg, L. Fei-Fei, ImageNet Large Scale Visual Recognition Challenge, in: IJCV
2015, arXiv, 2015. URL: http://arxiv.org/abs/1409.0575.
[32] J. Ahlert, T. Klein, F. Wichmann, R. Geirhos, How Aligned are Diferent Alignment Metrics?, 2024.</p>
      <p>URL: http://arxiv.org/abs/2407.07530. doi:10.48550/arXiv.2407.07530.
[33] J. Freeman, C. M. Ziemba, D. J. Heeger, E. P. Simoncelli, J. A. Movshon, A functional and perceptual
signature of the second visual area in primates, Nature Neuroscience 16 (2013) 974–981. URL:
https://www.nature.com/articles/nn.3402.
[34] N. Baker, J. H. Elder, Deep learning models fail to capture the configural nature of human
shape perception, iScience 25 (2022). URL: https://www.sciencedirect.com/science/article/pii/
S2589004222011853.
[35] R. Geirhos, K. Narayanappa, B. Mitzkus, T. Thieringer, M. Bethge, F. A. Wichmann, W. Brendel,
Partial success in closing the gap between human and machine vision, in: 35th Conference on
NeurIPS, NeurIPS, 2021. URL: http://arxiv.org/abs/2106.07411.
[36] A. Subramanian, E. Sizikova, N. J. Majaj, D. G. Pelli, Spatial-frequency channels, shape bias, and
adversarial robustness, in: Conference on Neural Information Processing Systems, NeurIPS, 2023.
[37] I. Sucholutsky, T. L. Grifiths, Alignment with human representations supports
robust few-shot learning, in: Advances in Neural Information Processing Systems,
volume 36, 2023, pp. 73464–73479. URL: https://proceedings.neurips.cc/paper_files/paper/2023/hash/
e8ddc03b001d4c4b44b29bc1167e7fdd-Abstract-Conference.html.
[38] B. Hoak, R. Sheatsley, P. McDaniel, Err on the Side of Texture: Texture Bias on Real Data, in:
2025 IEEE Conference on Secure and Trustworthy Machine Learning (SaTML), IEEE Computer
Society, 2025, pp. 661–680. URL: https://www.computer.org/csdl/proceedings-article/satml/2025/
171100a661/26Vnph1kKxW.
[39] B. Hoak, P. McDaniel, Explorations in Texture Learning, in: ICLR 2024, Tiny Papers Track, 2024.</p>
      <p>URL: http://arxiv.org/abs/2403.09543.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>I.</given-names>
            <surname>Sucholutsky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Muttenthaler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Weller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Peng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bobu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. C.</given-names>
            <surname>Love</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. J.</given-names>
            <surname>Cueva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Grant</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Groen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Achterberg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. B.</given-names>
            <surname>Tenenbaum</surname>
          </string-name>
          ,
          <string-name>
            <surname>K. M. Collins</surname>
            ,
            <given-names>K. L.</given-names>
          </string-name>
          <string-name>
            <surname>Hermann</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Oktar</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Gref</surname>
            ,
            <given-names>M. N.</given-names>
          </string-name>
          <string-name>
            <surname>Hebart</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Cloos</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Kriegeskorte</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Jacoby</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Marjieh</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Geirhos</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Kornblith</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Rane</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Konkle</surname>
          </string-name>
          ,
          <string-name>
            <surname>T. P. O'Connell</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Unterthiner</surname>
            ,
            <given-names>A. K.</given-names>
          </string-name>
          <string-name>
            <surname>Lampinen</surname>
          </string-name>
          ,
          <string-name>
            <surname>K.-R. Müller</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Toneva</surname>
            ,
            <given-names>T. L.</given-names>
          </string-name>
          <string-name>
            <surname>Grifiths</surname>
          </string-name>
          ,
          <source>Getting aligned on representational alignment</source>
          ,
          <year>2024</year>
          . URL: http://arxiv.org/abs/2310.13018. doi:
          <volume>10</volume>
          .48550/arXiv.2310.13018, arXiv:
          <fpage>2310</fpage>
          .13018 [q-bio].
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>J.</given-names>
            <surname>Dapello</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Marques</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Schrimpf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Geiger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Cox</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. J. DiCarlo</surname>
          </string-name>
          ,
          <article-title>Simulating a Primary Visual Cortex at the Front of CNNs Improves Robustness to Image Perturbations</article-title>
          ,
          <source>in: Advances in Neural Information Processing Systems</source>
          , volume
          <volume>33</volume>
          ,
          <string-name>
            <surname>Curran</surname>
            <given-names>Associates</given-names>
          </string-name>
          , Inc.,
          <year>2020</year>
          , pp.
          <fpage>13073</fpage>
          -
          <lpage>13087</lpage>
          . URL: https: //proceedings.neurips.cc/paper/2020/hash/98b17f068d5d9b7668e19fb8ae470841-Abstract.html.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Brendel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Walker</surname>
          </string-name>
          , E. Cobos,
          <string-name>
            <given-names>T.</given-names>
            <surname>Muhammad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Reimer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bethge</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Sinz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Pitkow</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Tolias</surname>
          </string-name>
          ,
          <article-title>Learning from brains how to regularize machines</article-title>
          ,
          <source>in: Advances in Neural Information Processing Systems</source>
          , volume
          <volume>32</volume>
          ,
          <string-name>
            <surname>Curran</surname>
            <given-names>Associates</given-names>
          </string-name>
          , Inc.,
          <year>2019</year>
          . URL: https://proceedings.neurips. cc/paper/2019/hash/70117ee3c0b15a2950f1e82a215e812b-Abstract.html.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>A.</given-names>
            <surname>Madry</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Makelov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Schmidt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Tsipras</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Vladu</surname>
          </string-name>
          , Towards Deep Learning Models Resistant to Adversarial Attacks,
          <year>2019</year>
          . URL: http://arxiv.org/abs/1706.06083.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>M.</given-names>
            <surname>Schrimpf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Kubilius</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Hong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. J.</given-names>
            <surname>Majaj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Rajalingham</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. B.</given-names>
            <surname>Issa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bashivan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Prescott-Roy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Geiger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Schmidt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. L. K.</given-names>
            <surname>Yamins</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. J. DiCarlo</surname>
          </string-name>
          , Brain-Score:
          <article-title>Which Artificial Neural Network for Object Recognition is most Brain-Like?</article-title>
          ,
          <year>2018</year>
          . URL: http://biorxiv.org/lookup/ doi/10.1101/407007. doi:
          <volume>10</volume>
          .1101/407007.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>F.</given-names>
            <surname>Croce</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hein</surname>
          </string-name>
          ,
          <article-title>Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks</article-title>
          ,
          <source>in: Proceedings of the 37th International Conference on Machine Learning, PMLR</source>
          ,
          <year>2020</year>
          . URL: https://proceedings.mlr.press/v119/croce20b.html.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>D. L.</given-names>
            <surname>Yamins</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Hong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Cadieu</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. J. DiCarlo</surname>
          </string-name>
          ,
          <article-title>Hierarchical Modular Optimization of Convolutional Networks Achieves Representations Similar to Macaque IT and Human Ventral Stream</article-title>
          ,
          <source>in: Advances in Neural Information Processing Systems</source>
          , volume
          <volume>26</volume>
          ,
          <string-name>
            <surname>Curran</surname>
            <given-names>Associates</given-names>
          </string-name>
          , Inc.,
          <year>2013</year>
          . URL: https://papers.nips.cc/paper_files/paper/2013/hash/ 9a1756fd0c741126d7bbd4b692ccbd91-Abstract.html.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>N.</given-names>
            <surname>Kriegeskorte</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mur</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. A.</given-names>
            <surname>Bandettini</surname>
          </string-name>
          ,
          <article-title>Representational similarity analysis - connecting the branches of systems neuroscience</article-title>
          ,
          <source>Frontiers in Systems Neuroscience</source>
          <volume>2</volume>
          (
          <year>2008</year>
          ). URL: https: //www.frontiersin.org/journals/systems-neuroscience/articles/10.3389/neuro.06.004.
          <year>2008</year>
          /full.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>S.</given-names>
            <surname>Kornblith</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Norouzi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Hinton</surname>
          </string-name>
          ,
          <source>Similarity of Neural Network Representations Revisited</source>
          ,
          <year>2019</year>
          . URL: http://arxiv.org/abs/
          <year>1905</year>
          .00414. doi:
          <volume>10</volume>
          .48550/arXiv.
          <year>1905</year>
          .
          <volume>00414</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>A.</given-names>
            <surname>Dosovitskiy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Beyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kolesnikov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Weissenborn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Unterthiner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Dehghani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Minderer</surname>
          </string-name>
          , G. Heigold,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gelly</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Uszkoreit</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Houlsby</surname>
          </string-name>
          ,
          <article-title>An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale</article-title>
          ,
          <year>2021</year>
          . URL: http://arxiv.org/abs/
          <year>2010</year>
          .11929. doi:
          <volume>10</volume>
          . 48550/arXiv.
          <year>2010</year>
          .
          <volume>11929</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>L.</given-names>
            <surname>Muttenthaler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Linhardt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Dippel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. A.</given-names>
            <surname>Vandermeulen</surname>
          </string-name>
          , K. Hermann,
          <string-name>
            <given-names>A. K.</given-names>
            <surname>Lampinen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kornblith</surname>
          </string-name>
          ,
          <article-title>Improving neural network representations using human similarity judgments</article-title>
          ,
          <year>2023</year>
          . URL: http://arxiv.org/abs/2306.04507.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12] Y.-A. Cheng,
          <string-name>
            <given-names>I. F.</given-names>
            <surname>Rodriguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Watanabe</surname>
          </string-name>
          , T. Serre,
          <source>RTify: Aligning Deep Neural Networks with Human Behavioral Decisions</source>
          ,
          <year>2024</year>
          . URL: http://arxiv.org/abs/2411.03630. doi:
          <volume>10</volume>
          . 48550/arXiv.2411.03630.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Mao</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.-Y. Wu</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Feichtenhofer</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Darrell</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Xie</surname>
          </string-name>
          ,
          <article-title>A ConvNet for the 2020s</article-title>
          ,
          <source>in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>11976</fpage>
          -
          <lpage>11986</lpage>
          . URL: https://openaccess.thecvf.com/content/CVPR2022/html/Liu_A_
          <article-title>ConvNet_for_ the_2020s_CVPR_2022_paper</article-title>
          .html.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>A.</given-names>
            <surname>Krizhevsky</surname>
          </string-name>
          , I. Sutskever,
          <string-name>
            <given-names>G. E.</given-names>
            <surname>Hinton</surname>
          </string-name>
          ,
          <article-title>ImageNet classification with deep convolutional neural networks</article-title>
          ,
          <source>Communications of the ACM</source>
          <volume>60</volume>
          (
          <year>2017</year>
          ). URL: https://dl.acm.org/doi/10.1145/3065386.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>K.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , S. Ren,
          <string-name>
            <given-names>J.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <article-title>Deep Residual Learning for Image Recognition</article-title>
          ,
          <source>in: CVPR</source>
          <year>2016</year>
          ,
          <year>2015</year>
          . URL: http://arxiv.org/abs/1512.03385.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>N.</given-names>
            <surname>Carlini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Wagner</surname>
          </string-name>
          ,
          <source>Towards Evaluating the Robustness of Neural Networks</source>
          ,
          <year>2017</year>
          . URL: http: //arxiv.org/abs/1608.04644. doi:
          <volume>10</volume>
          .48550/arXiv.1608.04644.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>I. J.</given-names>
            <surname>Goodfellow</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Shlens</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Szegedy</surname>
          </string-name>
          , Explaining and Harnessing Adversarial Examples,
          <year>2015</year>
          . URL: http://arxiv.org/abs/1412.6572. doi:
          <volume>10</volume>
          .48550/arXiv.1412.6572.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>R.</given-names>
            <surname>Sheatsley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Hoak</surname>
          </string-name>
          , E. Pauley,
          <string-name>
            <given-names>P.</given-names>
            <surname>McDaniel</surname>
          </string-name>
          ,
          <article-title>The Space of Adversarial Strategies</article-title>
          ,
          <source>in: 32nd USENIX Security Symposium</source>
          ,
          <year>2023</year>
          . URL: https://www.usenix.org/conference/usenixsecurity23/ presentation/sheatsley.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>A.</given-names>
            <surname>Athalye</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Carlini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Wagner</surname>
          </string-name>
          ,
          <article-title>Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples</article-title>
          ,
          <source>in: Proceedings of the 35th ICML, PMLR</source>
          ,
          <year>2018</year>
          . URL: https://proceedings.mlr.press/v80/athalye18a.html.
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>N.</given-names>
            <surname>Carlini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Athalye</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Papernot</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Brendel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Rauber</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Tsipras</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Goodfellow</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Madry</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kurakin</surname>
          </string-name>
          , On Evaluating Adversarial Robustness,
          <year>2019</year>
          . URL: http://arxiv.org/abs/
          <year>1902</year>
          .06705. doi:
          <volume>10</volume>
          .48550/arXiv.
          <year>1902</year>
          .
          <volume>06705</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <surname>K. O'Shea</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Nash</surname>
          </string-name>
          , An Introduction to Convolutional Neural Networks,
          <year>2015</year>
          . URL: http://arxiv. org/abs/1511.08458.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>K.</given-names>
            <surname>Simonyan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Vedaldi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Zisserman</surname>
          </string-name>
          ,
          <article-title>Deep Inside Convolutional Networks: Visualising Image Classification Models</article-title>
          and
          <string-name>
            <given-names>Saliency</given-names>
            <surname>Maps</surname>
          </string-name>
          ,
          <year>2014</year>
          . URL: http://arxiv.org/abs/1312.6034. doi:
          <volume>10</volume>
          .48550/ arXiv.1312.6034.
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>L.</given-names>
            <surname>Beyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Royer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Markeeva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Anil</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kolesnikov</surname>
          </string-name>
          ,
          <article-title>Knowledge Distillation: A Good Teacher Is Patient</article-title>
          and Consistent,
          <year>2022</year>
          , pp.
          <fpage>10925</fpage>
          -
          <lpage>10934</lpage>
          . URL: https://openaccess.thecvf.com/content/CVPR2022/html/Beyer_Knowledge_Distillation_
          <string-name>
            <surname>A</surname>
          </string-name>
          _ Good_Teacher_Is_Patient_and_Consistent_CVPR_
          <year>2022</year>
          <article-title>_paper</article-title>
          .html.
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>H.</given-names>
            <surname>Touvron</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Cord</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Douze</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Massa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Sablayrolles</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Jegou</surname>
          </string-name>
          ,
          <article-title>Training data-eficient image transformers &amp; distillation through attention</article-title>
          ,
          <source>in: Proceedings of the 38th International Conference on Machine Learning, PMLR</source>
          ,
          <year>2021</year>
          . URL: https://proceedings.mlr.press/v139/touvron21a.html.
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>A.</given-names>
            <surname>Radford</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. W.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Hallacy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ramesh</surname>
          </string-name>
          , G. Goh,
          <string-name>
            <given-names>S.</given-names>
            <surname>Agarwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Sastry</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Askell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Mishkin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Clark</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Krueger</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Sutskever</surname>
          </string-name>
          ,
          <article-title>Learning Transferable Visual Models From Natural Language Supervision</article-title>
          ,
          <source>in: Proceedings of the 38th ICML, PMLR</source>
          ,
          <year>2021</year>
          . URL: https://proceedings.mlr.press/ v139/radford21a.html.
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>R.</given-names>
            <surname>Geirhos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rubisch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Michaelis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bethge</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. A.</given-names>
            <surname>Wichmann</surname>
          </string-name>
          , W. Brendel,
          <article-title>ImageNet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness</article-title>
          , in: ICLR,
          <year>2019</year>
          . URL: http://arxiv.org/abs/
          <year>1811</year>
          .12231.
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>R.</given-names>
            <surname>Geirhos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. R. M.</given-names>
            <surname>Temme</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Rauber</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. H.</given-names>
            <surname>Schütt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bethge</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. A.</given-names>
            <surname>Wichmann</surname>
          </string-name>
          ,
          <article-title>Generalisation in humans and deep neural networks</article-title>
          ,
          <source>in: NeurIPS</source>
          <year>2018</year>
          ,
          <year>2020</year>
          . URL: http://arxiv.org/abs/
          <year>1808</year>
          .08750.
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <surname>K. L. Hermann</surname>
            , T. Chen,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Kornblith</surname>
          </string-name>
          ,
          <article-title>The Origins and Prevalence of Texture Bias in Convolutional Neural Networks</article-title>
          ,
          <source>in: NeurIPS</source>
          <year>2020</year>
          ,
          <year>2020</year>
          . URL: http://arxiv.org/abs/
          <year>1911</year>
          .09071.
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>F.</given-names>
            <surname>Croce</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Andriushchenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Sehwag</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Debenedetti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Flammarion</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Mittal</surname>
          </string-name>
          , M. Hein,
          <article-title>RobustBench: a standardized adversarial robustness benchmark</article-title>
          , in: NeurIPS, arXiv,
          <year>2021</year>
          . URL: http://arxiv.org/abs/
          <year>2010</year>
          .09670.
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>H.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <surname>Torchattacks: A PyTorch Repository for Adversarial Attacks</surname>
          </string-name>
          ,
          <year>2021</year>
          . URL: http://arxiv.org/ abs/
          <year>2010</year>
          .
          <year>01950</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>