<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>IFBiD: Inference-Free Bias Detection</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Ignacio Serna</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Daniel DeAlcala</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Aythami Morales</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Julian Fierrez</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Javier Ortega-Garcia</string-name>
          <email>javier.ortegag@uam.es</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Biometrics and Data Pattern Analytics Lab (BiDA-Lab), Autonomous University of Madrid</institution>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2018</year>
      </pub-date>
      <volume>32</volume>
      <fpage>818</fpage>
      <lpage>833</lpage>
      <abstract>
        <p>This paper is the first to explore an automatic way to detect bias in deep convolutional neural networks by simply looking at their weights, without the model inference for a specific input. Furthermore, it is also a step towards understanding neural networks and how they work. We analyze how bias is encoded in the weights of deep networks through a toy example using the Colored MNIST database and we also provide a realistic case study in gender detection from face images using state-of-the-art methods and experimental resources. To do so, we generated two databases with 36K and 48K biased models each. In the MNIST models we were able to detect whether they presented strong or low bias with more than 99% accuracy, and we were also able to classify between four levels of bias with more than 70% accuracy. For the face models, we achieved 83% accuracy in distinguishing between models biased towards Asian, Black, or Caucasian ethnicity.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Artificial intelligence is generating more and more
expectations. But is it really living up to those expectations? Its use
is being reviewed in all areas, from natural language
processing for virtual assistants, to computer vision for citizen
monitoring systems or medical follow-up
        <xref ref-type="bibr" rid="ref34">(Stone et al. 2016)</xref>
        .
Deep Neural Networks play a key role in the deployment of
machine learning models in these applications. But although
these algorithms achieve impressive prediction accuracies,
their structure makes them very opaque. Data-driven
learning processes make it difficult to control the factors and
understand the information from the input data that actually
drive their decisions. In this environment new efforts are
being devoted to making systems more understandable and
interpretable by humans
        <xref ref-type="bibr" rid="ref21 ref22 ref33 ref6 ref8">(Mahendran and Vedaldi 2015;
Montavon, Samek, and Mu¨ ller 2018; Bau et al. 2020)</xref>
        . More
concretely, new techniques are being developed to understand
and visualize what machine learning models learn
        <xref ref-type="bibr" rid="ref18">(Zeiler
and Fergus 2014; Koh and Liang 2017)</xref>
        , as well as models
that generate text-based explanations of the decisions they
make
        <xref ref-type="bibr" rid="ref27">(Barredo Arrieta et al. 2020; Ortega et al. 2021)</xref>
        .
      </p>
      <p>On the other hand, thanks to adequate public outreach and
debate, more and more investigations are emerging that
un</p>
      <p>ො = 0.99</p>
      <sec id="sec-1-1">
        <title>Model i</title>
        <p>ො = 0.87</p>
      </sec>
      <sec id="sec-1-2">
        <title>Model i</title>
        <p>ො = 0.99</p>
      </sec>
      <sec id="sec-1-3">
        <title>Model i</title>
        <p>
          ො = 0.99
Labels image 3 ={Male, Black}
cover some erratic and biased behaviors of these artificial
intelligence systems. These errors and biases are calling into
question the safety of AI systems, both because of privacy
issues
          <xref ref-type="bibr" rid="ref10 ref24 ref27">(Fierrez, Morales, and Ortega-Garcia 2021)</xref>
          and
unintended side effects
          <xref ref-type="bibr" rid="ref31">(Serna et al. 2020)</xref>
          .
        </p>
        <p>
          One way to build trust into AI systems is to relate their
inner workings to human-interpretable concepts
          <xref ref-type="bibr" rid="ref6">(Bau et al.
2020)</xref>
          . But research is showing that not all representations
in the convolutional layers of a DNN correspond to natural
parts, raising the possibility of a different decomposition of
the world than humans might expect, requiring further study
Labels image 1 ={Male, Caucassian}
0.99 0.87
% = min 0.87 , 0.99
= 0,88
with glasses
Labels image 1 ={Male, Caucassian}
into the exact nature of the learned representations
          <xref ref-type="bibr" rid="ref11">(Yosinski
et al. 2015; Geirhos et al. 2019)</xref>
          .
        </p>
        <p>
          In this regard, bias detection is a major challenge to
ensure trust in machine learning and its applications
          <xref ref-type="bibr" rid="ref26 ref35">(Ntoutsi
et al. 2020; Terhorst et al. 2022)</xref>
          . Recent approaches for
bias detection focus on the analysis of model outcomes or
the visualization of learned features at the data input level
          <xref ref-type="bibr" rid="ref2 ref33 ref33 ref8 ref8">(Alvi, Zisserman, and Nella˚ker 2018; Zhang, Wang, and Zhu
2018)</xref>
          . That is, they are data-bound and need inference to
gain insight (see Fig. 1). We propose a novel approach
focused solely on what the Neural Networks learn (i.e., the
weights of the network), freeing our method from the
pitfalls of possible conflated biases in the considered datasets
used for inference. The main contributions of this work can
be summarized as:
• We propose IFBiD, a novel bias detector trained with
weights of biased and unbiased learned models.
• We analyze how bias is encoded in the weights1 of deep
networks through two different Case Studies in image
recognition: A) digit classification, and B) gender
detection from face biometrics.
• Our results demonstrate that bias can be detected in the
learned weights of Neural Networks. This work opens a
new research line to improve the transparency of these
algorithms.
• We present two novel databases composed by 84K
models trained with different types of biases. These databases
are unique in the field and can be used to further research
on bias analysis in machine learning.
        </p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>Related Works</title>
      <p>
        To the best of our knowledge there are no prior works that
have attempted to detect the bias of a network by modeling it
from learned weights. Existing literature in bias analysis
focuses on the performance (outcome)
        <xref ref-type="bibr" rid="ref11 ref2 ref33 ref7 ref8 ref9">(Bolukbasi et al. 2016;
Alvi, Zisserman, and Nella˚ker 2018; Geirhos et al. 2019;
Chen et al. 2019)</xref>
        and those focused on learned
representations are few
        <xref ref-type="bibr" rid="ref32 ref33 ref8">(Stock and Cisse 2018; Serna et al. 2021)</xref>
        .
      </p>
      <sec id="sec-2-1">
        <title>Bias Explainability</title>
        <p>There is significant work on understanding neural networks
learning processes, which has been useful for diagnosing
CNN representations to gain a deep understanding of the
biased features encoded in a CNN. In general, they map an
abstract concept (e.g. a predicted class) into a domain that
the human can make sense of, e.g. images or text; or they
collect features of the interpretable domain that have
contributed for a given example to produce a decision.</p>
        <p>When an attribute often appears alongside other specific
visual features in training images, the CNN may use these
features to represent the attribute. Thus, features that appear
together, but are not semantically related to the target
attribute, are considered biased representations. Zhang, Wang,
and Zhu (2018) presented a method to discover such
potentially biased representations of a CNN.</p>
        <p>1We used the terms parameters and weights indistinctly to refer
the learned filters of a Neural Network.</p>
        <p>Nagpal et al. (2019) used Class Activation Maps (CAMs
(Zhou et al. 2016)) to obtain the most discriminative regions
of interest for input face images in deep face recognition
models, and observed that activation maps vary significantly
across races.</p>
        <p>
          Also, some investigations show how psychology-inspired
approaches can help elucidate bias in DNNs. Examples
include Ritter et al. (2017);
          <xref ref-type="bibr" rid="ref11">Geirhos et al. (2019)</xref>
          , who found
that CNNs trained on ImageNet exhibit a strong bias towards
recognizing textures rather than shapes or color. This
contrasted sharply with evidence from human behavior, and
revealed fundamentally different classification strategies
between CNNs and humans.
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>Bias Detection</title>
        <p>
          Research on bias analysis focuses largely on detecting
causal connections between attributes in the input data and
outcomes of the learned models
          <xref ref-type="bibr" rid="ref3">(Balakrishnan et al. 2021)</xref>
          .
This kind of research relies primarily on observational
studies where the main conclusions are drawn from
benchmarking the learned models. However, in real life applications, it
is highly difficult to measure the impact of different
covariates on the outcome of a learned model (i.e., it is necessary
to demonstrate that correlation implies causation).
          <xref ref-type="bibr" rid="ref3">Balakrishnan et al. (2021)</xref>
          proposed the use of Generative Models
to develop causal benchmarks applied to face analysis
algorithms. These Generative Models allow manipulation of
attributes in the input data, but as the authors mentioned, the
synthesis methods are far from being fully controllable and
there are still hidden confounders to be considered in these
benchmarks.
        </p>
        <p>
          <xref ref-type="bibr" rid="ref33">Stock and Cisse (2018)</xref>
          used an adversarial example
approach to model critique
          <xref ref-type="bibr" rid="ref14">(Kim, Khanna, and Koyejo 2016)</xref>
          by feeding the model with a carefully hand-selected subset
of examples to subsequently determine whether or not it is
biased.
          <xref ref-type="bibr" rid="ref29">Schaaf et al. (2021)</xref>
          introduced different metrics to
reliably measure several attribution maps’ techniques
(GradCAM, Score-CAM, Integrated Gradients, and LRP- )
capability to detect data biases. Glu¨ge et al. (2020) attempted to
quantify racial bias by clustering the embeddings obtained
from the model, but observed no correlation between
separation in embedding space and bias.
        </p>
        <p>
          Our work goes beyond proposals that seek to model bias
through the observation of the model outcome in response to
particular inputs.
          <xref ref-type="bibr" rid="ref1">Adebayo et al. (2018)</xref>
          already reported the
inconsistency of some widely deployed saliency methods,
as they are not independent of the data. The present work
follows a similar strategy to
          <xref ref-type="bibr" rid="ref32">Serna et al. (2021)</xref>
          , who uses
the information learned from the model to discover bias by
observing the activation of neurons to particular attributes
in the inputs. The present work, however, relies solely on
the information encoded in the model, without looking at
particular input/outputs of the model, thus in an
InferenceFree way, with the significant benefits that this represents
with respect to all previous works.
        </p>
        <p>The hypothesis behind our proposed Inference-Free Bias
Detection (IFBiD) is that bias is encoded in the parameters
of a learned model and it can be detected. IFBiD is an
interesting and noteworthy effort to contribute to tackling the
bias problem. The inference of IFBiD is performed directly
over the weights of a learned model, and therefore does not
require a causal benchmark based on input/output analysis.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Problem Statement and Proposed Approach</title>
      <p>
        In this work we adopt the formulation proposed by
        <xref ref-type="bibr" rid="ref17">Kleinberg et al. (2019)</xref>
        and
        <xref ref-type="bibr" rid="ref31">Serna et al. (2020)</xref>
        for the
discrimination of groups of people, but applicable to any type of bias.
The formalization follows:
      </p>
      <p>Definition 1 (Data). D is a dataset (collection of
multiple samples from different classes) used for training and/or
evaluating a model M. Samples in D can be classified
according to some criterion d. The set Ddc D represents all
the samples corresponding to class c of criterion d.</p>
      <p>Definition 2 (Learned Model). The learned model M is
trained according to input data I D, a Target function T
(e.g., digit classification or gender detection), and a learning
strategy that maximizes a goodness criterion G on that task
(e.g., typically a performance function) based on the output
O of the model and the Target function T for the input data
I.</p>
      <p>Definition 3 (Biased Model). A learned model M is
biased with respect to a specific class c of criterion d if the
goodness G on task T when considering the full set of data
c
D is significantly different to the goodness G(Dd) on the
subset of data corresponding to class c of the criterion d.</p>
      <p>Typically, as in our case, the model M is a Neural
Network ( ), parameterized by , and the goodness-of-fit
criterion consists in minimizing an objective function (e.g. the
cross-entropy loss function).</p>
      <p>The training process of Neural Networks is usually not
deterministic and the resulting parameters depend on several
elements: training data, learning architecture (e.g., number
of layers, number of neurons per layer, etc.), training
hyperparameters (e.g., loss function, number of epochs, batch
size, learning rate, etc.), initialization parameters, and
optimization algorithm.</p>
      <p>
        The existing literature on bias analysis is mainly focused
on the inputs I
        <xref ref-type="bibr" rid="ref33 ref37 ref8">(Tommasi et al. 2017; Zhang, Wang, and
Zhu 2018; Wang, Narayanan, and Russakovsky 2020)</xref>
        and
the outputs O to given inputs
        <xref ref-type="bibr" rid="ref2 ref31 ref33 ref33 ref8 ref8">(Buolamwini and Gebru 2018;
Alvi, Zisserman, and Nella˚ker 2018; Serna et al. 2020)</xref>
        . We
propose a novel approach to detect bias in the learned
parameters , regardless of the particular input I or the output
O = (Ij ) (see Fig. 2).
      </p>
      <sec id="sec-3-1">
        <title>IFBiD: Inference-Free Bias Detection Learning</title>
        <p>The aim of the bias detection model is to find patterns in</p>
        <p>associated with biased outcomes. We designed the bias
detector as a Neural Network ( ) represented by its
parameters .</p>
        <p>In our approach (detailed in next sections), we train the
bias detector using a dataset of biased and unbiased models.</p>
        <p>The models ( j ) for task T , are biased by training them
with biased subsets of the database S1; :::; Sn D. To build
a training set for the detector ( j ), we train a number of</p>
        <p>i i=1;:::;n
models ( j ) with each subset Si, forming f j gj=1;:::;B
(see Fig. 2), where n is the number of biased subsets and
B is the number of models trained with each biased subset.</p>
        <p>ij denotes the jth instance of a learned model trained with
biased subset i.</p>
        <p>
          Because of the non-deterministic nature of the training
process of the network ( j ), the same training subset i is
likely to give rise to different (i.e., ij 6= ik). The reason
for this is that since the solution space is very large, the
solution (which is iteratively approximated) typically arrives
at a local minimum that depends on the initialization, the
particular training configuration, and the order of the data
          <xref ref-type="bibr" rid="ref19 ref21">(LeCun, Bengio, and Hinton 2015)</xref>
          . In CNNs, this translates
into the fact that filters tend to differ between networks. The
bias detector, therefore, has to be able to detect similar
filters in different positions and configurations. The problem is
analogous to detecting patterns in images, where one can be
in different parts of an image.
        </p>
        <p>It is important to underline that the approach requires the
target DNNs (i.e., Input in Fig. 2) to have exactly the same
architecture as the DNNs used to train the detector. This does
not detract from the fact that the task is still challenging,
since, as we have explained before, the filters learned by a
convolutional network never appear in the same place and
are never identical due to the randomness of data
presentation and weight initialization (which is, incidentally, the
reason why we have also used convolutions in the detector
architecture).</p>
        <p>Learned Model  (⋅ȁ Ω)</p>
        <p>Bias Detection Model  (Ω ȁ Θ)
Weights Layer 1
Weights Layer 
Ω1
Ω</p>
        <p>Convolutional Block</p>
        <p>Convolutional Block
Weights Layer 
Ω</p>
        <p>E.g., r×5×5×3</p>
        <p>Convolutional Block
Conv2D
filter1(1,1) Θ1,1,1</p>
        <p>MaxPooling2D
 1 = [ 11,  21, … ,   1]
  = [ 1 ,  2 , … ,    ]
filter1(2,1) Θ1,2,1</p>
        <p>Conv1D</p>
        <p>1-conv module variant. The architecture depends on the number of
layers k of the model</p>
        <p>to be audited. The depth of the module filters depends on the depth of the input weights. Module variant
1
1-conv consists of the subsequent layers: 1
1 convolution followed by d
d MaxPooling, then again a one-dimensional
convolution with kernel size of 1 followed by a MaxPooling with pool size equal to the number of input filters. Do not confuse
the suffixes in this figure (k,i indicates layer) with those in Fig. 2 (j indicates model number).</p>
      </sec>
      <sec id="sec-3-2">
        <title>The Detector</title>
        <p>We evaluated many different learning architectures for
IFBiD. The design of the possible architectures has not only
taken into account the number and types of layers, but has
also depended on the selection of the parameters
of the
model
( j</p>
        <p>) used as input for the detector</p>
        <p>The detector architecture consists of a module to process
the weights/filters of each layer, and then a dense layer that
concatenates all the outputs of each module. The bias
detector architecture consists of multiple modules to process
the weights/filters of each layer (thus one module for each
layer), and then a fully connected layer that concatenates all
the outputs of each module. Fig. 3 shows the general
architecture designed for a specific module variant (see below).
The components of a module are the same for all layers
(convolution, maxpooling, etc.), as well as their order; the only
thing that changes are their parameters, which depend on the
( j ).
size of the input weights.</p>
        <p>We have developed different approaches, in which the
general architecture remains stable, and what changes are
the modules. The module variants we analyzed were the
following (where d</p>
        <p>d is the dimension of the input
filter weights, c is the number of input channels, and r is the
number of input filters):
• MLP: Flatten ! Dense(r)
and there is always 0:1 dropout afterwards (we have seen
that it works best among the values: 0.0, 0.1, 0.2 and 0.3).</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Experiments</title>
      <sec id="sec-4-1">
        <title>Datasets of Biased Models</title>
        <p>for further research.2
We have created two databases for experimenting in
automatic bias detection: DigitWdb and GenderWdb. The
databases contain the weights
of the models
in our experiments for the tasks of digit and gender
classification. The databases include 84K models trained with
different types of biases (each model has an associated label
identifying the bias). These databases are publicly available
( ; j
) used
2https://github.com/BiDAlab/IFBiD/</p>
      </sec>
      <sec id="sec-4-2">
        <title>Case Study A: Digit Classifier (DigitWdb). We have put</title>
        <p>
          together a database that contains the weights of 48K digit
classification networks ( ; j ). For this we have used the
colored MNIST database
          <xref ref-type="bibr" rid="ref15">(Kim et al. 2019)</xref>
          , which consists
of seven replicas of the MNIST database, each with a
different level of color bias (of which we have only used four).
        </p>
        <p>To synthesize the color bias, ten different colors were
selected and assigned to each digit category as its mean color.
Then, for each training image, a color was randomly
sampled from the normal distribution of the corresponding mean
color, and the digit was colorized. The level of bias of each
replica depends on the value of the variance used in the
normal distribution: the lower the more bias.</p>
        <p>The architecture is the same for all models: a CNN ( j )
with three convolutional layers with relu activation, each
followed by a maxpool, and two fully connected layers at the
end (with 128 and 10 neurons, a relu and a softmax
activation function respectively), with a dropout layer of 0.3
between the two. Each of the trained models results in a total
of 50K parameters.</p>
        <p>
          All model parameters have been initialized randomly
with Glorot uniform
          <xref ref-type="bibr" rid="ref12">(Glorot and Bengio 2010)</xref>
          to avoid
possible commonalities. A diagram showing the general
construction of a weight database is shown in Fig. 2. The
composition is as follows:
• Train: 40K models classified by bias level into four
groups, with 10K models per level (B = 10K). The
models were trained using the first 30K training digits from
Colored MNIST. The models have been categorized into
four groups depending on the replica subset with which
they have been trained (n = 4). The level of bias of the
replica subset is what determines the level of bias of the
model. Groups are: very high bias (color jitter variance of
0:02), high bias (color jitter variance of 0:03), low bias
(color jitter variance of 0:04), and very low bias (color
jitter variance of 0:05).
• Test: 8K models classified by bias level into four groups
(2K models for each level). The models were trained
using the last 30K training digits from Colored MNIST and
categorized in the same way as the training ones (i.e.,
from very high bias to very low bias).
        </p>
        <p>Each Colored MNIST biased subset has 60K training
digits, so the 30K for train and 30K for test are independent.
This means that the DigitWdb models assigned to test have
learned with different data than the DigitWdb models
assigned to train.</p>
        <p>Properties. All models have the same architecture and
similar class performance. In this case study, the bias is
determined by the color jitter variance of the digit images.</p>
        <p>Table 1 shows the average digit classification accuracy of
all models trained with the four subsets of different bias
levels, from very low bias (subset S1 with variance of 0:05) to
very high bias (subset S4 with variance of 0:02). The table
shows performance on the Colored MNIST’s test set, that is,
a set of randomly colored numbers, and therefore not biased.</p>
        <p>Although not shown, all models exceeded 99% accuracy
during training (i.e. in their respective training subsets Si).
This means that our models learn as far as the training set
allows. But then, in the (unbiased) test set (i.e., with all
digits colored randomly) the number of correct classifications
drops considerably. The reason is that the color (present in
the training set in a biased way) has been learned as a
differentiating element when classifying digits. Thus, the
network was not only learning to associate a number to a shape,
but also to a color. This is why, subsequently, when finding
a digit with a random color, it has much more difficulty in
classifying it correctly.</p>
        <p>Also, Table 1 shows a clear difference in the performance
of the models as a function of the level of bias of the dataset
with which it has been trained. This difference is the basis
for the experiments carried out in this work.</p>
        <p>Case Study B: Gender Classifier (GenderWdb). We
have gathered together a database that contains the weights
of 36K gender classification networks ( j ). We are
aware there are more gender categories other than male and
female. Since establishing ground-truth genetic sex is not
possible, we use gender as a proxy for sex. We use it as a
simplified application of a real-life based problem.</p>
        <p>
          To train the gender classification models that constitute
this database, we used DiveFace
          <xref ref-type="bibr" rid="ref10 ref24">(Morales et al. 2021)</xref>
          .
DiveFace is a face dataset contaning 24K identities and three
images per identity. Identities are evenly distributed
according to gender: male and female, and three categories related
to ethnic physical characteristics: Asia, African/Indian, and
Caucasian.
        </p>
        <p>As in the DigitWdb database, we have separated the data
for training into two independent sets of the same size. (This
serves to ensure the future independence of the bias detector
training and testing.)</p>
        <p>We have trained 36K gender classification models ( j ),
divided into 30K for training and 6K for testing. The
architecture is the same for all models: a CNN with six
convolutional layers with relu activation, each followed by a
maxpool, and two fully connected layers at the end (with 128
and two neurons, a relu and a softmax activation function
respectively). The result is a model with a total of 100K
parameters.</p>
        <p>
          All model parameters have been initialized randomly
with Glorot uniform
          <xref ref-type="bibr" rid="ref12">(Glorot and Bengio 2010)</xref>
          to avoid
possible commonalities. The composition is as follows:
• Train: 30K models belonging to three classes of bias
(n = 3), depending on the subset Si with which the
model has been trained, with 10K models per class (B =
10K). The models were trained using the first 12K faces
of each ethnic group of DiveFace: S1 is asian biased, S2
is black biased, and S3 is caucasian biased.
• Test: 6K models with the same three types of bias as
train. The models were trained using the last 12K faces
of each ethnic group of DiveFace.
        </p>
        <p>Properties. All models have the same architecture and
similar class performance. Table 2 shows the average
accuracy in gender classification of all models, separated by
bias. Bias has been introduced through the subset Si with
which the model has been trained. In this case study the bias
is determined by the ethnicity of the face images. What
becomes clear from looking at the table is the strong bias in
the performance of the models in each of the groups. Note
that ethnicity attributes include the color of the skin, but also
more complex anthropomorphic face features.</p>
      </sec>
      <sec id="sec-4-3">
        <title>Results</title>
      </sec>
      <sec id="sec-4-4">
        <title>Bias in digit classification models (Case Study A). First of</title>
        <p>all we have attempted a binary classification problem: detect
strong bias against minimal or no bias.</p>
        <p>In this first case we have used models with very high bias
and very low bias. Fig. 4 shows the accuracy of bias
detection in digit classification models ( j ) for the different
architectures given the number of samples the detector ( j )
was trained with. It can be seen that the convolutional
architectures show a saturation of classification performance and
that it does not take many samples to get great performance.
In fact, with the best architectures, 100 training samples are
sufficient to achieve a performance of around 90%. These
initial results suggest that bias is encoded in the weights
of the learned models ( ) and it can be detected.</p>
        <p>A second experiment has been trying to detect the level of
bias of a model ( ), or in other words, to classify the
models according to their level of bias. This is a more complex
problem and has required us to test more architectures for
the detector ( ).</p>
        <p>Fig. 5 shows the classification accuracy of the 4 bias
levels (cf. initial subsection within Experiments describing the
Datasets for Case Study A) for the digit classification
models ( j ) and the different architectures given the number
of samples with which the detector ( j ) was trained. We
see that distinguishing the level of bias in digit classification
models is more complicated than simply stating bias-no bias,
and that in this case the maximum success rate we achieve
in the classification is 70% (note that random chance is 25%
for this task). Another important thing to note is the tendency
(of good architectures) to keep improving as the training set
is increased, they do not seem to be reaching their
performance limit.</p>
      </sec>
      <sec id="sec-4-5">
        <title>Bias in gender classification models (Case Study B).</title>
        <p>After seeing positive results, we made the leap to a more
complex problem (i.e., more covariates): detecting ethnic
bias in models trained for gender recognition.</p>
        <p>Fig. 6 shows the bias classification accuracy of the biased
gender recognition models ( ) for the different
architectures and the number of samples used for training.</p>
        <p>
          The curves show that after a certain number of training
samples, the accuracy is no longer increasing in the model
with more parameters (that containing Conv3D: 1 1
1max). However, it can be seen that when trained with
little data, it performs similarly to the rest. The hypothesis
that best seems to explain this behavior is that, since there
are so many parameters , the solution space is so large
that the choice of a better architectural configuration occurs
automatically, leaving unnecessary parameters unchanged,
as if they were not present
          <xref ref-type="bibr" rid="ref30">(Schmidt, Kraaijveld, and Duin
1992)</xref>
          . With little training data the model ( j ) adjusts
very quickly to those data (losses are practically nil) and in
just a couple of epochs it no longer needs to adjust those
weights . On the other hand, when the number of training
samples increases, it needs to modify more parameters in
order to correlate the training data well, thus losing the
generalization capability equivalent to architectures with fewer
parameters.
        </p>
        <p>The gender classifier ( ) has more layers than the digit
classifier ( ), twice as many. So the bias detection network
( ) for these models has more parameters, and thus the
performance is different. The best performance is obtained with
the same architecture that also obtains the best performance
in the digit models, the architecture with two convolutions:
1 1-conv; reaching 83% detection accuracy. The
improvement in the MLP, that has the most parameters, seems to
be growing steadily. The rest of the architectures seems that
from 15k training samples onwards is when doubling the
samples does not increase its performance so much. But it
would be necessary to keep doubling the number to check if
the trend holds.</p>
        <p>We have dealt with many more architectures that are not
worth describing here: using two dense layers at the end,
adding a dense layer after each convolution, replacing
convolutions with dense layers, playing with dropout, etc.; all
resulting in worse performance than the learning
architectures reported here.</p>
      </sec>
      <sec id="sec-4-6">
        <title>SOTA Comparison</title>
        <p>
          Table 3 shows the comparison with a recent state-of-the-art
bias detection method
          <xref ref-type="bibr" rid="ref32">(Serna et al. 2021)</xref>
          , which consists of
measuring the activation of the last layer upon image input.
We also use an SVM with radial basis function (RBF)
kernel as a baseline, trained in the same way as our detector.
The table shows the percentage of biased models detected
by InsideBias, the RBF SVM, and by our method. The 6K
GenderWdb test models were used, being 2K of each type
of bias.
        </p>
        <p>In order to apply InsideBias, we used 60 images from
DiveFace, with 20 of each ethnicity and the same number
of men and women. Separately, as input to the SVM, we
used the parameters of the models put together as a vector
of length 97K.</p>
        <p>Our method shows considerable superiority: it has a good
hit performance on models with all biases, whereas the other
methods only detect well a single type of bias in the models.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Conclusion</title>
      <p>
        We presented a novel approach called IFBiD (Inference-Free
Bias Detection) to analyze biases in neural networks: by
auditing the models through their weights. Our experiments
demonstrate the existence of identifiable patterns associated
with bias in the weights of a trained Neural Network
        <xref ref-type="bibr" rid="ref35">(Terhorst et al. 2022)</xref>
        . We conducted experiments in two
computer vision use cases: digit and face gender classification
        <xref ref-type="bibr" rid="ref32">(Serna et al. 2021)</xref>
        . This involved generating two databases
with thousands of biased models each. The first, DigitWdb,
with models trained on the Colored MNIST database
        <xref ref-type="bibr" rid="ref15">(Kim
et al. 2019)</xref>
        ; and the second, GenderWdb, with models
trained on a face database, DiveFace
        <xref ref-type="bibr" rid="ref10 ref24">(Morales et al. 2021)</xref>
        .
      </p>
      <p>We used each database to train bias detectors following
the proposed IFBiD principles. We have evaluated a number
of architectures and have found that in both cases it is
possible to achieve a good performance in bias detection. In the
digit classification models we were able to detect whether
they presented strong or low bias with more than 99%
accuracy, and we were also able to classify between four levels
of bias with more than 70% accuracy. For the face models,
we achieved 83% accuracy in distinguishing between
models biased towards Asian, Black, or Caucasian ethnicity. In
both cases the experiments are open-ended in the absence
of increasing both databases. This has been evident in the
plots with the experiments carried out for different sizes of
the training set.</p>
      <p>We evaluated our approach by varying the nature of the
data (i.e., digits and face images), type of architecture (i.e.,
number of layers, units), and optimization strategy (e.g., loss
function). For future work, the generalization capabilities of
our proposed bias detection approach should be studied in
more depth. The training process of the biased models used
for training IFBiD can be affected by hidden confounders
that need to be considered.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>This work has been supported by projects:
TRESPASSETN (MSCA-ITN-2019-860813), PRIMA
(MSCAITN-2019-860315), BIBECA (RTI2018-101248-B-I00
MINECO/FEDER), and BBforTAI
(PID2021-127641OBI00 MICINN/FEDER). I. Serna is supported by a FPI
fellowship from UAM.</p>
      <p>Yosinski, J.; Clune, J.; Nguyen, A.; Fuchs, T.; and Lipson, H.
2015. Understanding Neural Networks Through Deep
Visualization. In Intenational Conference on Machine Learning
(ICML) Deep Learning Workshop. Lille, France.</p>
      <p>Zhou, B.; Khosla, A.; Lapedriza, A.; Oliva, A.; and Torralba,
A. 2016. Learning Deep Features for Discriminative
Localization. In Conference on Computer Vision and Pattern
Recognition (CVPR), 2921–2929. Las Vegas, Nevada, USA:
IEEE.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <surname>Adebayo</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Gilmer</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ; Muelly,
          <string-name>
            <given-names>M.</given-names>
            ;
            <surname>Goodfellow</surname>
          </string-name>
          ,
          <string-name>
            <surname>I.</surname>
          </string-name>
          ; Hardt,
          <string-name>
            <given-names>M.</given-names>
            ; and
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <surname>B.</surname>
          </string-name>
          <year>2018</year>
          .
          <article-title>Sanity Checks for Saliency Maps</article-title>
          .
          <source>In Advances in Neural Information Processing Systems (NIPS)</source>
          , volume
          <volume>31</volume>
          ,
          <fpage>9525</fpage>
          -
          <lpage>9536</lpage>
          . Montre´al, Canada: Curran Associates Inc.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <surname>Alvi</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Zisserman</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ; and Nella˚ker,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <year>2018</year>
          .
          <article-title>Turning a Blind Eye: Explicit Removal of Biases and Variation from Deep Neural Network embeddings</article-title>
          .
          <source>In European Conference on Computer Vision (ECCV)</source>
          ,
          <fpage>556</fpage>
          -
          <lpage>572</lpage>
          . Munich, Germany.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <surname>Balakrishnan</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ; Xiong,
          <string-name>
            <given-names>Y.</given-names>
            ;
            <surname>Xia</surname>
          </string-name>
          ,
          <string-name>
            <surname>W.</surname>
          </string-name>
          ; and Perona,
          <string-name>
            <surname>P.</surname>
          </string-name>
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <article-title>Towards Causal Benchmarking of Bias in Face Analysis Algorithms</article-title>
          .
          <source>In Deep Learning-Based Face Analytics</source>
          ,
          <fpage>327</fpage>
          -
          <lpage>359</lpage>
          . Springer.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          2020.
          <article-title>Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI</article-title>
          .
          <source>Information Fusion</source>
          ,
          <volume>58</volume>
          :
          <fpage>82</fpage>
          -
          <lpage>115</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <surname>Bau</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Zhu</surname>
          </string-name>
          , J.-Y.;
          <string-name>
            <surname>Strobelt</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Lapedriza</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Zhou</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Torralba</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <year>2020</year>
          .
          <article-title>Understanding the Role of Individual Units in a Deep Neural Network</article-title>
          .
          <source>Proceedings of the National Academy of Sciences, 1-8.</source>
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <surname>Bolukbasi</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Chang</surname>
            , K.-W.; Zou,
            <given-names>J.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Saligrama</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Kalai</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <year>2016</year>
          .
          <article-title>Man Is to Computer Programmer as Woman Is to Homemaker? Debiasing Word Embeddings</article-title>
          .
          <source>In Advances in Neural Information Processing Systems</source>
          ,
          <volume>4349</volume>
          -
          <fpage>4357</fpage>
          . Barcelona, Spain.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <surname>Buolamwini</surname>
            , J.; and Gebru,
            <given-names>T.</given-names>
          </string-name>
          <year>2018</year>
          .
          <article-title>Gender Shades: Intersectional Accuracy Disparities in Commercial Gender Classification</article-title>
          . In Friedler, S. A.; and Wilson, C., eds., Conference on Fairness,
          <source>Accountability and Transparency</source>
          , volume
          <volume>81</volume>
          <source>of Proceedings of Machine Learning Research</source>
          ,
          <volume>77</volume>
          -
          <fpage>91</fpage>
          . New York, NY, USA.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Cruz</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Ramsey</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Dickson</surname>
            ,
            <given-names>C. J.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Duca</surname>
            ,
            <given-names>J. S.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Hornak</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Koes</surname>
            ,
            <given-names>D. R.</given-names>
          </string-name>
          ; and Kurtzman,
          <string-name>
            <surname>T.</surname>
          </string-name>
          <year>2019</year>
          .
          <article-title>Hidden Bias in the DUD-E Dataset Leads to Misleading Performance of Deep Learning in Structure-based Virtual Screening</article-title>
          .
          <source>PLOS ONE</source>
          ,
          <volume>14</volume>
          (
          <issue>8</issue>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <surname>Fierrez</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Morales</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Ortega-Garcia</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <year>2021</year>
          .
          <article-title>Encyclopedia of Cryptography, Security</article-title>
          and Privacy, chapter Biometrics Security. Springer.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <surname>Geirhos</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ; Rubisch,
          <string-name>
            <given-names>P.</given-names>
            ;
            <surname>Michaelis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            ;
            <surname>Bethge</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ;
            <surname>Wichmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. A.</given-names>
            ; and
            <surname>Brendel</surname>
          </string-name>
          ,
          <string-name>
            <surname>W.</surname>
          </string-name>
          <year>2019</year>
          .
          <article-title>ImageNet-trained CNNs are Biased Towards Texture; Increasing Shape Bias Improves Accuracy and Robustness</article-title>
          .
          <source>In International Conference on Learning Representations (ICLR)</source>
          . New Orleans, Louisiana, USA.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <surname>Glorot</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ; and Bengio,
          <string-name>
            <surname>Y.</surname>
          </string-name>
          <year>2010</year>
          .
          <article-title>Understanding the Difficulty of Training Deep Feedforward Neural Networks</article-title>
          . In Teh, Y. W.; and Titterington, M., eds.,
          <source>Conference on Artificial Intelligence and Statistics</source>
          , volume
          <volume>9</volume>
          <source>of Proceedings of Machine Learning Research</source>
          ,
          <volume>249</volume>
          -
          <fpage>256</fpage>
          . Chia Laguna Resort, Sardinia, Italy: PMLR.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          2020.
          <article-title>How (Not) to Measure Bias in Face Recognition Networks</article-title>
          .
          <source>In Artificial Neural Networks in Pattern Recognition</source>
          ,
          <fpage>125</fpage>
          -
          <lpage>137</lpage>
          . Springer International Publishing.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Khanna</surname>
          </string-name>
          , R.; and
          <string-name>
            <surname>Koyejo</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          <year>2016</year>
          .
          <article-title>Examples Are Not Enough</article-title>
          , Learn to Criticize!
          <article-title>Criticism for Interpretability</article-title>
          .
          <source>In Advances in Neural Information Processing Systems (NIPS)</source>
          ,
          <fpage>2288</fpage>
          -
          <lpage>2296</lpage>
          . Barcelona, Spain.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Kim</surname>
          </string-name>
          , H.;
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          Learning Not to Learn:
          <article-title>Training Deep Neural Networks With Biased Data</article-title>
          .
          <source>In Conference on Computer Vision and Pattern Recognition (CVPR)</source>
          ,
          <fpage>9012</fpage>
          -
          <lpage>9020</lpage>
          . Las Vegas, Nevada, USA: IEEE.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <string-name>
            <surname>Kleinberg</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Ludwig</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ; Mullainathan,
          <string-name>
            <given-names>S.</given-names>
            ; and
            <surname>Sunstein</surname>
          </string-name>
          ,
          <string-name>
            <surname>C. R.</surname>
          </string-name>
          <year>2019</year>
          .
          <article-title>Discrimination in the Age of Algorithms</article-title>
          .
          <source>Journal of Legal Analysis</source>
          ,
          <volume>10</volume>
          :
          <fpage>113</fpage>
          -
          <lpage>174</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          <string-name>
            <surname>Koh</surname>
            ,
            <given-names>P. W.</given-names>
          </string-name>
          ; and Liang,
          <string-name>
            <surname>P.</surname>
          </string-name>
          <year>2017</year>
          .
          <article-title>Understanding Black-box Predictions via Influence Functions</article-title>
          .
          <source>In International Conference on Machine Learning (ICML)</source>
          , volume
          <volume>70</volume>
          ,
          <fpage>1885</fpage>
          -
          <lpage>1894</lpage>
          . PMLR.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          <string-name>
            <surname>LeCun</surname>
          </string-name>
          , Y.;
          <string-name>
            <surname>Bengio</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ; and Hinton,
          <string-name>
            <surname>G.</surname>
          </string-name>
          <year>2015</year>
          . Deep Learning.
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          <string-name>
            <surname>Nature</surname>
          </string-name>
          ,
          <volume>521</volume>
          (
          <issue>7553</issue>
          ):
          <fpage>436</fpage>
          -
          <lpage>444</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          <string-name>
            <surname>Mahendran</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Vedaldi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <year>2015</year>
          .
          <article-title>Understanding Deep Image Representations by Inverting Them</article-title>
          .
          <source>In Conference on Computer Vision and Pattern Recognition (CVPR)</source>
          ,
          <fpage>5188</fpage>
          -
          <lpage>5196</lpage>
          . Boston, MA, USA: IEEE.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          <string-name>
            <surname>Montavon</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Samek</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ; and Mu¨ller, K.-R.
          <year>2018</year>
          .
          <article-title>Methods for Interpreting and Understanding Deep Neural Networks</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          <source>Digital Signal Processing</source>
          ,
          <volume>73</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          <string-name>
            <surname>Morales</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Fierrez</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Vera-Rodriguez</surname>
            ,
            <given-names>R.</given-names>
            ; and Tolosana, R.
          </string-name>
          <year>2021</year>
          .
          <article-title>SensitiveNets: Learning Agnostic Representations with Application to Face Recognition</article-title>
          .
          <source>IEEE Trans. on Pattern Analysis and Machine Intelligence</source>
          ,
          <volume>43</volume>
          (
          <issue>6</issue>
          ):
          <fpage>2158</fpage>
          -
          <lpage>2164</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          2019.
          <article-title>Deep Learning for Face Recognition: Pride or Prejudiced? arXiv:</article-title>
          <year>1904</year>
          .
          <volume>01219</volume>
          ,
          <fpage>1</fpage>
          -
          <lpage>10</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          <string-name>
            <surname>Ntoutsi</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Fafalios</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Gadiraju</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Iosifidis</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Nejdl</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ; Vidal, M.-E.;
          <string-name>
            <surname>Ruggieri</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Turini</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Papadopoulos</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Krasanakis</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ; et al.
          <year>2020</year>
          .
          <article-title>Bias in Data-driven Artificial Intelligence Systems-An Introductory Survey</article-title>
          .
          <source>Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery</source>
          ,
          <volume>10</volume>
          (
          <issue>3</issue>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          <string-name>
            <surname>Ortega</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Fierrez</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Morales</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ; and Ribeiro,
          <string-name>
            <surname>T.</surname>
          </string-name>
          <year>2021</year>
          .
          <article-title>Symbolic AI for XAI: Evaluating LFIT Inductive Programming for Fair and Explainable Automatic Recruitment</article-title>
          .
          <source>In IEEE/CVF Winter Conf. on Applications of Computer Vision</source>
          Workshops (WACVw).
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          2017.
          <article-title>Cognitive Psychology for Deep Neural Networks: A Shape Bias Case Study</article-title>
          .
          <source>In International Conference on Machine Learning (ICML)</source>
          , volume
          <volume>70</volume>
          ,
          <fpage>2940</fpage>
          -
          <lpage>2949</lpage>
          . Sydney,
          <string-name>
            <surname>NSW</surname>
          </string-name>
          , Australia: PMLR.
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          <string-name>
            <surname>Schaaf</surname>
          </string-name>
          , N.;
          <string-name>
            <surname>de Mitri</surname>
          </string-name>
          , O.;
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>H. B.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Windberger</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Huber</surname>
            ,
            <given-names>M. F.</given-names>
          </string-name>
          <year>2021</year>
          .
          <article-title>Towards Measuring Bias in Image Classification</article-title>
          .
          <source>arXiv preprint arXiv:2107</source>
          .
          <fpage>00360</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          <string-name>
            <surname>Schmidt</surname>
            ,
            <given-names>W. F.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Kraaijveld</surname>
            ,
            <given-names>M. A.</given-names>
          </string-name>
          ; and Duin,
          <string-name>
            <surname>R. P.</surname>
          </string-name>
          <year>1992</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          <string-name>
            <surname>Serna</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Morales</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Fierrez</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ; Cebrian,
          <string-name>
            <surname>N.</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Obradovich;</article-title>
          and
          <string-name>
            <surname>Rahwan</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          <year>2020</year>
          .
          <article-title>Algorithmic Discrimination: Formulation and Exploration in Deep Learning-based Face Biometrics</article-title>
          .
          <source>In AAAI Workshop on Artificial Intelligence Safety (SafeAI)</source>
          . New York, NY, USA.
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          <string-name>
            <surname>Serna</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ;
          <article-title>Pen˜a, A.;</article-title>
          <string-name>
            <surname>Morales</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Fierrez</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <year>2021</year>
          .
          <article-title>InsideBias: Measuring Bias in Deep Networks and Application to Face Gender Biometrics</article-title>
          .
          <source>In IAPR Intl. Conf. on Pattern Recognition (ICPR)</source>
          ,
          <fpage>3720</fpage>
          -
          <lpage>3727</lpage>
          . IEEE.
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          <string-name>
            <surname>Stock</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ; and Cisse,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <year>2018</year>
          .
          <article-title>ConvNets and ImageNet Beyond Accuracy: Understanding Mistakes and Uncovering Biases</article-title>
          .
          <source>In European Conference on Computer Vision (ECCV)</source>
          ,
          <fpage>498</fpage>
          -
          <lpage>512</lpage>
          . Springer International Publishing.
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          <string-name>
            <surname>Stone</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Brooks</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Brynjolfsson</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Calo</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Etzioni</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Hager</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Hirschberg</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ; Kalyanakrishnan,
          <string-name>
            <given-names>S.</given-names>
            ;
            <surname>Kamar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            ;
            <surname>Kraus</surname>
          </string-name>
          ,
          <string-name>
            <surname>S.</surname>
          </string-name>
          ; et al.
          <source>2016. Artificial Intelligence and Life in 2030. One Hundred Year Study on Artificial Intelligence: Report of the 2015-2016 Study Panel</source>
          ,
          <volume>52</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          <string-name>
            <surname>Terhorst</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Kolf</surname>
            ,
            <given-names>J. N.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Huber</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Kirchbuchner</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Damer</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Morales</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Fierrez</surname>
          </string-name>
          , J.; and
          <string-name>
            <surname>Kuijper</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <year>2022</year>
          .
          <article-title>A Comprehensive Study on Face Recognition Biases Beyond Demographics</article-title>
          .
          <source>IEEE Trans. on Technology and Society.</source>
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          2017. A Deeper Look at Dataset Bias,
          <fpage>37</fpage>
          -
          <lpage>55</lpage>
          . Springer International Publishing.
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Narayanan</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Russakovsky</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          <year>2020</year>
          .
          <article-title>REVISE: A Tool for Measuring and Mitigating Bias in Visual Datasets</article-title>
          .
          <source>In European Conference on Computer Vision</source>
          ,
          <fpage>733</fpage>
          -
          <lpage>751</lpage>
          . Springer International Publishing.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>