<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Bias in Machine Learning - What is it Good for?</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Virginia Dignum</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Suna Bensch</string-name>
        </contrib>
      </contrib-group>
      <abstract>
        <p>In public media as well as in scientific publications, the term bias is used in conjunction with machine learning in many different contexts, and with many different meanings. This paper proposes a taxonomy of these different meanings, terminology, and definitions by surveying the, primarily scientific, literature on machine learning. In some cases, we suggest extensions and modifications to promote a clear terminology and completeness. The survey is followed by an analysis and discussion on how different types of biases are connected and depend on each other. We conclude that there is a complex relation between bias occurring in the machine learning pipeline that leads to a model, and the eventual bias of the model (which is typically related to social discrimination). The former bias may or may not influence the latter, in a sometimes bad, and sometime good way.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>INTRODUCTION</title>
      <p>
        Media, as well as scientific publications, frequently report on ‘Bias in
Machine Learning’, and how systems based on AI or machine
learning are ‘sexist’ 3 or ‘discriminatory’ 4 [
        <xref ref-type="bibr" rid="ref11 ref38">10,37</xref>
        ]. In the field of machine
learning, the term bias has an established historical meaning that, at
least on the surface, totally differs from how the term is used in
typical news reporting. Furthermore, even within machine learning, the
term is used in very many different contexts and with very many
different meanings. Definitions are not always given, and if they are, the
relation to other usages of the word is not always clear. Furthermore,
definitions sometimes overlap or contradict each other [
        <xref ref-type="bibr" rid="ref9">8</xref>
        ].
      </p>
      <p>The main contribution of this paper is a proposed taxonomy of
the various meanings of the term bias in conjunction with machine
learning. When needed, we suggest extensions and modifications to
promote a clear terminology and completeness. We argue that this
is more than a matter of definitions of terms. Terminology shapes
how we identify and approach problems, and furthermore how we
communicate with others. This is particularly important in
multidisciplinary work, such as application-oriented machine learning.</p>
      <p>The taxonomy is based on a survey of published research in
several areas, and is followed by a discussion on how different types of
biases are connected and depend on each other.</p>
      <p>Since humans are involved in both the creation of bias, and in
the application of, potentially biased, systems, the presented work
is related to several of the AI-HLEG recommendations for building
Human-Centered AI systems.</p>
      <p>Machine learning is a wide research field with several distinct
approaches. In this paper we focus on inductive learning, which is a
corner stone in machine learning. Even with this specific focus, the
amount of relevant research is vast, and the aim of the survey is not
to provide an overview of all published work, but rather to cover the
wide range of different usages of the term bias.</p>
      <p>This paper is organized as follows. Section 2 briefly summarizes
related earlier work. In Section 3 we survey various sources of bias,
as it appears in the different steps in the machine learning process.
Section 4 contains a survey of various ways of defining bias in the
model that is the outcome of the machine learning process. In
Section 5 we provide a taxonomy of bias, and discuss the different types
of found biases and how they relate to each other. Section 6 concludes
the paper.
2</p>
    </sec>
    <sec id="sec-2">
      <title>RELATED WORK</title>
      <p>
        A number of reviews, with varying focuses related to bias have been
published recently. Barocas and Selbst [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] give as good overview
of various kinds of biases in data generation and preparation for
machine learning. Loftus et al. [
        <xref ref-type="bibr" rid="ref32">31</xref>
        ] review a number of both non
causal and causal notions on fairness, which is closely related to bias.
Suresh and Guttag [
        <xref ref-type="bibr" rid="ref46">45</xref>
        ] identify a number of sources of bias in the
machine learning pipeline. Olteanu et al. [
        <xref ref-type="bibr" rid="ref36">35</xref>
        ] investigate bias and
usage of data from a social science perspective. Our analysis is
complementary to the work cited above, by focusing on bias in conjunction
with machine learning, and by examining a wider range of biases.
We also provide a novel analysis and discussion on the connections
and dependencies between the different types of biases.
3
      </p>
    </sec>
    <sec id="sec-3">
      <title>SOURCES OF BIAS</title>
      <p>Our survey of sources of bias is organized in sections corresponding
to the major steps in the machine learning process (see Figure 1). In
Section 3.1 we describe bias as a major concept in the learning step.
In Section 3.2 we focus on our biased world, which is the source
of information for the learning process. Section 3.3 describes the
plethora of biases related terms used in the data generation process.
3.1</p>
    </sec>
    <sec id="sec-4">
      <title>Bias in learning</title>
      <p>In inductive learning, the aim is to use a data set f(xi; yi)giN=1 to find
a function f (x) such that f (xi) approximates yi in a good way.
Each xi is a vector of features, while y is denoted the output, target
variable or label. An often-discussed case is when machine learning
is used to build decision support systems that outputs
recommendations, for example on whether to accept or deny loan applications in a
bank. The features may in this case be income, property magnitude,
family status, credit history, gender, and criminal record for a loan
applicant, while the output y is a recommendation, 0 or 1.</p>
      <p>Without further restrictions, infinitely many functions perfectly
match any given data set, but most of them are typically useless since
they simply memorize the given data set but generalize poorly for
other data from the same application. Therefore, the search for a good
f has to be limited to a certain space of functions. For example,
the function may be assumed to be linear, which is the assumption in
linear regression. While this may be sufficient in some cases, more
complex function spaces, such as high-order polynomials, or
artificial neural networks are often chosen. Each specific function in is
typically defined by a set of numerical weights.</p>
      <p>
        This preference of certain functions over others was denoted bias
by Tom Mitchell in his paper from 1980 with the title The Need for
Biases in Learning Generalizations [
        <xref ref-type="bibr" rid="ref35">34</xref>
        ], and is a central concept in
statistical learning theory. The expression inductive bias (also known
as learning bias) is used to distinguish it from other types of biases.
      </p>
      <p>In general, inductive learning can be expressed as the
minimization problem
f
= arg min L(f );</p>
      <p>f2
where L(f ) is a costf unction quantifying how well f matches the
data. The most common loss function is defined as</p>
      <p>L(f ) =</p>
      <p>N
X (f (xi)
i=1
yi)2:
(1)
(2)</p>
      <p>The process of solving the minimization problem in Eq. 1 is called
training, and uses a training set f(xi; yi)giN=1 to search through for
the function f that minimizes the loss function in Eq. 2.</p>
      <p>All machine learning techniques for inductive learning (for
example neural networks, support vector machines, and K-nearest
neighbor), need some kind of inductive bias to work, and the choice of
is often a critical design parameter. Having too low inductive bias
( too big) may lead to overfit, causing noise in data to affect the
choice of f . This leads to bad generalization, meaning that f does
not approximate new data that was not used during learning. On the
other hand, having too high inductive bias ( too small) may lead to
underfit, meaning that f approximates both the used data and new
data in an equally poor way.</p>
      <p>
        The learning step includes other types of bias than the inductive
bias described above. Many machine learning algorithms, in
particular within deep learning, contain a large number of hyper-parameters
that are not learned during training but have to be chosen by the
user [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. For neural networks, the choice of number of hidden nodes
and layers and type of activation functions are strictly part of the
definition of , but are often seen as hyper-parameters [
        <xref ref-type="bibr" rid="ref52">51</xref>
        ]. Other
hyper-parameters are related to the way the optimization problem
in Eq. 1 is solved. Besides the choice of algorithm (for example
back propagation, Levenberg-Marquardt, or Gauss-Newton),
learning rate, batch size, number of epochs, and stopping criteria are all
important choices that affect which function f is finally chosen. The
shape of the loss function L(f ) is another hyper-parameter that not
necessarily have to be as in Eq. 2. Examples of other possible choices
are Mean Absolute Error, Hinge Loss, and Cross-entropy Loss. The
choice of meta-parameters usually have a large impact on the
resulting model. We propose to denote this particular type of bias as
hyperparameter bias.
      </p>
      <p>
        The learning step involves more possible sources of bias. One
example is denoted uncertainty bias [
        <xref ref-type="bibr" rid="ref20">19</xref>
        ], and has to do with the
probability values that are often computed together with each produced
classification in a machine learning algorithm5. The probability
represents uncertainty, and typically has to be above a set threshold for a
classification to be considered. For example, a decision support
system for bank loan applications may reject an application although it is
classified as ‘approve’, because the probability is below the
threshold. This threshold is usually manually set, and may create a bias
against underrepresented demographic groups, since less data
normally leads to higher uncertainty.
3.2
      </p>
    </sec>
    <sec id="sec-5">
      <title>A Biased World</title>
      <p>
        The word ‘bias’ has an established normative meaning in legal
language, where it refers to ‘judgement based on preconceived notions
or prejudices, as opposed to the impartial evaluation of facts’ [
        <xref ref-type="bibr" rid="ref9">8</xref>
        ].
The world around us is often described as biased in this sense, and
since most machine learning techniques simply mimic large amounts
of observations of the world, it should come as no surprise that the
resulting systems also express the same bias.
      </p>
      <p>
        This bias of the world is sometimes denoted historical bias.
Historical bias has to do with data that in itself has unwanted properties
that are regarded as biased: ‘Historical bias arises even if the data is
perfectly measured and sampled, if the world as it is or was leads
a model to produce outcomes that are not wanted’ [
        <xref ref-type="bibr" rid="ref46">45</xref>
        ]. The bias of
the world obviously has many dimensions, each one describing some
unwanted aspect of the world.
      </p>
      <p>
        Sometimes, the bias in the world is analyzed by looking at
correlations between features, and between features and the label. The
authors of [
        <xref ref-type="bibr" rid="ref53">52</xref>
        ] show that in a certain data set, the label cooking
co-occurs unproportionally often with woman, as compared to man.
Since most machine learning techniques depend on correlations, such
biases may propagate to learned models or classifiers. The authors
of [
        <xref ref-type="bibr" rid="ref53">52</xref>
        ] show examples of this, and present techniques to detect and
quantify bias related to correlations. They define a subset of output
variables G that reflect a demographic attribute such as gender or
race (e.g. G = fman; womang), and o as a variable that is
potentially correlated with the elements of G (e.g. o = cooking). To
identify unwanted correlations, a bias score for o, with respect to a
demographic variable g 2 G, is defined as
b(o; g) =
      </p>
      <p>c(o; g)
P c(o; g0)
g02G
;
where c(o; g) is the number of occurrences of o and g in a corpus.
If b(o; g) &gt; 1=jjGjj, then o is positively correlated with g, which
indicates that data is biased in this respect. To identify this particular
notion of bias, we propose using the term co-occurrence bias.</p>
      <p>
        A large body of research investigate bias properties of text, at
sentence level, paragraph level, article level, or entire corpora such as
Wikipedia news. In some published work, the word ‘bias’ simply
denotes general, usually unwanted, properties of text [
        <xref ref-type="bibr" rid="ref26 ref41">25, 40</xref>
        ]. Framing
bias refers to how a text expresses a particular opinion on a topic. The
connection between framing bias and gender/race bias is investigated
in [
        <xref ref-type="bibr" rid="ref28">27</xref>
        ], which presents a corpus with sentences expressing negative
5 For many classification algorithms for example K-nearest neighbor,
Artificial Neural Networks, and Na¨ıve Bayes, the output is the probabilities for
each class yi, given the input x: P (yi j x). The predicted class is simply
the class with the highest such probability.
bias towards certain races and genders. Another example of text
related bias is epistemological bias, which refers to the degree of belief
expressed in a proposition. For example, the word claimed expresses
an epistemological bias towards doubts, as compared to stated.
      </p>
      <p>
        As the authors of [
        <xref ref-type="bibr" rid="ref25">24</xref>
        ] conclude, text related bias depends not only
on individual words, but also on the context in which they appear.
The authors use the broader term language bias with reference to
to the guidelines for Neutral Point of View6 (NPOV). Aimed for
Wikipedia editors writing on controversial topics, NPOV suggests
to ‘(i) avoid stating opinions as facts, (ii) avoid stating seriously
contested assertions as facts, (iii) avoid stating facts as opinions,
(iv) prefer nonjudgemental language, and (v) indicate the relative
prominence of opposing views’. Wagner et al. [
        <xref ref-type="bibr" rid="ref48">47</xref>
        ] present and apply
several measures for assessing gender bias in Wikipedia. Coverage
bias is computed by comparing the proportions of notable men and
women that are covered by Wikipedia (the somewhat surprising
result is that this proportion is higher for women).
3.3
      </p>
    </sec>
    <sec id="sec-6">
      <title>Bias in Data Generation</title>
      <p>In machine learning, data generation is responsible for acquiring and
processing observations of the real world, and deliver the resulting
data for learning. Several sub-steps can be identified, each one with
potential bias that will affect the end result. As a first step, the data
of interest has to be specified. The specification guides the
measurement step, which may be automatic sensor based data acquisition, or
manual observations of phenomena of interest. For inductive
learning, data is then usually manually labelled. In the following, possible
sources of bias in each of these sub-steps will be surveyed.
3.3.1</p>
      <sec id="sec-6-1">
        <title>Specification bias</title>
        <p>We propose the term specification bias to denote bias in the choices
and specifications of what constitutes the input and output in a
learning task, i.e.:</p>
        <p>
          The features in the vectors xi in Eq. 2, for example ‘income’,
‘property magnitude’, ‘family status’, ‘credit history’, and
‘gender’ in a decision support system for bank loan approvals.
The output y in Eq. 2, for example ‘approve’ as target variable.
Sometimes the choice of target variable involves creation of new
concepts, such as ‘creditworthiness’, which adds extra bias.
In the case of categorical features and output, discrete classes
related to both x and y, for example ‘low’, ‘medium’, and ‘high’.
These specifications are typically done by the designer of the
system, and require good understanding of the problem, and an ability
to convert this understanding into appropriate entities [
          <xref ref-type="bibr" rid="ref10">9</xref>
          ].
Unintentionally or intentionally biased choices may negatively affect
performance, and also systematically disadvantage protected classes in
systems building on these choices [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ].
        </p>
        <p>
          Related to the selection of features, the notion of proxies deserves
some comments. When designing a decision support system, one
approach to prevent bias with respect to a protected attribute, such as
race, is to simply remove race from the features used for training.
One problem with this approach is that the result may still be biased
with respect to race, if other features are strongly correlated with race
and therefor act as proxies for race in the learning [
          <xref ref-type="bibr" rid="ref15 ref46">14, 45</xref>
          ]. Proxies
for race could, for example, be area code, length, and hairstyle.
3.3.2
        </p>
      </sec>
      <sec id="sec-6-2">
        <title>Measurement bias</title>
        <p>
          In epidemiology, Measurement bias, Observational bias, and
Information bias refer to bias arising from measurement errors [
          <xref ref-type="bibr" rid="ref43">42</xref>
          ], i.e.
errors occurring in the process of making observations of the world.
In the reviewed material on bias and machine learning, such bias was
rarely mentioned, although this process can be biased in very many
ways. In epidemiology and medicine, the data gathering process is
central, and the Dictionary of Epidemiology [
          <xref ref-type="bibr" rid="ref40">39</xref>
          ] lists 37 different
types of biases that may influence data. While most of the listed
biases are specific for medicine and epidemiology, we identified the
following fundamental types of measurement related bias that are
highly relevant also for machine learning. Bias due to instrument
error is a ‘Systematic error due to faulty calibration, inaccurate
measuring instruments, contaminated reagents, incorrect dilution or mixing
of reagents, etc.’. Observer bias is defined as ‘Systematic difference
between a true value and the value actually observed due to observer
variation’. The related investigator bias is defined as ‘Bias on the
part of the investigators of a study toward a particular research
result, exposure or outcome, or the consequences of such bias’. Hence,
a measurement bias can occur either due to the used equipment, or
due to human error or conscious bias.
3.3.3
        </p>
      </sec>
      <sec id="sec-6-3">
        <title>Sampling bias</title>
        <p>
          Sampling bias occurs when there is an underrepresentation or
overrepresentation of observations from a segment of the population [
          <xref ref-type="bibr" rid="ref16">15</xref>
          ].
Such bias, which is sometimes called selection bias [
          <xref ref-type="bibr" rid="ref9">8</xref>
          ], or
population bias [
          <xref ref-type="bibr" rid="ref36">35</xref>
          ], may result in a classifier that performs bad in general,
or bad for certain demographic groups. One example of
underrepresentation is a reported case where a New Zealand passport robot
rejected an Asian man’s eyes because ‘subject eyes are closed’7. A
possible reason could have been that the robot was trained with too
few pictures of Asian men, and therefor made bad predictions on this
demographic group.
        </p>
        <p>
          There are many reasons for sampling bias in a dataset. One kind
is denoted self-selection bias [
          <xref ref-type="bibr" rid="ref16">15</xref>
          ] and can be exemplified with an
online survey about computer use. Such a survey is likely to attract
people more interested in technology than is typical for the entire
population and therefor creates a bias in data. Another example is
a system that predicts crime rates in different parts of a city. Since
areas with more crimes typically have more police present, the
number of reported arrests would become unfairly high in these areas. If
such a system would be used to determine the distribution of police
presence, a viscous circle may even be created [
          <xref ref-type="bibr" rid="ref12 ref42">11, 41</xref>
          ].
        </p>
        <p>
          An opposite example demonstrates how the big data era with its
automatic data gathering can create ‘dark zones or shadows where
some citizens and communities are overlooked’ [
          <xref ref-type="bibr" rid="ref13">12</xref>
          ]. The author
Kate Crawford points to Street Bump, a phone app that uses the
phone’s built in accelerometer to detect and report information about
road problems to the city. Due to the uneven distribution of
smartphones across different parts of the city, data from Street Bump will
have a sampling bias.
        </p>
        <p>
          It is important to note that sampling bias does not only refer to
unbalanced categories of humans, and furthermore not even to
unbalanced categories. Unbalances may also concern features that have
to appear in a balanced fashion. One example is given in [
          <xref ref-type="bibr" rid="ref47">46</xref>
          ], and
is there denoted dataset bias. Focusing on image data, the authors
argue that ‘... computer vision datasets are supposed to be a
representation of the world’, but in reality, many commonly used datasets
6 Wikipedia:Neutral point of view. Accessed July 29, 2020.
represent the world in a very biased way. Objects may, for example,
always appear in the center of the image. This bias makes it hard for a
classifier to recognize objects that are not centered in the image [
          <xref ref-type="bibr" rid="ref47">46</xref>
          ].
The authors compared six common datasets of images used for
object detection, and found that performance on another dataset than
the one used during training in average was cut to less than half. A
similar effect is reported in [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]. If all images in a dataset containing
a snowmobile also contain snow, a machine learning algorithm may
find snow cues useful to detect snowmobiles. While this may work
fine for images in the dataset used for training, it becomes
problematic to analyze images with snowmobiles placed indoors.
        </p>
        <p>
          Another kind of sampling bias is survivorship bias [
          <xref ref-type="bibr" rid="ref16">15</xref>
          ]. It occurs
when the sampled data does not represent the population of interest,
since some data items ‘died’. One example is when a bank’s stock
fund management is assessed by sampling the performance of the
bank’s current funds. This leads to a biased assessment since
poorlyperforming funds are often removed or merged into other funds [
          <xref ref-type="bibr" rid="ref33">32</xref>
          ].
3.3.4
        </p>
      </sec>
      <sec id="sec-6-4">
        <title>Annotator bias</title>
        <p>
          Annotator bias refers to the manual process of labelling data, e.g.
when human annotators assign ‘approve’ or ‘do not approve’ to
each yi to be used to build a classifier for approval of loan
applications [
          <xref ref-type="bibr" rid="ref44">43</xref>
          ]. During this process, the annotators may transfer their
prejudices to the data, and further to models trained with the data.
Sometimes, labelling is not manual, and the annotations are read from the
real world, such as manual decisions for real historical loan
applications. In such cases bias rather falls into the category historical bias
(see Section 3.2).
3.3.5
        </p>
      </sec>
      <sec id="sec-6-5">
        <title>Inherited bias</title>
        <p>
          It is quite common that tools built with machine learning are used to
generate inputs for other machine learning algorithms. If the output
of the tool is biased in any way, this bias may be inherited by
systems using the output as input to learn other models. One example
is if the output of a smile detector based on images is used as
input to a machine learning algorithm. If the smile detection is biased
with respect to age, this bias will propagate into the machine
learning algorithm. We suggest the term inherited bias to refer to this
type of bias. The authors of [
          <xref ref-type="bibr" rid="ref45">44</xref>
          ] identify a number of Natural
Language Processing tasks that may cause such inherited bias: machine
translation, caption generation, speech recognition, sentiment
analysis, language modelling, and word embeddings. For example, tools
for sentiment analysis have been shown to generate different
sentiment for utterances depending on the gender of the subject in the
utterance [
          <xref ref-type="bibr" rid="ref28">27</xref>
          ]. Another example is word embeddings, which are
numerical vector representations of words, learned from data [
          <xref ref-type="bibr" rid="ref34 ref39">33, 38</xref>
          ].
Word embeddings are often used as input to other machine learning
algorithms, and usually provide a powerful way to generalize since
word embeddings for semantically close words are close also in the
word vector space. However, it has been shown that several
common word embeddings are gender biased. The authors in [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] show
that word embeddings trained on Google News articles exhibit
female/male gender stereotypes to a disturbing extent. For example, in
the embedding space, the word ‘nurse’ is closer to ‘female’ than to
‘male’. The authors in [
          <xref ref-type="bibr" rid="ref8">7</xref>
          ] propose a method called WEAT (Word
Embedding Association Test) to measure such bias. Two sets of
socalled target words (e.g. programmer, engineer, scientist &amp; nurse,
teacher, librarian) and two set of so-called attribute words (e.g. man,
male &amp; woman, female) are considered. The null hypothesis is that
(4)
(5)
(6)
(7)
there is no difference between the two sets of target words in terms
of their relative similarity to the two sets of attribute words.
        </p>
        <p>
          Methods that reduce this kind of bias in word embeddings have
been suggested, and either modify already trained word embeddings
[
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] or remove parts of the data used to train the embeddings [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ].
However, bias may still remain [
          <xref ref-type="bibr" rid="ref19">18</xref>
          ] after applying these methods,
and may propagate to models generated by other machine learning
algorithms that rely on word embeddings as input.
4
        </p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>ASSESSING MODEL BIAS</title>
      <p>
        The result from an inductive learning process, i.e. the function f in
Eq. 1), is often referred to as a ‘model’. As described in the previous
sections, bias may propagate from the biased world, through a biased
data generation, to the learning step with its inevitable inductive bias
and other biases. The observed bias of a resulting model is often
simply denoted ‘bias’ [
        <xref ref-type="bibr" rid="ref11 ref12 ref22 ref9">8, 10, 11, 21</xref>
        ]. To distinguish this from other types
of bias discussed in this paper, we propose using the term model bias
to refer to bias as it appears and is analyzed in the final model. An
alternative would be the existing term algorithmic bias [
        <xref ref-type="bibr" rid="ref14 ref7">13</xref>
        ]. However,
typical usage of that term usually refers to the societal effects of
biased systems [
        <xref ref-type="bibr" rid="ref37">36</xref>
        ], while our notion of bias is broader. Nevertheless,
most suggestions on how to define model bias statistically consider
such societal effects: how classification rates differ for groups of
people with different values on a protected attribute such as race, color,
religion, gender, disability, or family status [
        <xref ref-type="bibr" rid="ref22">21</xref>
        ]. As we will see in
the following, classification rates may differ in very many respects,
and a large number of bias types have been defined based on the
condition that should hold for a model not to have that particular type
of bias. For a binary classifier we can for example require that the
overall misclassification rate (OMR) is independent of a certain
protected attribute A (that takes the values 0 or 1). The corresponding
condition for a classifier not being biased in this respect is [
        <xref ref-type="bibr" rid="ref50">49</xref>
        ]:
P (Yb 6= yjA = 0) = P (Yb 6= yjA = 1);
(3)
where Yb is the classifier output f (x) (see Eq. 2), and y is the correct
classification for input x. Both Yb and y take the values 0 or 1. For
example, the fact that a person is female (A = 0) should not increase
or decrease the risk of incorrectly being refused, or allowed, to
borrow money at the bank. Several similar conditions can be defined to
describe other types of unwanted bias in a classifier model [
        <xref ref-type="bibr" rid="ref50">49</xref>
        ]:
false positive rate (FPR):
      </p>
      <p>P (Yb 6= yjA = 0; y = 0) = P (Yb 6= yjA = 1; y = 0);
P (Yb 6= yjA = 0; y = 1) = P (Yb 6= yjA = 1; y = 1);
P (Yb 6= yjA = 0; Yb = 0) = P (Yb 6= yjA = 1; Yb = 0);
false negative rate (FNR):
false omission rate (FOR):
false discovery rate (FDR):</p>
      <p>P (Yb 6= yjA = 0; Yb = 1) = P (Yb 6= yjA = 1; Yb = 1):
Each one of these equations focuses on that an incorrect (Yb 6= y)
classification should be independent of A and a specific value of Yb
or y. The advantageous classifier output (for example being accepted
a loan) is here coded as 1. A classifier that does not satisfy one of
these equations is said to be biased in the corresponding sense8. For
example, a classifier is biased with respect to FDR if the value of A
affects the probability of incorrectly being allowed to borrow money.</p>
      <p>
        A related condition is the equalized odds, which appears in the
literature with slightly different definitions (see [
        <xref ref-type="bibr" rid="ref22">21</xref>
        ] and [
        <xref ref-type="bibr" rid="ref32">31</xref>
        ]). In [
        <xref ref-type="bibr" rid="ref22">21</xref>
        ],
equalized odds is defined by the following two conditions (slightly
modified notation):
and
      </p>
      <p>P (Yb = 1jA = 0; y = 0) = P (Yb = 1jA = 1; y = 0);
P (Yb = 1jA = 0; y = 1) = P (Yb = 1jA = 1; y = 1):
Note that Eq. 8 is equivalent to FPR in Eq. 4, and Eq. 9 is equivalent
to TPR in Eq. 5.</p>
      <p>
        Several other indicators of model bias have been proposed. Loftus
et al. [
        <xref ref-type="bibr" rid="ref32">31</xref>
        ] define Calibration, Demographic Parity/Disparate Impact,
and Individual Fairness. For a binary classification Yb , and a binary
protected group A, demographic parity is defined as follows:
P (Yb = 1jA = 0) = P (Yb = 1jA = 1):
(10)
That is, Yb should be independent of A, such that the classifier in
average gives the same predictions to different groups. If the equality
does not hold, this is referred to as disparate impact. An example
is a software company that wants to reach a better gender balance
among their, mainly male, programmers. By following the principle
of demographic parity, when recruiting, the same proportion of
female applicants as male applicants are hired.
      </p>
      <p>
        Taken all together we conclude that there is a large number of
different types of model biases, each one with its own focus on
unwanted behavior of a classifier. Furthermore, many of these biases are
related, and it can also be shown that several of them are conflicting
in the sense that they cannot be avoided simultaneously [
        <xref ref-type="bibr" rid="ref11 ref29 ref50">10, 28, 49</xref>
        ].
Hence, it is problematic to talk about ‘fair’ or ‘unbiased’ classifiers,
at least without clearly defining the meaning of the terms. It can also
be argued that a proper notion of fairness must be task-specific [
        <xref ref-type="bibr" rid="ref17">16</xref>
        ].
5
      </p>
    </sec>
    <sec id="sec-8">
      <title>TOWARDS A TAXONOMY OF BIAS</title>
      <p>In this section we summarize and discuss the various notions of bias
found in the survey, and propose a taxonomy, illustrated in Figure 1.
5.1</p>
    </sec>
    <sec id="sec-9">
      <title>Terminology</title>
      <p>While it used to be the case that ‘Bias in machine learning’ usually
referred to the inductive bias we describe in Section 3.1, this is no
longer the case. As the survey shows, there is a multitude of usages
with different meanings of bias in the context of machine learning.
We summarize our proposed taxonomy in Figure 1, with different
types of biases organized in the three categories A biased world, Data
generation, and Learning. In several cases the meaning of terms
differed between surveyed papers, and in some cases specific and
important types of biases were only referred to as ‘bias’. In these cases,
we propose descriptive names.</p>
      <p>
        In the Biased world category, the main term is historical bias.
We identify five named types of historical bias. If we define bias
as things that ‘produce outcomes that are not wanted’ [
        <xref ref-type="bibr" rid="ref46">45</xref>
        ], this list
8 In practise, the requirement is usually that the left and right hand side of the
equation should be approximate equal.
(8)
(9)
could of course be made considerably longer. We suggest the term
co-occurrence bias for cases when a word occurs disproportionately
often together with certain other words in texts (see Section 3.2).
      </p>
      <p>In the Data generation category, we found five types of sources
of bias. This list should also not be taken as complete, but rather
as containing some of the most common and representative
examples used in the literature. Several sub-types were also identified (see
Section 3.3). We propose the term specification bias to denote bias in
the specifications of what constitutes the input and output in a
learning task (see Section 3.3.1), and we suggest the term inherited bias
to refer to existing bias in previously computed inputs to a machine
learning algorithm (see Section 3.3.5).</p>
      <p>In the Learning category, we have the classical inductive bias, but
also what we name hyper-parameter bias, the bias caused by, often
manually set, hyper-parameters in the learning step (see Section 3.1).</p>
      <p>We propose using the term model bias to distinguish the bias
detected in the outcome of a machine learning system, from the
possible reasons for this bias. Specific remarks concerning model bias are
presented below.
5.2</p>
    </sec>
    <sec id="sec-10">
      <title>On model bias</title>
      <p>
        A wast majority of published research refer to social discrimination
when talking about bias in machine learning. A typical, and
frequently discussed, example of such model bias is COMPAS, a
computer program used for bail and sentencing decisions. It has been
labeled biased against black defendants [
        <xref ref-type="bibr" rid="ref27">26</xref>
        ] 9.
      </p>
      <p>
        Model bias is caused by bias propagating through the machine
learning pipeline. Bias in the data generation step may, for example,
influence the learned model, as in the previously described example
of sampling bias, with snow appearing in most images of
snowmobiles. This may cause an object classification algorithm to use
irrelevant features as shortcuts when learning to recognize snowmobiles
(in this case snow cues) [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. This in turn leads to a classifier that is
biased against snowmobiles placed indoors, and biased for
snowmobiles placed outdoors. While this, at first, may not be seen as a case of
social discrimination, an owner of a snowmobile shop may feel
discriminated against if Google does not even find the shop’s products
when searching for ‘snowmobiles’.
      </p>
      <p>
        In our survey we identified nine aspects of model bias, defined by
statistical conditions that should hold for a model not being biased
in a specific way. In addition, several causal versions exist. Some of
the identified conditions are contradictory such that any attempt to
decrease one bias will increase another. This is not totally surprising
since the conditions are related to common performance measures
for classifiers, such as precision and recall, which are known to have
the same contradictory relation [22, pp. 405]. The contradictory
conditions is not a statistical peculiarity, but a very real phenomenon.
The COMPAS system mentioned above is indeed biased by certain
conditions, but fair by others10,11 [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
      </p>
      <p>
        In some cases, certain types of bias violates intuitive notions of
fairness, and may even be prohibited by law. One example is
demographic parity (Eq. 10), which aims at classifiers with the same
predictions to different groups. As noted in [
        <xref ref-type="bibr" rid="ref32">31</xref>
        ], this may require
positive discrimination, where individuals having different protected
attributes are treated very differently. In some cases, this may be a
consciously chosen strategy to change societal imbalances, for
example gender balance in certain occupations. However, it would
probably not be seen as a good idea to apply the same reasoning to correct
arrest rates for violent crimes, where men are significantly
overrepresented as a group.
      </p>
      <p>
        Given this complex situation, one should view the different
aspects of model bias as dimensions of a multi dimensional concept.
They should, together with traditional performance measures, be
selected, prioritized and used to guide the design of an optimal, albeit
not necessarily statistically ‘unbiased’ machine learning system. As
noted in [
        <xref ref-type="bibr" rid="ref11">10</xref>
        ], ‘... it is important to bear in mind that fairness itself ...
is a social and ethical concept, not a statistical one’.
      </p>
      <p>
        Most used notions of model bias share a fundamental
shortcoming: they do not take the underlying causal mechanism that generated
data into account. This is serious not least since the legal system
defines discrimination as an identified causal process which is deemed
unfair by society [
        <xref ref-type="bibr" rid="ref22">21</xref>
        ]). Furthermore, the importance of causality in
this context is widely recognized among ethicists and social choice
theorists [
        <xref ref-type="bibr" rid="ref32">31</xref>
        ]. Unfortunately, correlations between observed entities
can alone not be used to identify causal processes without further
assumptions or additional information. Several researchers have
recently developed causal approaches to bias detection. A causal
version of equalized odds, denoted Counterfactual Direct Error Rate,
is proposed in [
        <xref ref-type="bibr" rid="ref51">50</xref>
        ], together with causal versions of other types
of model biases. Causal versions of additional types are suggested
in [
        <xref ref-type="bibr" rid="ref22 ref32">21,31</xref>
        ]. Due to space constraints we will not discuss these further,
although causal reasoning is seen as critical both for identification
and reduction of model bias.
5.3
      </p>
    </sec>
    <sec id="sec-11">
      <title>The world as it should be vs. the world as it is</title>
      <p>It is important, but not always recognized, that most statistical
measures and definitions of model bias, such as Equations 3-9, use the
correct classifications y as baseline when determining whether a
model is biased or not. If y are observations of humans’ biased
decisions in the real world (such as historical loan approvals), or
humans’ biased manual labels created in the data generation process,
Eq. 3 could be perfectly satisfied, which may be interpreted as the
model being free of bias (with respect to overall misclassification
rate). However, a more correct interpretation would be that the model
is no more, or less, biased than the real world. Assessing the ‘true’
degree of biasedness of a model, requires a notion of an ideal ‘world
as it should be’, as opposed to the observed ‘world as it is’.
Demographic parity (Eq. 10) has such a notion built in, namely that the
classifier output should be independent of the protected attribute.</p>
      <p>
        There are at least two fundamentally different approaches to
address the problem with a biased model. We may debias the computed
model, based on an understanding of what ‘the world as it should be’
looks like. For example, word embeddings may be transformed such
that the distance between words describing occupations are
equidistant between gender pairs such as ‘he’ and ‘she’ [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Another
approach to address biased models is to debias the data used to train the
model, for example by removing biased parts, such as suggested for
word embeddings [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], by oversampling [
        <xref ref-type="bibr" rid="ref18">17</xref>
        ], or by resampling [
        <xref ref-type="bibr" rid="ref31">30</xref>
        ].
Debiasing input data can be seen as a technical introduction of (good)
bias in the data generation process, but it can also be seen as an
attempt to model an ideal ‘world as it should be’ rather than the biased
‘world as it is’.
      </p>
      <p>
        The distinction between these two worlds is related to when a
specific model is useful or not. If the model is going to be used to predict
‘the world as it is’, model bias may not be problem. Such a model
may, for example, be used to predict whether a given loan
application will be accepted or not by the bank. Good predictions should
model also biased decisions made by the bank. On the other hand, if
the model is going to be used in a decision support system, we may
want it to mimic ‘the world as it should be’, and bias is then highly
relevant to detect and avoid in the design of the system. One
example of how ‘the world as it should be’ is chosen as norm, is Google’s
image search algorithm. Since only 5% of Fortune 500 CEOs were
women (2018), a search for ‘CEO’ resulted in images of mostly men.
Since then, Google has reportedly changed the algorithm to display
a higher proportion of women [
        <xref ref-type="bibr" rid="ref46">45</xref>
        ].
5.4
      </p>
    </sec>
    <sec id="sec-12">
      <title>What’s wrong with discrimination?</title>
      <p>
        The necessity of inductive bias in machine learning was mentioned
in Section 3.1. The same holds at the level of human learning, as
discussed in the area of philosophical hermeneutics [
        <xref ref-type="bibr" rid="ref24">23</xref>
        ]. In [
        <xref ref-type="bibr" rid="ref21">20</xref>
        ] the
author argues that we always need some form of prejudice (or bias)
to understand and learn about the world. Returning to the example in
Section 3.1, a decision support system for approval of bank loans is
sometimes described as biased and discriminating if it treats certain
groups of people differently. It is important to realize that this
difference in treatment, in a general sense, is inevitable and rather the
main purpose of such a decision support system: to approves some
people’s applications, and reject others12. For example, it may be the
bank’s policy to not approve applications by people with very low
income. While this technically is the same as rejecting people based
on ethnicity, the former may be accepted or even required, while the
latter is often referred to as ‘unwanted’ [
        <xref ref-type="bibr" rid="ref22">21</xref>
        ], ‘racial’ [
        <xref ref-type="bibr" rid="ref44">43</xref>
        ], or
‘discriminatory’ [
        <xref ref-type="bibr" rid="ref11 ref38">10, 37</xref>
        ] (the terms classifier fairness [
        <xref ref-type="bibr" rid="ref11 ref17 ref50">10, 16, 49</xref>
        ] and
demographic parity [
        <xref ref-type="bibr" rid="ref22">21</xref>
        ] are sometimes used in this context). The
difference between features such as ‘income’ and ‘ethnicity’ has to
do with the, already cited, normative meaning of the word bias
expressed as ‘an identified causal process which is deemed unfair by
society’ [
        <xref ref-type="bibr" rid="ref9">8</xref>
        ]. This is further reflected in the notions of protected groups
and protected attributes [
        <xref ref-type="bibr" rid="ref22">21</xref>
        ], which simply define away features such
as ‘income’, while including features that are viewed as important for
equal and fair treatment in our society.
5.5
      </p>
    </sec>
    <sec id="sec-13">
      <title>Fighting bad bias with good</title>
      <p>With the possible exception of inductive bias, the various types of
biases described in this paper are usually used with negative
connotations - to describe unwanted behavior of a machine learning system.
However, several of the types of biases described are not
necessarily bad. For example, some kind of specification bias is necessary to
setup a machine learning task. The alternative would be to observe
everything observable in the real world, which would make learning
extremely hard, if not impossible. The choice of features to include
in the learning constitute a (biased) decision, that may be either good
or bad from the point of view of the bias of the final model.</p>
      <p>
        Likewise, annotator bias is usually regarded as a bad thing, where
human annotators inject their prejudices into the data, for example by
rejecting loan applications in a way that discriminates members of a
certain demographic group. However, there is of course also a
possibility for the human annotators, to consciously or unconsciously,
12 In machine learning this general ability to distinguish between varying
input data is even called ‘discrimination’, but without any negative
connotations (see for example [
        <xref ref-type="bibr" rid="ref30">29</xref>
        ]).
      </p>
      <p>Observations
A BIASED WORLD DATA GENERATION
Historical bias: • Specification bias
• Co-occurrence bias • Annotation bias
• Framing bias • Measurement bias
• Epistemological bias • Sampling bias
• Language bias • Inherited bias
• Coverage bias ...
...</p>
      <p>Data</p>
      <p>Model
LEARNING EVALUATION
• Inductive bias Model bias:
• Hyper-parameter bias • Overall
mis• Uncertainty bias classification rate
... • False positive rate
• False negative rate
• False omission rate
• False discovery rate
• Equalized odds
• Calibration
• Demographic Parity
• Individual Fairness
+ Causal versions
...
inject ‘kindness’ by approving loan applications by the same
members ‘too often’. Depending on the context, this could be described
as a good annotator bias.</p>
      <p>
        Increasing the inductive bias in the learning step can even be
shown to be a general way to reduce an unwanted model bias.
Imposing requirements on f , such as Eq. 3, can be expressed as constrained
minimization [
        <xref ref-type="bibr" rid="ref50">49</xref>
        ] in the inductive learning. Eq. 1 may be rewritten
as
N
X (f (xi)
yi)2:
(11)
f =
arg min
s. t. f2 ;
P (f(x)6=yjA=0)= i=1
      </p>
      <p>P (f(x)6=yjA=1)</p>
      <p>
        While the minimization problems 1 and 11 seem to be identical,
the latter is unfortunately much harder to solve. The constraints are
non convex, as opposed to the normal concave case which can be
solved by several efficient algorithms. The authors in [
        <xref ref-type="bibr" rid="ref50">49</xref>
        ]
approximate the additional constraints such that they can be solved
efficiently by convex-concave programming [
        <xref ref-type="bibr" rid="ref49">48</xref>
        ].
      </p>
      <p>However, the imposed requirements on f can also be seen as
unconstrained minimization over a restricted function space 0
f</p>
      <p>N
= arg min X (f (xi)
f2 0; i=1
yi)2;
(12)
where 0 is the original , with all functions not satisfying the
imposed requirements removed. Hence, in order to decrease unwanted
(bad) model bias, we increase the inductive (good) bias by restricting
the function space appropriately.
6</p>
    </sec>
    <sec id="sec-14">
      <title>FINAL REMARKS</title>
      <p>Our survey and resulting taxonomy show that ‘bias’ used in
conjunction with machine learning can mean very many different things,
even if the most common usage of the word refers to social
discrimination in the behavior of a learned model. Even this specific
meaning of the word deserves careful usage, since it comes in a variety
of types that sometimes even contradict each other. Regarding bias
in the steps leading to a model in the machine learning pipeline, it
may or may not influence the model bias, in a sometimes bad, and
sometimes good way.</p>
      <p>A final remark is that humans are deeply involved in all parts of
the machine learning process illustrated in Figure 1: the biased world,
the data generation process, the learning, and the evaluation of bias
in the final model. Cognitive biases are systematic, usually
undesirable, patterns in human judgment and are studied in psychology
and behavioral economics. They come in a large variety of shades,
and the Wikipedia page13 lists more than 190 different types. Only
a small number of them are directly applicable to machine learning,
but the size of the list suggests caution when claiming that a machine
learning system is ‘non-biased’.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Kristin</given-names>
            <surname>Bechtel Anthony W. Flores</surname>
          </string-name>
          and Christopher T. Lowenkamp, '
          <article-title>False positives, false negatives, and false analyses: A rejoinder to ”machine bias: There's software used across the country to predict future criminals. and it's biased against blacks”'</article-title>
          ,
          <source>Federal Probation Journal</source>
          ,
          <volume>80</volume>
          (
          <issue>2</issue>
          ), (
          <year>September 2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Hyojin</given-names>
            <surname>Bahng</surname>
          </string-name>
          , Sanghyuk Chun, Sangdoo Yun, Jaegul Choo, and Seong Joon Oh, '
          <article-title>Learning de-biased representations with biased representations'</article-title>
          , ArXiv, abs/
          <year>1910</year>
          .02806, (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Solon</given-names>
            <surname>Barocas</surname>
          </string-name>
          and
          <string-name>
            <given-names>Andrew D.</given-names>
            <surname>Selbst</surname>
          </string-name>
          , '
          <article-title>Big data's disparate impact'</article-title>
          ,
          <source>California law Review</source>
          ,
          <volume>671</volume>
          , (
          <year>2014</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Hadrien</given-names>
            <surname>Bertrand</surname>
          </string-name>
          ,
          <article-title>Hyper-parameter optimization in deep learning and transfer learning : applications to medical imaging</article-title>
          , Theses, Universite´ Paris-Saclay,
          <year>January 2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Tolga</given-names>
            <surname>Bolukbasi</surname>
          </string-name>
          ,
          <string-name>
            <surname>Kai-Wei</surname>
            <given-names>Chang</given-names>
          </string-name>
          , James Zou, Venkatesh Saligrama, and Adam Kalai, '
          <article-title>Man is to computer programmer as woman is to homemaker? debiasing word embeddings'</article-title>
          ,
          <source>in Proceedings of the 30th Conference on Neural Information Processing Systems (NIPS</source>
          <year>2016</year>
          ), (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Marc-Etienne</surname>
            <given-names>Brunet</given-names>
          </string-name>
          , Colleen Alkalay-Houlihan,
          <string-name>
            <given-names>Ashton</given-names>
            <surname>Anderson</surname>
          </string-name>
          , and Richard Zemel, '
          <article-title>Understanding the origins of bias inword embeddings'</article-title>
          ,
          <source>in Proc. of the 36th Int. Conf. on Machine Learning</source>
          , (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <article-title>13 Wikipedia List of cognitive biases</article-title>
          .
          <source>Accessed July 29</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Aylin</given-names>
            <surname>Caliskan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Joanna J.</given-names>
            <surname>Bryson</surname>
          </string-name>
          , and Arvind Narayanan, '
          <article-title>Semantics derived automatically from language corpora necessarily contain human biases'</article-title>
          ,
          <source>Science</source>
          ,
          <volume>356</volume>
          (
          <issue>6334</issue>
          ),
          <fpage>183</fpage>
          -
          <lpage>186</lpage>
          , (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Alex</given-names>
            <surname>Campolo</surname>
          </string-name>
          , Madelyn Sanfilippo, Meredith Whittaker, and Kate Crawford, '
          <article-title>Ai now 2017 report'</article-title>
          ,
          <source>in AI Now 2017 Symposium and Workshop</source>
          , (
          <year>January 2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Peter</given-names>
            <surname>Chapman</surname>
          </string-name>
          , Janet Clinton, Randy Kerber, Tom Khabaza, Thomas Reinartz,
          <string-name>
            <given-names>C. Russell H.</given-names>
            <surname>Shearer</surname>
          </string-name>
          , and
          <article-title>Robert Wirth, CRISP-DM 1.0: Step-by-step data mining guide</article-title>
          ,
          <source>SPSS</source>
          ,
          <year>2000</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Alexandra</surname>
            <given-names>Chouldechova</given-names>
          </string-name>
          , '
          <article-title>Fair prediction with disparate impact: A study of bias in recidivism prediction instruments'</article-title>
          ,
          <source>Big data, 5</source>
          <volume>2</volume>
          ,
          <fpage>153</fpage>
          -
          <lpage>163</lpage>
          , (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Ignacio</surname>
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Cofone</surname>
          </string-name>
          , '
          <article-title>Algorithmic discrimination is an information problem'</article-title>
          ,
          <source>Hastings Law Journal</source>
          ,
          <volume>70</volume>
          ,
          <fpage>1389</fpage>
          -
          <lpage>1444</lpage>
          , (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Kate</surname>
            <given-names>Crawford</given-names>
          </string-name>
          , '
          <article-title>Think again: Big data'</article-title>
          ,
          <string-name>
            <surname>Foreign</surname>
            <given-names>Policy</given-names>
          </string-name>
          , (May
          <year>2013</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>David</given-names>
            <surname>Danks</surname>
          </string-name>
          and Alex John London, '
          <article-title>Algorithmic bias in autonomous systems'</article-title>
          ,
          <source>in Proceedings of the 26th International Joint Conference on Artificial Intelligence, IJCAI'17</source>
          , p.
          <fpage>4691</fpage>
          -
          <lpage>4697</lpage>
          . AAAI Press, (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Anupam</surname>
            <given-names>Datta</given-names>
          </string-name>
          , Matt Fredrikson, Gihyuk Ko, Piotr Mardziel, and Shayak Sen, '
          <article-title>Proxy non-discrimination in data-driven systems'</article-title>
          , CoRR, abs/1707.08120, (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>Rice</given-names>
            <surname>University David M. Lane</surname>
          </string-name>
          , '
          <article-title>Chapter 6 research design - sampling bias'</article-title>
          ,
          <source>in Online Statistics Education: A Multimedia Course of Study</source>
          ,
          <volume>235</volume>
          , Rice University.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Cynthia</surname>
            <given-names>Dwork</given-names>
          </string-name>
          , Moritz Hardt, Toniann Pitassi, Omer Reingold, and Richard Zemel, '
          <article-title>Fairness through awareness'</article-title>
          ,
          <source>in Proc. of the 3rd Innovations in Theoretical Computer Science Conference, ITCS '12</source>
          , p.
          <fpage>214</fpage>
          -
          <lpage>226</lpage>
          , New York, NY, USA, (
          <year>2012</year>
          ). Ass. for Comp. Machinery.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Robert</surname>
            <given-names>Geirhos</given-names>
          </string-name>
          , Patricia Rubisch, Claudio Michaelis, Matthias Bethge, Felix A.
          <string-name>
            <surname>Wichmann</surname>
          </string-name>
          , and Wieland Brendel, '
          <article-title>Imagenet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness</article-title>
          .',
          <source>in Int. Conf. on Learning Representations</source>
          , (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>Hila</given-names>
            <surname>Gonen</surname>
          </string-name>
          and Yoav Goldberg, '
          <article-title>Lipstick on a pig: Debiasing methods cover up systematic gender biases in word embeddings but do not remove them'</article-title>
          ,
          <source>in Proc. of the 2019 Conf. of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          , (
          <year>June 2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>Bryce</given-names>
            <surname>Goodman</surname>
          </string-name>
          and Seth Flaxman, '
          <article-title>European union regulations on algorithmic decision-making and a “right to explanation”'</article-title>
          ,
          <source>AI Magazine</source>
          ,
          <volume>38</volume>
          ,
          <fpage>50</fpage>
          -
          <lpage>57</lpage>
          , (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>Gadamer</given-names>
            <surname>Hans-Georg</surname>
          </string-name>
          , Truth and Method, Continuum,
          <year>1975</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [21]
          <string-name>
            <surname>Moritz</surname>
            <given-names>Hardt</given-names>
          </string-name>
          , Eric Price, and Nati Srebro, '
          <article-title>Equality of opportunity in supervised learning'</article-title>
          ,
          <source>in Advances in Neural Information Processing Systems</source>
          <volume>29</volume>
          , eds.,
          <string-name>
            <given-names>D. D.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sugiyama</surname>
          </string-name>
          ,
          <string-name>
            <given-names>U. V.</given-names>
            <surname>Luxburg</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Guyon</surname>
          </string-name>
          , and
          <string-name>
            <given-names>R.</given-names>
            <surname>Garnett</surname>
          </string-name>
          ,
          <volume>3315</volume>
          -
          <fpage>3323</fpage>
          , Curran Associates, Inc., (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>Aboul</given-names>
            <surname>Ella</surname>
          </string-name>
          <string-name>
            <surname>Hassanien</surname>
          </string-name>
          , Aboul Ella Hassanien, and Tarek Gaber,
          <source>Handbook of Research on Machine Learning Innovations and Trends</source>
          , IGI Global, USA, 1st edn.,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [23]
          <string-name>
            <surname>Mireille</surname>
            <given-names>Hildebrandt</given-names>
          </string-name>
          , '
          <article-title>Privacy as protection of the incomputable self: From agnostic to agonistic machine learning'</article-title>
          ,
          <source>Theoretical Inquiries in Law</source>
          ,
          <volume>20</volume>
          (
          <issue>1</issue>
          ),
          <fpage>83</fpage>
          -
          <lpage>121</lpage>
          , (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>Christoph</given-names>
            <surname>Hube</surname>
          </string-name>
          and Besnik Fetahu, '
          <article-title>Detecting biased statements in wikipedia'</article-title>
          ,
          <source>in WWW'18: Companion Proceedings of the The Web Conference</source>
          <year>2018</year>
          , pp.
          <fpage>1779</fpage>
          -
          <lpage>1786</lpage>
          , (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [25]
          <string-name>
            <surname>Christoph</surname>
            <given-names>Hube</given-names>
          </string-name>
          , Besnik Fetahu, and Robert Ja¨schke, '
          <article-title>Towards bias detection in online text corpora'</article-title>
          , in International Workshop on Bias in Information, Algorithms, and
          <string-name>
            <surname>Systems</surname>
          </string-name>
          (BIAS),
          <source>number 2103 in CEUR Workshop Proceedings</source>
          , pp.
          <fpage>19</fpage>
          -
          <lpage>23</lpage>
          , Aachen, (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>Surya</given-names>
            <surname>Mattu Julia Angwin</surname>
          </string-name>
          , Jeff Larson and Lauren Kirchner, '
          <article-title>Machine bias: There's software used across the country to predict future criminals. and it's biased against blacks'</article-title>
          ,
          <source>ProPublica</source>
          , (May
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>Svetlana</given-names>
            <surname>Kiritchenko and Saif M. Mohammad</surname>
          </string-name>
          , '
          <article-title>Examining gender and race bias in two hundred sentiment analysis systems'</article-title>
          , ArXiv, abs/
          <year>1805</year>
          .04508, (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [28]
          <string-name>
            <surname>Jon</surname>
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Kleinberg</surname>
          </string-name>
          , Sendhil Mullainathan, and Manish Raghavan, '
          <article-title>Inherent trade-offs in the fair determination of risk scores'</article-title>
          , CoRR, abs/1609.05807, (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>Tim</given-names>
            <surname>Kovacs and Andy J. Wills</surname>
          </string-name>
          , '
          <article-title>Generalization versus discrimination in machine learning'</article-title>
          ,
          <source>in Encyclopedia of the Sciences of Learning</source>
          , ed.,
          <string-name>
            <surname>Norbert</surname>
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Seel</surname>
          </string-name>
          ,
          <volume>1352</volume>
          -
          <fpage>1355</fpage>
          ,
          <string-name>
            <surname>Springer</surname>
            <given-names>US</given-names>
          </string-name>
          , Boston, MA, (
          <year>2012</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>Yi</given-names>
            <surname>Li</surname>
          </string-name>
          and
          <string-name>
            <given-names>Nuno</given-names>
            <surname>Vasconcelos</surname>
          </string-name>
          , 'Repair:
          <article-title>Removing representation bias by dataset resampling'</article-title>
          ,
          <source>2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</source>
          ,
          <fpage>9564</fpage>
          -
          <lpage>9573</lpage>
          , (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [31]
          <string-name>
            <surname>Joshua</surname>
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Loftus</surname>
            , Chris Russell,
            <given-names>Matt J.</given-names>
          </string-name>
          <string-name>
            <surname>Kusner</surname>
          </string-name>
          , and Ricardo Silva, '
          <article-title>Causal reasoning for algorithmic fairness'</article-title>
          , CoRR, abs/
          <year>1805</year>
          .05859, (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          [32]
          <string-name>
            <surname>Burton</surname>
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>Malkiel</surname>
          </string-name>
          , '
          <article-title>Returns from investing in equity mutual funds 1971 to</article-title>
          <year>1991</year>
          ',
          <source>The Journal of Finance</source>
          ,
          <volume>50</volume>
          (
          <issue>2</issue>
          ),
          <fpage>549</fpage>
          -
          <lpage>572</lpage>
          , (
          <year>1995</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          [33]
          <string-name>
            <surname>Tomas</surname>
            <given-names>Mikolov</given-names>
          </string-name>
          , Kai Chen, Gregory S. Corrado, and
          <string-name>
            <given-names>Jeffrey</given-names>
            <surname>Dean</surname>
          </string-name>
          , '
          <article-title>Efficient estimation of word representations in vector space'</article-title>
          , CoRR, abs/1301.3781, (
          <year>2013</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          [34]
          <string-name>
            <surname>Tom</surname>
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Mitchell</surname>
          </string-name>
          , '
          <article-title>The need for biases in learning generalizations'</article-title>
          ,
          <source>Technical report</source>
          , Rutgers University, New Brunswick, NJ, (
          <year>1980</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          [35]
          <string-name>
            <surname>Alexandra</surname>
            <given-names>Olteanu</given-names>
          </string-name>
          , Carlos Castillo, Fernando Diaz, and Emre Kıcıman, '
          <article-title>Social data: Biases, methodological pitfalls, and ethical boundaries'</article-title>
          ,
          <source>Frontiers in Big Data</source>
          ,
          <volume>2</volume>
          ,
          <fpage>13</fpage>
          , (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          [36]
          <string-name>
            <surname>Trishan</surname>
            <given-names>Panch</given-names>
          </string-name>
          , Heather Mattie, and Rifat Atun, '
          <article-title>Artificial intelligence and algorithmic bias: implications for health systems'</article-title>
          ,
          <source>Journal of global health</source>
          ,
          <volume>9</volume>
          (
          <issue>2</issue>
          ),
          <fpage>010318</fpage>
          -
          <lpage>010318</lpage>
          , (12
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          [37]
          <string-name>
            <surname>Dino</surname>
            <given-names>Pedreshi</given-names>
          </string-name>
          , Salvatore Ruggieri, and Franco Turini, '
          <article-title>Discriminationaware data mining'</article-title>
          ,
          <source>in Proc. of the 14th ACM SIGKDD Int. Conf. on Knowledge Discovery and Data Mining, KDD '08</source>
          , p.
          <fpage>560</fpage>
          -
          <lpage>568</lpage>
          , New York, NY, USA, (
          <year>2008</year>
          ). ACM.
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          [38]
          <string-name>
            <surname>Jeffrey</surname>
            <given-names>Pennington</given-names>
          </string-name>
          , Richard Socher, and
          <string-name>
            <given-names>Christopher D.</given-names>
            <surname>Manning</surname>
          </string-name>
          , 'Glove:
          <article-title>Global vectors for word representation'</article-title>
          ,
          <source>in Empirical Methods in Natural Language Processing (EMNLP)</source>
          , pp.
          <fpage>1532</fpage>
          -
          <lpage>1543</lpage>
          , (
          <year>2014</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref40">
        <mixed-citation>
          [39]
          <string-name>
            <given-names>A</given-names>
            <surname>Dictionary</surname>
          </string-name>
          of Epidemiology, ed.,
          <source>Miquel Porta</source>
          , Oxford University Press, USA,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref41">
        <mixed-citation>
          [40]
          <string-name>
            <surname>Marta</surname>
            <given-names>Recasens</given-names>
          </string-name>
          ,
          <article-title>Cristian Danescu-Niculescu-</article-title>
          <string-name>
            <surname>Mizil</surname>
          </string-name>
          , and Dan Jurafsky, '
          <article-title>Linguistic models for analyzing and detecting biased language'</article-title>
          ,
          <source>in Proceedings of the 51st Annual Meeting of the Association for Computational Linguistics</source>
          , pp.
          <fpage>650</fpage>
          -
          <lpage>1659</lpage>
          . ACL, (
          <year>2013</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref42">
        <mixed-citation>
          [41]
          <string-name>
            <surname>Rashida</surname>
            <given-names>Richardson</given-names>
          </string-name>
          , Jason Schultz, and Kate Crawford, '
          <article-title>Dirty data, bad predictions: How civil rights violations impact police data'</article-title>
          ,
          <source>Predictive Policing Systems, and Justice</source>
          , (
          <year>February 2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref43">
        <mixed-citation>
          [42]
          <string-name>
            <surname>K.J. Rothman</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Greenland</surname>
            , and
            <given-names>T.L.</given-names>
          </string-name>
          <string-name>
            <surname>Lash</surname>
          </string-name>
          , Modern Epidemiology, Wolters Kluwer Health/Lippincott Williams &amp; Wilkins,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref44">
        <mixed-citation>
          [43]
          <string-name>
            <surname>Maarten</surname>
            <given-names>Sap</given-names>
          </string-name>
          , Dallas Card, Saadia Gabriel, Yejin Choi, and
          <string-name>
            <surname>Noah</surname>
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Smith</surname>
          </string-name>
          , '
          <article-title>The risk of racial bias in hate speech detection'</article-title>
          ,
          <source>in Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics</source>
          , pp.
          <fpage>1668</fpage>
          -
          <lpage>1678</lpage>
          , Florence, Italy, (
          <year>July 2019</year>
          ).
          <article-title>Association for Computational Linguistics</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref45">
        <mixed-citation>
          [44]
          <string-name>
            <surname>Tony</surname>
            <given-names>Sun</given-names>
          </string-name>
          , Andrew Gaut, Shirlyn Tang, Yuxin Huang,
          <string-name>
            <surname>Mai</surname>
            <given-names>ElSherief</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jieyu</surname>
            <given-names>Zhao</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Diba</given-names>
            <surname>Mirza</surname>
          </string-name>
          , Elizabeth Belding,
          <string-name>
            <surname>Kai-Wei</surname>
            <given-names>Chang</given-names>
          </string-name>
          , and William Yang Wang, '
          <article-title>Mitigating gender bias in natural language processing: Literature review'</article-title>
          ,
          <source>in Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics</source>
          , pp.
          <fpage>1630</fpage>
          -
          <lpage>1640</lpage>
          . Association for Computational Linguistics, (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref46">
        <mixed-citation>
          [45]
          <string-name>
            <given-names>Harini</given-names>
            <surname>Suresh</surname>
          </string-name>
          and John V. Guttag, '
          <article-title>A framework for understanding unintended consequences of machine learning'</article-title>
          , ArXiv, abs/
          <year>1901</year>
          .10002, (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref47">
        <mixed-citation>
          [46]
          <string-name>
            <given-names>A.</given-names>
            <surname>Torralba</surname>
          </string-name>
          and
          <string-name>
            <given-names>A. A.</given-names>
            <surname>Efros</surname>
          </string-name>
          , '
          <article-title>Unbiased look at dataset bias'</article-title>
          ,
          <source>in Proc. of the 2011 IEEE Conf. on Computer Vision</source>
          and Pattern Recognition,
          <source>CVPR '11</source>
          , p.
          <fpage>1521</fpage>
          -
          <lpage>1528</lpage>
          , USA, (
          <year>2011</year>
          ).
          <source>IEEE Comp. Soc.</source>
        </mixed-citation>
      </ref>
      <ref id="ref48">
        <mixed-citation>
          [47]
          <string-name>
            <surname>Claudia</surname>
            <given-names>Wagner</given-names>
          </string-name>
          , David Garc´ıa, Mohsen Jadidi, and Markus Strohmaier, '
          <article-title>It's a man's wikipedia? assessing gender inequality in an online encyclopedia'</article-title>
          ,
          <string-name>
            <surname>in</surname>
            <given-names>ICWSM</given-names>
          </string-name>
          , (
          <year>2015</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref49">
        <mixed-citation>
          [48]
          <string-name>
            <surname>Shen</surname>
            <given-names>Xinyue</given-names>
          </string-name>
          , Diamond Steven, Gu Yuantao, and Boyd Stephen, '
          <article-title>Disciplined convex-concave programming'</article-title>
          ,
          <source>in 2016 IEEE 55th Conference on Decision and Control (CDC)</source>
          , pp.
          <fpage>1009</fpage>
          -
          <lpage>1014</lpage>
          , (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref50">
        <mixed-citation>
          [49]
          <string-name>
            <given-names>Muhammad</given-names>
            <surname>Bilal</surname>
          </string-name>
          <string-name>
            <surname>Zafar</surname>
          </string-name>
          , Isabel Valera, Manuel Gomez-Rodriguez, and Krishna P. Gummadi, '
          <article-title>Fairness beyond disparate treatment &amp; disparate impact: Learning classification without disparate mistreatment'</article-title>
          ,
          <source>in Proceedings of the 26th International Conference on World Wide Web, WWW</source>
          <year>2017</year>
          , Perth, Australia, April 3-
          <issue>7</issue>
          ,
          <year>2017</year>
          , pp.
          <fpage>1171</fpage>
          -
          <lpage>1180</lpage>
          , (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref51">
        <mixed-citation>
          [50]
          <string-name>
            <given-names>Junzhe</given-names>
            <surname>Zhang</surname>
          </string-name>
          and Elias Bareinboim, '
          <article-title>Equality of opportunity in classification: A causal approach'</article-title>
          ,
          <source>in Proceedings of the 32nd Conference on Neural Information Processing Systems (NIPS</source>
          <year>2018</year>
          ), (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref52">
        <mixed-citation>
          [51]
          <string-name>
            <surname>Xu</surname>
            <given-names>Zhang</given-names>
          </string-name>
          , Xiaocong Chen,
          <string-name>
            <given-names>Lina</given-names>
            <surname>Yao</surname>
          </string-name>
          , Chang Ge, and Manqing Dong, '
          <article-title>Deep neural network hyperparameter optimization with orthogonal array tuning'</article-title>
          ,
          <string-name>
            <surname>in</surname>
            <given-names>ICONIP</given-names>
          </string-name>
          , (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref53">
        <mixed-citation>
          [52]
          <string-name>
            <surname>Jieyu</surname>
            <given-names>Zhao</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tianlu</surname>
            <given-names>Wang</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Mark</given-names>
            <surname>Yatskar</surname>
          </string-name>
          , Vicente Ordonez, and KaiWei Chang, '
          <article-title>Men also like shopping: Reducing gender bias amplification using corpus-level constraints'</article-title>
          ,
          <source>in Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing</source>
          , pp.
          <fpage>2979</fpage>
          -
          <lpage>2989</lpage>
          , (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>