<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>and Ajay Chander</institution>
        </aff>
      </contrib-group>
      <abstract>
        <p>From the data that goes into the AI pipeline to the choice of models and target application, AI safety demands examination at various levels of abstraction. Additionally, the notion of safety has to be assessed across different types of humans-in-theloop involved in the AI pipeline-from AI scientists and software engineers to various types of consumers. Much of the research on AI safety has focused on catering to the needs of AI scientists (such as in design of systems robust to adversarial attacks and ethically grounded algorithms) and consumers (such as in engendering trust and facilitating model interpretation). Choosing the right AI model, tuning various parameters, and processing datasets are some of the many issues that engineers face. A wrong choice in any of these steps can aggravate safety issues in an inconspicuous manner and can harm the interests of the consumer. There is thus a need to provide software engineers with a much more accessible tool whereby they can be better aware of their decisions and the consequences those decisions bear to the consumer. In this paper, we propose a persistence homology based visualization that can aid software engineers in understanding bias in datasets. Unlike other machine learning methods, this topological data analysis method imposes less burden in the sense that the human-inthe-loop does not need to select the right metric or tune parameters, and can determine the bias based on the data before choosing any model. Experiments on the German credit dataset demonstrates the effectiveness of the proposed method in identifying the bias in the dataset due to age. Corresponding Author</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Amidst the massive upsurge of AI-based applications, there
has been growing concern amongst regulatory bodies,
policy makers, and consumers about AI being a black-box
technology. Beyond just AI scientists, today, a variety of
stakeholders are involved in AI based decision pipelines. These
stakeholders could include business executives, software
engineers, and consumers. Thus, successful adoption and safety
of AI systems relies on how much different types of
stakeholders can trust the AI based decision and understand its
functionality.</p>
      <p>Most existing works study AI safety from the perspective
of either an AI scientist or the consumer. Some such efforts
include building systems robust to adversarial attacks,
designing ethically grounded and fair algorithms, and creating
explainable AI models to facilitate model interpretation and to
engender trust in the consumer. However, there are
significant knowledge gaps between an AI scientist who designs a
model, a software engineer who implements and integrates
various models, and a consumer who uses a model for their
custom applications. These gaps could affect safety of AI
systems in inconspicuous ways.</p>
      <p>Consider the role of software engineers in the AI pipeline.
People in these roles are responsible for building software.
These engineers could choose some off-the-shelf AI
modules and integrate them with various APIs and other software
components. They do have to understand various parameters
involved, tune them, and perhaps re-configure various
architectural blocks to suit the requirements of an application. At
the outset, the job of such developers may seem to be mostly
engineering oriented without having to worry about the
consequences of how the model’s decision affects the consumer.
However, this is not the case.</p>
      <p>For one, there are several models available, which is the
right model for a particular application? Next, there are
millions of parameters involved, which ones need to be tuned
and how to set their values? Even before choosing the right
model, the data has to be processed and set into a format
that is amenable for the model to process. The data itself
could be biased or limited. How to ensure accurate data
preprocessing? Such questions have to be carefully examined
and appropriately addressed. Failure to do so can aggravate
safety issues in an inconspicuous manner and can cause
serious consequences to the consumer.</p>
      <p>As an instance, consider prediction of loan defaulting
using AI. A software engineer unaware of bias in training data
may inadvertently use it to train models basing his judgement
merely on validation and test accuracy. Suppose the
training data was biased- say there were too many young people
who defaulted- then the model is likely to predict the same
on test data. This can have serious consequences on a young
applicant who actually may not default. Thus, there are
several safety critical issues that need to be addressed from the
perspective of a software engineer. Engineering trust-worthy
AI software architectures necessitates accessible and
explainable methods that allow software engineers to seamlessly
preprocess the data, select the right model, and integrate various
AI modules into the use cases of their interest. Given the
widespread adoption of AI and the scarcity of people skilled
in AI, there is an even greater need for building such
accessible tools in order to ensure AI safety.</p>
      <p>With the number of biased systems expected to increase
within the next five years, understanding safety in the
context of bias and ensuring fair decisions has been a major area
of interest across several AI based systems used in banking,
insurance, hiring, to name a few. In this paper, we propose
a method based on topological data analysis (TDA) to
enable software engineers to visualize the bias in datasets prior
to even applying any bias mitigation algorithm. Specifically,
we leverage a technique called persistence homology which
can be viewed as a complement to standard feature
representation techniques used in AI. Unlike other feature
representation techniques which require guidance from a machine
learning expert regarding the choice of algorithm, model
architecture and parameters, this technique does not require the
human-in-the-loop to select any metric or tune parameters,
and can work with sparse datasets as well. Experiments on
the German credit dataset demonstrates the effectiveness of
the proposed method in uncovering the bias due to age in the
prediction of loan defaulting. The specific contributions of
the paper can be summarized as follows:
1.1</p>
    </sec>
    <sec id="sec-2">
      <title>Contributions</title>
      <p>We study AI safety from the perspective of software
engineers who form a critical link in the AI pipeline. In
particular, we describe an accessible method by which
software engineers can understand bias in datasets.
We elucidate a novel application of topological data
analysis to quantify bias due to different attributes in a
dataset. Presence of bias is visualized by means of
persistence barcodes and is also validated through
permutation tests (Section 3).</p>
      <p>We demonstrate the effectiveness of the proposed
method in detecting bias due to age on the German credit
dataset (Section 4).</p>
      <p>We provide an accessible guideline to facilitate software
engineers to easily use the method.</p>
      <p>The rest of the paper is organized as follows. A review of
related work is provided in Section 2. Section 3 provides
the details of the method. Results are discussed in Section
4. Section 5 lists some common questions and answers in
order to enhance comprehension about the accessibility of the
approach. Conclusions are provided in Section 6.
2</p>
      <sec id="sec-2-1">
        <title>Related work</title>
        <p>We review related works concerning TDA and its applications
in machine learning. We also review works concerning AI
safety, and bias and accessibility of AI systems.
2.1
TDA is an interdisciplinary field spanning topology and data
analysis, and is used as a tool to uncover patterns in data.
TDA is based on the philosophy that data has shape, and that
shape has meaning. Persistence homology (PH) is a
technique from TDA that can identify clusters, holes, and voids
within a set of points. Persistence homology can be viewed as
a complement to standard feature representation techniques
used in artificial intelligence, and offers the advantage of
being applicable to sparse datasets as well. Unlike other
feature representation techniques which require guidance from
a machine learning expert regarding the choice of algorithm,
model architecture and parameters, this technique does not
require any parameter tuning.</p>
        <p>TDA can be used as an independent tool to uncover
patterns in data or it can also be used in conjunction with
machine learning (ML) techniques as a feature extractor. TDA
has been used in several computer vision applications such as
for shape analysis [Zhou et al., 2017; Wang et al., 2011], for
texture analysis [Zeppelzauer et al., 2018], for medical
imaging [Pachauri et al., 2011], for pose estimation [Nguyen et al.,
2018], and for structure recognition [Li et al., 2014]. It has
also been used in NLP applications for detecting structural
similarity in texts [Zhu, 2013]. TDA can also help in time
series data analysis such as in [Umeda, 2017]. Recently, TDA
has also been used to shed light about the workings of
convolutional neural networks through works such as [Gabrielsson
and Carlsson, 2018; Carlsson and Gabrielsson, 2018] wherein
the authors perform TDA on the weight matrices of the deep
networks to study what is being learnt at each step of
training. In a recent blog post [Gunnar, 2018], it is also mentioned
that TDA on weight matrices of CNNs can be used to
understand dataset variability, correlation between accuracy and
persistence barcodes, etc. However, we have not come across
works that use TDA and PH to understand bias due to various
attributes in a dataset, which is the focus of this work.
2.2</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>AI safety</title>
      <p>The safety of AI models could be compromised in several
ways—a decision may not be ethically justified [Bostrom and
Yudkowsky, 2018], there could be bias in the system
[Srivastava and Rossi, 2018], the system could cause hazardous
effects [Pettigrew et al., 2018], or the privacy and security of
individuals may be at stake [Al-Rubaie and Chang, 2018].
Several works have studied the impact of AI based decisions
in safety critical applications such as healthcare [Challen1 et
al., 2019], judiciary [Kleinberg et al., 2018], transport
[Stilgoe, 2017], finance [FSB, 2017], amongst others. These
studies have examined the impact AI based decisions can have on
various consumers (such as doctors, patients, judges, etc.),
and how the AI model can be made explainable to address
the needs of these consumers. In addition, several other
excellent works [Kurutach et al., 2018; Goodfellow et al., 2014;
Ramakrishan and Shah, 2016] have explored explainable AI
methods to cater to the needs of AI scientists in
understanding the underpinnings of various models. However, there is a
pressing need to address AI safety from the perspective of a
software engineer. This paper is one such effort.
Recognizing the need to develop tools that are accessible
across a broader set of users, IBM released AIFairness360
which is an excellent tool to compute bias along various
metrics [Bellamy et al., 2018]. Accenture also released a similar
fairness assessment tool [Peters, 2018]. It was also reported
that Microsoft is creating an oracle to catch biased AI
algorithms [Knight, 2018]. Google introduced AI bias
visualization with the What-If tool and TensorBoard [Wexler, 2018].
There have also been several academic works in this area.
In a recent paper, MIT researchers detailed what they call as
a toolbox for helping machine learning engineers figure out
what questions to ask of their data in order to diagnose why
their systems may be making unfair predictions [Chen et al.,
2018]. Guidelines have also been proposed to reduce the
potential for bias in AI. These include “factsheets for datasets”
from IBM, and “Datasheets for Datasets”, an approach for
sharing essential information about datasets used to train AI
models [Gebru et al., 2018]. In this work, we propose a
complementary perspective of analyzing bias using topological
data analysis. This method offers the advantage of being
applicable across sparse datasets and can be used for efficient
feature representations and visualizations. Furthermore, a
software engineer does not have to tune parameters or deal
with metrics of evaluation, thereby enhancing accessibility.
3</p>
      <sec id="sec-3-1">
        <title>Methodology</title>
        <p>We begin by reviewing some necessary definitions. More
details about the same can be found in any book on algebraic
topology such as [Weintraub, 2014].
3.1</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Definitions</title>
      <p>Point cloud: A point cloud is often defined as a finite set
of points in some Euclidean space, but may be taken to
be any finite metric space.</p>
      <p>Simplex: A simplex is a generalization of a triangle or a
tetrahedron to their higher dimensional counterparts.
Simplicial complex: A simplicial complex is a
combination of simplexes such that any face (subset) of a simplex
from K is also in K, and the intersection of any two
simplices in K is either empty or shares faces.</p>
      <p>Vietoris-Rips complex: Also known as the Rips
complex, this is a simplicial complex with radius r that
consists of the set of all points (and simplicial complexes)
such that the largest Euclidean distance between any of
its points is at most 2r.</p>
      <p>A natural question may then arise as to what is the best value
of r to use for a dataset. The answer is provided by
persistence homology, which is defined next.</p>
      <p>Persistence Homology (PH): This is a method for
computing topological features of a space at different
spatial resolutions. Such topological features could include
clusters, holes and voids in a dataset. More persistent
topological features are detected over a wide range of
spatial scales and are deemed more likely to represent
true topological features of the underlying space rather
than artifacts of sampling, noise, or other factors. For
example, in a 2D space, as the radius r is gradually
increased, points that are initially disconnected could get
connected and higher order topological features such as
holes could appear. The holes could later disappear as
the entire space gets connected. The process of varying
the radius is referred to as “filtration”. The best value
of r is one that can reveal persistent topological features
in the dataset. This value is automatically computed by
PH softwares.</p>
      <p>Persistence diagrams: The appearance and
disappearance of clusters, holes and other such topological
features can be captured by means of persistence diagrams
(bottom right in Figure 2) and persistence barcodes.
Persistence diagram is a plot of the birth time (i.e., the value
of the radius at which a topological feature appears) and
death time (i.e., the value of the radius at which a
topological feature disappears) of a topological feature as the
the radius is varied. Any point on the diagonal of this
plot is insignificant as it does not persists long enough
(i.e. it disappears soon after it appears). Points above
the diagonal are topological features that persist.</p>
      <p>Persistence barcodes: This captures the interval
between the birth and death of a topological feature. It
is another way of looking at the persistence diagram.
With the above background, we are now in a position to
understand the intuition behind using PH in analyzing dataset
bias.
3.2</p>
    </sec>
    <sec id="sec-5">
      <title>Intuition</title>
      <p>Persistent barcodes as computed by PH is a collection of
intervals along various dimensions. In dimension=0, the
barcode output reflects the decomposition of the dataset into
clusters or components. In clustering, a threshold is
chosen, and any two points are connected by an edge if their
distance is less than this threshold. As the threshold grows, more
points will be connected, and there will be fewer clusters. The
barcode is a way of tracking this behavior.</p>
      <p>The author in [Gunnar, 2018] nicely illustrates the intuition
behind barcodes. We leverage a similar example to explain
the concept of barcodes. Consider some toy data as shown in
Figure 1. On the left, we see a dataset that consists of two
clusters close to each other and on the right we see another
dataset that consists of two clusters which are relatively
farther apart. The corresponding barcodes beneath each dataset
represents two lines, one longer than the other. The presence
of two lines indicates that there are two clusters in both these
datasets. However, we notice that in the left dataset, the lines
are shorter compared to the ones on the right. This is because,
the clusters on the left are closer to each other, as a result, the
two initial clusters get merged into a single large cluster and
there is only one line (top one) after the merger happens. For
the dataset on the right, this merger happens little later at a
larger value of radius r as the clusters are farther apart.</p>
      <p>The aforementioned illustration is for dimension =0,
wherein PH captures connected components or clusters. The
length of the barcode is indicative of how well connected the
clusters are, and the number of barcodes is indicative of the
number of such connected components. For higher
dimensions, the barcodes captures the presence of holes, voids, etc.
Further, lengths of barcodes are used as indicators of
variations in training data. For example, the authors in [Carlsson
and Gabrielsson, 2018] use the length of barcodes as
indicators of the accuracy of convolutional neural networks.
Now, imagine a 2D point cloud of predictors (attributes) and
target variables, i.e., each predictor (say age) constitutes the
x-coordinate and the y-coordinate is the target variable (say
loan defaulting). We will use the words attributes and
predictors interchangeably in the paper. In dimension 0, the barcode
captures the connected components of the age variable with
respect to loan defaulting. So, if the data consists of a
particular age group of people who defaulted, it will be evident
in the form of an isolated cluster, not connected to other age
groups. In other words, if there is bias in the dataset,
barcodes provide a visualization of the same. A barcode that is
significantly longer than others is indicative of bias due to that
predictor. If all barcodes are of same length, there is no bias
due to that predictor.
3.4</p>
    </sec>
    <sec id="sec-6">
      <title>Validation</title>
      <p>The aforementioned hypothesis can be verified by means of
statistical hypothesis tests. Since the distribution of
topological features has not been well characterized yet,
statistical inference on persistent homology must be non-parametric
tests [Wadhwa et al., 2018]. For our purposes, we use
nonparametric permutation test.</p>
      <p>
        If we define a function T that returns the persistent
homology of a point cloud, then given two point clouds, C and D,
we can use a permutation test to conduct statistical inference
with the following null and alternative hypotheses:
HA : T (C) = T (D)
H0 : T (C) 6= T (D)
(
        <xref ref-type="bibr" rid="ref5">1</xref>
        )
(2)
We then use the Wasserstein distance (Earth-mover’s
distance) as a similarity metric between persistent homologies
of two point clouds [Vallender, 1974].
      </p>
      <p>Going back to the example considered, suppose we have
two point clouds corresponding to two predictors, say age and
gender, with respect to loan defaulting. We can now run a
permutation test on the two point clouds to confirm that the
persistent homologies of the two are, in fact, distinct. We set
the null hypothesis that the two persistent homologies are not
distinct. The resulting p value from the test indicates whether
the null hypothesis can be rejected or not. Typically if p &lt; ,
the null hypothesis is rejected. The paramaeter is known
as the significance value of the test and a standard value of
0.05 is typically chosen for . We use off the shelf TDAstats
package to test the hypothesis [Wadhwa et al., 2018].
3.5</p>
    </sec>
    <sec id="sec-7">
      <title>Algorithm</title>
      <p>The aforementioned procedure can be summarized as
follows:
1. Create point clouds of individual predictors and the target
variable.
2. Compute Rips complex for each of the point clouds
created in step 1.
3. Compute persistance homology for the Rips complexes
created in step 2 and plot persistance barcodes.
4. For each PH, the length of the longest barcode is a way
of visualizing the bias due to individual predictors. If all
barcodes are of same length, then there is no bias. If there are
barcodes that are considerably longer than others, then there
is bias.
5. Compute p values from permutation tests setting the null
hypothesis that the resulting PHs from the two point clouds
under consideration are not distinct. A rejection of null
hypothesis indicates that the two PH are indeed distinct and thus
there is bias in the attribute which has a long barcode.
4</p>
      <sec id="sec-7-1">
        <title>Results</title>
        <p>We demonstrate the method on German credit dataset [Dua
and Graff, 2019]. This dataset contains 1000 data points
wherein the goal is to predict loan defaulting based on twenty
predictors such as credit history, savings, checking account
status, property, housing, job, etc. “Age” and “gender” are the
protected attributes with “old” and “male” being previlaged
attributes and “young” and “female” being unprevilaged
attributes.</p>
        <p>As described earlier, different topological features can be
detected at different dimensions. Dimension 0 reveals the
existence of clusters or connected components. Figure 2
provides the persistence barcodes of age with respect to loan
defaulting in dimension 0. The x-axis represents the variation
of the radius parameter r. The y axis does not have a physical
interpretation, it represents the set of all connected
components. We see that there are several individual clusters when
the radius is 1, these merge and there are two clusters until
r = 2. Beyond r = 2, there is a single large cluster that
persists. The length of the longest barcode is 4.</p>
        <p>Now, consider the plot of persistence barcodes of gender
with respect to defaulting as shown in Figure 3. We see a
few clusters which persist upto r = 1, thus the length of the
longest barcode is 1. From Figures 2 and 3, it can be inferred
that the persistence homology due to age and gender are
distinct. We can objectively validate the hypothesis that the two
PHs are different by means of permutation test as described
earlier. We obtained a p value of 0, thus leading to the
rejection of the null hypothesis that the two PHs are not distinct.
There is a single large cluster as evident by the long barcode
in the PH of age, thus the bias due to age is significant.
Furthermore, if we consider the length of the longest barcode as
an indicator of the bias in the dataset, then the bias due to age
is four times the bias due to gender (if there is any due to
gender). In fact, since there is no single persistent bar in the PH
of gender, we can conclude that there is no significant bias
due to gender. In dimension 1, we did not observe a
statistically significant difference between the two PHs. However,
to detect bias, it suffices if there is a statistically significantly
difference between the PHs in any one dimension.</p>
        <p>The aforementioned result can also be validated using the
AIfairness 360 tool as can be observed from Figure 4.
AIfairness 360 also shows that there is no bias due to gender. Four
out of the five metrics (statistical parity difference, equal
opportunity difference, disparate impact, and average odds
difference) show bias with respect to age. Theil’s index,
however does not show any bias. The presence of multiple
metrics and varying amount of bias across those metrics might
be little confusing to a software engineer not aware of the
details of these metrics. The suitability of a particular metric
may be dependent on the type of data amongst other factors.
The burden of choosing a suitable metric might thus fall on
the software engineer, who may not necessarily be equipped
with the knowledge to do so. On the other hand, persistence
homology based bias visualization method provides the
software engineer with a universal tool that is applicable across
datasets without having to choose any parameter. Intrigued by
how PH of other attributes compare with respect to protected
attributes, we also plotted the PH of attribute “job” with
respect to the target variable. The PH of “job” was the same as
that of gender indicating that there is no bias due to job as can
be observed from Figure 5.
5</p>
      </sec>
      <sec id="sec-7-2">
        <title>Accessibility for the software engineer</title>
        <p>Engineering trust-worthy software architectural pipelines is
an integral aspect of AI safety. There is a pressing need to
create accessible AI interfaces and tools to ensure that software
engineers who may not necessarily possess technical depth in
AI are able to appropriately pre-process data, select the right
AI model, and tune it. In this paper, we described how TDA
can aid software engineers in understanding bias in datasets,
a pre-processing step that is very important in the context of
AI safety.</p>
        <p>In this section, we summarize what a software engineer
needs to know to leverage this method and how they can use
the same. The mathematical background discussed earlier
may give an impression that this method is not simple enough
to be comprehended by a software engineer. However, the
nice thing about TDA is that those details are not necessary
to actually detect bias. Furthermore, the software engineer is
not burdened to choose a bias metric. Below, we enlist some
common questions and simple answers to further enhance the
accessibility of this method.</p>
        <p>What to know about TDA and PH: Topology is the study
of shapes. These shapes may be viewed as
generalizations of triangle in higher dimensions and are referred to
as simplicial complexes. Different simplicial complexes
(triangles, tetrahedrons, etc.) appear based on the
resolution at which the data is analyzed. PH is a method for
computing topological features of a space at different
spatial resolutions. Topological features could include
clusters, holes and voids in a dataset.</p>
        <p>How can TDA and PH help in detecting bias: If there is
bias due to an attribute in the dataset, it will be evident
in the form of an isolated topological feature such as a
cluster or hole that persists for a considerable interval.
How to choose a method to construct simplicial
complex: This is based on data type. For example, Rips
complex is typically chosen for point clouds, Morse complex
is chosen for images, etc.</p>
        <p>How to prepare data: For categorical datasets, create
point clouds of individual predictors and the target
variable.</p>
        <p>How to choose radius parameter : The radius parameter
is automatically chosen by PH based software to detect
topological features of interest.</p>
        <p>How to compute PH: PH is visualized in terms of
persistence diagrams and barcodes. Persistence diagram is
a plot of the birth time (i.e., the value of the radius at
which a topological feature appears) and death time (i.e.,
the value of the radius at which a topological feature
disappears) of a topological feature. Persistence barcodes
capture the interval between the birth and death of a
topological feature. There are several off-the-shelf
software packages available to compute PH like GHUDHI,
R-TDA, Dipha, etc [Pun et al., 2018] which can be
chosen based on the language of preference.</p>
        <p>How to interpret the barcodes for bias detection: If there
is a single long barcode, there is bias. If all barcodes are
of same length, there is no bias. See Figure 6.</p>
        <p>How to validate bias: Use non-parametric permutation
test to show that the PH of the predictor contributing to
bias is distinct from other predictors.
6</p>
      </sec>
      <sec id="sec-7-3">
        <title>Conclusions</title>
        <p>Topological data analysis offers promising alternate feature
representation techniques. In this work, we described a novel
way of quantifying bias in datasets. Specifically, we used
persistence homology to determine bias due to different
attributes in the German credit dataset and validated the same
using non-parametric statistical permutation tests. The
proposed visualization can serve as a useful pre-processing tool
for software engineers to understand which attributes need
to be accounted for and mitigated when ensuring fairness in
classification.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <source>[Al-Rubaie and Chang</source>
          , 2018]
          <string-name>
            <given-names>Mohammad</given-names>
            <surname>Al-Rubaie</surname>
          </string-name>
          and
          <string-name>
            <given-names>Morris</given-names>
            <surname>Chang</surname>
          </string-name>
          .
          <article-title>Privacy preserving machine learning - threats and solutions</article-title>
          .
          <source>IEEE Security and Privacy Magazine</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [Bellamy et al.,
          <year>2018</year>
          ]
          <string-name>
            <given-names>Rachel</given-names>
            <surname>Bellamy</surname>
          </string-name>
          , Kuntal Dey, Michael Hind, Samuel C. Hoffman, Stephanie Houde, Kalapriya Kannan, Pranay Lohia, Jacquelyn Martino, Sameep Mehta, Aleksandra Mojsilovic, Seema Nagar, Karthikeyan Natesan Ramamurthy, John Richards, Diptikalyan Saha, Prasanna Sattigeri, Moninder Singh,
          <string-name>
            <given-names>Ramazon</given-names>
            <surname>Kush</surname>
          </string-name>
          , and Yunfeng Zhang.
          <source>Ai</source>
          fairness
          <volume>360</volume>
          :
          <article-title>An extensible toolkit for detecting, understanding, and mitigating unwanted algorithmic bias</article-title>
          .
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <source>[Bostrom and Yudkowsky</source>
          , 2018]
          <string-name>
            <given-names>Nick</given-names>
            <surname>Bostrom</surname>
          </string-name>
          and
          <string-name>
            <given-names>Eliezer</given-names>
            <surname>Yudkowsky</surname>
          </string-name>
          .
          <source>The ethics of artificial intelligence. Cambridge Handbook of Artificial Intelligence</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <source>[Carlsson and Gabrielsson</source>
          , 2018] Gunnar Carlsson and Richard Gabrielsson.
          <article-title>Topological approaches to deep learning</article-title>
          .
          <source>ArXiv</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <source>[Challen1</source>
          et al.,
          <year>2019</year>
          ]
          <article-title>Robert Challen1</article-title>
          ,
          <string-name>
            <surname>Joshua</surname>
            <given-names>Denny</given-names>
          </string-name>
          , Martin Pitt, Luke Gompels, Tom Edwards, and
          <string-name>
            <surname>Krasimira</surname>
          </string-name>
          Tsaneva-Atanasova1.
          <article-title>Artificial intelligence, bias and clinical safety</article-title>
          .
          <source>BMJ Quality and Safety</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [Chen et al.,
          <year>2018</year>
          ]
          <string-name>
            <surname>Irene</surname>
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>Fredrik D.</given-names>
          </string-name>
          <string-name>
            <surname>Johansson</surname>
          </string-name>
          , and David Sontag.
          <article-title>Why is my clasifier discriminatory</article-title>
          .
          <source>NeurIPS</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <source>[Dua and Graff</source>
          , 2019]
          <string-name>
            <given-names>D.</given-names>
            <surname>Dua</surname>
          </string-name>
          and
          <string-name>
            <surname>C. Graff.</surname>
          </string-name>
          <article-title>Uci machine learning repository</article-title>
          . University of California, School of Information and Computer Science, Irvine,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <source>[FSB</source>
          ,
          <year>2017</year>
          ] FSB.
          <article-title>Artificial intelligence and machine learning in financial services: Market developments and financial stability implications</article-title>
          .
          <source>Technical Report: Financial Stability Board</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <source>[Gabrielsson and Carlsson</source>
          , 2018]
          <article-title>Richard Gabrielsson and Gunnar Carlsson. Exposition and interpretation of the topology of neural networks</article-title>
          .
          <source>ArXiv</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [Gebru et al.,
          <year>2018</year>
          ]
          <string-name>
            <given-names>Timnit</given-names>
            <surname>Gebru</surname>
          </string-name>
          , Jamie Morgenstern, Briana Vecchione, Jennifer Wortman Vaughan, Hanna Wallach,
          <article-title>Hal Daumee´ III, and Kate Crawford. Datasheets for datasets</article-title>
          .
          <source>Proceedings of the 5 th Workshop on Fairness, Accountability, and Transparency in Machine Learning</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [Goodfellow et al.,
          <year>2014</year>
          ]
          <string-name>
            <given-names>I.</given-names>
            <surname>Goodfellow</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Shlens</surname>
          </string-name>
          , and
          <string-name>
            <given-names>C.</given-names>
            <surname>Szegedy</surname>
          </string-name>
          .
          <article-title>Explaining and harnessing adversarial examples</article-title>
          .
          <source>ICLR</source>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <source>[Gunnar</source>
          , 2018]
          <string-name>
            <given-names>Carlsson</given-names>
            <surname>Gunnar</surname>
          </string-name>
          .
          <article-title>Going deeper: Understanding how convolutional neural networks learn using tda</article-title>
          .
          <source>Ayasdi Blog</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [Kleinberg et al.,
          <year>2018</year>
          ]
          <string-name>
            <given-names>Jon</given-names>
            <surname>Kleinberg</surname>
          </string-name>
          , Himabindu Lakkaraju, Jure Leskovec, Jens Ludwig, and
          <string-name>
            <given-names>Sendhil</given-names>
            <surname>Mullainathan</surname>
          </string-name>
          .
          <article-title>Human decisions and machine predictions</article-title>
          .
          <source>The quarterly journal of economics</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <source>[Knight</source>
          , 2018]
          <string-name>
            <given-names>Will</given-names>
            <surname>Knight</surname>
          </string-name>
          .
          <article-title>Microsoft is creating an oracle for catching biased ai algorithms</article-title>
          .
          <source>MIT Technology Review</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [Kurutach et al.,
          <year>2018</year>
          ]
          <string-name>
            <given-names>Thanard</given-names>
            <surname>Kurutach</surname>
          </string-name>
          , Aviv Tamar, Ge Yang, Stuart Russell, and
          <string-name>
            <given-names>Pieter</given-names>
            <surname>Abbeel</surname>
          </string-name>
          .
          <article-title>Learning plannable representations with causal infogan</article-title>
          .
          <source>NeurIPS</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <string-name>
            <surname>[Li</surname>
          </string-name>
          et al.,
          <year>2014</year>
          ]
          <string-name>
            <given-names>C.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ovsjanikov</surname>
          </string-name>
          , and
          <string-name>
            <given-names>F.</given-names>
            <surname>Chazal</surname>
          </string-name>
          .
          <article-title>Persistence-based structural recognition</article-title>
          .
          <source>CVPR</source>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [Nguyen et al.,
          <year>2018</year>
          ]
          <string-name>
            <given-names>D. D.</given-names>
            <surname>Nguyen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z. X.</given-names>
            <surname>Cang</surname>
          </string-name>
          ,
          <string-name>
            <surname>K. D. Wu</surname>
            ,
            <given-names>M. L.</given-names>
          </string-name>
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Cao</surname>
            , and
            <given-names>G. W.</given-names>
          </string-name>
          <string-name>
            <surname>Wei</surname>
          </string-name>
          .
          <article-title>Mathematical deep learning for pose and binding affinity prediction and ranking in d3r grand challenges</article-title>
          .
          <source>ArXiv</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [Pachauri et al.,
          <year>2011</year>
          ]
          <string-name>
            <given-names>D.</given-names>
            <surname>Pachauri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Hinrichs</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.K.</given-names>
            <surname>Chung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.C.</given-names>
            <surname>Johnson</surname>
          </string-name>
          , and
          <string-name>
            <given-names>V.</given-names>
            <surname>Singh</surname>
          </string-name>
          .
          <article-title>Topology-based kernels with application to inference problems in alzheimer's disease</article-title>
          .
          <source>IEEE Transactions on Medical Imaging</source>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          <source>[Peters</source>
          , 2018]
          <string-name>
            <given-names>Adele</given-names>
            <surname>Peters</surname>
          </string-name>
          .
          <article-title>This tool lets you see-and correct-the bias in an algorithm</article-title>
          .
          <source>Fast Company</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [Pettigrew et al.,
          <year>2018</year>
          ]
          <string-name>
            <given-names>Simone</given-names>
            <surname>Pettigrew</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Lin</given-names>
            <surname>Fritschi</surname>
          </string-name>
          , and Richard Norman2.
          <article-title>The potential implications of autonomous vehicles in and around the workplace</article-title>
          .
          <source>Int J Environ Res Public Health.</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [Pun et al.,
          <year>2018</year>
          ]
          <string-name>
            <given-names>Chi</given-names>
            <surname>Pun</surname>
          </string-name>
          , Kelin Xia, and
          <string-name>
            <given-names>Si</given-names>
            <surname>Lee</surname>
          </string-name>
          .
          <article-title>Persistenthomology-based machine learning and its applications - a survey</article-title>
          .
          <source>ArXiv</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          <source>[Ramakrishan and Shah</source>
          , 2016]
          <string-name>
            <given-names>R.</given-names>
            <surname>Ramakrishan</surname>
          </string-name>
          and
          <string-name>
            <given-names>J.</given-names>
            <surname>Shah</surname>
          </string-name>
          .
          <article-title>Towards interpretable explanations for transfer learning in sequential tasks</article-title>
          .
          <source>AAAI Spring Symposium Series</source>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          <source>[Srivastava and Rossi</source>
          , 2018]
          <string-name>
            <given-names>B.</given-names>
            <surname>Srivastava</surname>
          </string-name>
          and
          <string-name>
            <given-names>F.</given-names>
            <surname>Rossi</surname>
          </string-name>
          .
          <article-title>Towards composable bias rating of ai systems</article-title>
          .
          <source>AIES</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          <source>[Stilgoe</source>
          , 2017]
          <string-name>
            <given-names>Jack</given-names>
            <surname>Stilgoe</surname>
          </string-name>
          .
          <article-title>Machine learning, social learning and the governance of self-driving cars</article-title>
          .
          <source>Social Studies of Science</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          <source>[Umeda</source>
          , 2017]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Umeda</surname>
          </string-name>
          .
          <article-title>Time series classification via topological data analysis</article-title>
          .
          <source>Information and Media Technologies</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          <source>[Vallender</source>
          , 1974]
          <string-name>
            <given-names>S.S.</given-names>
            <surname>Vallender</surname>
          </string-name>
          .
          <article-title>Calculation of the wasserstein distance between probability distributions on the line</article-title>
          .
          <source>Theory of Probability and its Applications</source>
          ,
          <year>1974</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [Wadhwa et al.,
          <year>2018</year>
          ]
          <string-name>
            <given-names>Raoul</given-names>
            <surname>Wadhwa</surname>
          </string-name>
          , Drew Williamson,
          <string-name>
            <given-names>Andrew</given-names>
            <surname>Dhawan</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Jacob</given-names>
            <surname>Scott</surname>
          </string-name>
          . Tdastats:
          <article-title>R pipeline for computing persistent homology in topological data analysis</article-title>
          .
          <source>Journal of open source software</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          <string-name>
            <surname>[Wang</surname>
          </string-name>
          et al.,
          <year>2011</year>
          ]
          <string-name>
            <given-names>B.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Summa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Pascucci</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.</given-names>
            <surname>Vejdemo-Johansson</surname>
          </string-name>
          .
          <article-title>Branching and circular features in high dimensional data</article-title>
          .
          <source>IEEE Transactions on Visualization and Computer Graphics</source>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          <source>[Weintraub</source>
          , 2014]
          <string-name>
            <given-names>Steven</given-names>
            <surname>Weintraub</surname>
          </string-name>
          .
          <source>Fundamentals of algebraic topology</source>
          . Springer,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          <source>[Wexler</source>
          , 2018]
          <string-name>
            <given-names>James</given-names>
            <surname>Wexler</surname>
          </string-name>
          .
          <article-title>The what-if tool: Code-free probing of machine learning models</article-title>
          .
          <source>Google AI blog</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [Zeppelzauer et al.,
          <year>2018</year>
          ]
          <string-name>
            <given-names>M.</given-names>
            <surname>Zeppelzauer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Zielinski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Juda</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.</given-names>
            <surname>Seidl</surname>
          </string-name>
          .
          <article-title>A study on topological descriptors for the analysis of 3d surface texture</article-title>
          .
          <source>CVIU</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          <string-name>
            <surname>[Zhou</surname>
          </string-name>
          et al.,
          <year>2017</year>
          ]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y. Z.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Wang</surname>
          </string-name>
          , and
          <string-name>
            <given-names>T. N.</given-names>
            <surname>Tan</surname>
          </string-name>
          .
          <article-title>Exploring generalized shape analysis by topological representations</article-title>
          .
          <source>Pattern Recognition Letters</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          <source>[Zhu</source>
          ,
          <year>2013</year>
          ]
          <string-name>
            <given-names>X. J.</given-names>
            <surname>Zhu</surname>
          </string-name>
          .
          <article-title>Persistent homology: An introduction and a new text representation for natural language processing</article-title>
          .
          <source>IJCAI</source>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>