<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>March</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Making Transparency Clear</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>The Dual Importance of Explainability</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Auditability</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Aaron Springer Computer Science University of California Santa Cruz Santa Cruz</institution>
          ,
          <addr-line>CA</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Steve Whittaker Psychology University of California Santa Cruz Santa Cruz</institution>
          ,
          <addr-line>CA</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2019</year>
      </pub-date>
      <volume>20</volume>
      <issue>2019</issue>
      <abstract>
        <p>Algorithmic transparency is currently invoked for two separate purposes: to improve trust in systems and to provide insight into problems like algorithmic bias. Although transparency can help both problems, recent results suggest these goals cannot be accomplished simultaneously by the same transparency implementation. Providing enough information to diagnose algorithmic bias will overwhelm users and lead to poor experiences. On the other hand, scaffolding user mental models with selective transparency will not provide enough information to audit these systems for fairness. This paper argues that if we want to address both problems we must separate two distinct aspects of transparency: explainability and auditability. Explainability improves user experience by facilitating mental model formation and building user trust. It provides users with sufficient information to form accurate mental models of system operation. Auditability is more exhaustive; providing third-parties with the ability to test algorithmic outputs and diagnose biases and unfairness. This conceptual separation provides a path forward for designers to make systems both usable and free from bias.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Transparency</kwd>
        <kwd>trust</kwd>
        <kwd>explanation</kwd>
        <kwd>bias</kwd>
        <kwd>auditability</kwd>
        <kwd>algorithms</kwd>
        <kwd>intelligent systems</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>CCS CONCEPTS</title>
      <p>• Human-centered computing~Human computer interaction (HCI)</p>
      <p>IUI Workshops'19, March 20, 2019, Los Angeles, USA.
Copyright © 2019 for the individual papers by the papers' authors. Copying
permitted for private and academic purposes. This volume is published and
copyrighted by its editors.
1</p>
    </sec>
    <sec id="sec-2">
      <title>Introduction</title>
      <p>We are at a pivotal time in the use of machine learning as
intelligent systems increasingly impact our daily lives. Machine
learning algorithms underlie the many intelligent systems we
routinely use. These systems provide information ranging from
routes to work to recommendations about criminal parole [2,4].
As humans with limited time and attention, we increasingly defer
responsibility to these systems with little reflection or oversight.
For example, as of February 2018, over 50% of adults in the
United States report using a range of voice assistants on a daily
basis to accomplish tasks such as navigating to work, answering
queries, and automating actions [27]. Improvements to the
increasing use of voice assistants are largely driven by
improvements in underlying algorithms.</p>
      <p>Compounding these advances in machine learning is the fact
that many people have difficulty understanding current intelligent
systems [38]. Here, we use ‘intelligent systems’ to mean systems
that use machine learned models and/or data derived from user
context to make predictions. The machine learning models that
often power these intelligent systems are complex and trained
upon massive troves of data, making it difficult for even experts to
form accurate mental models. For example, many Facebook users
did not know that the service curated their newsfeed using
machine learning, they simple thought that they saw a feed of all
their connections posts [15]. More recently, users of Facebook
and other systems have been shown to generate simple “folk
theories” that explain how such systems are working [14,38].
Although users cannot validate such folk theories that does not
stop users from acting upon them. [14] demonstrated that users
went so far as to modify how they interacted with Facebook to try
to force the system to present a certain outcome consistent with
their user folk theory. There is potential for danger in other
contexts when users are willing to act upon their folk hypotheses
when not given the ability to understand the system. Furthermore,
there are many challenges regarding the best ways to effectively
communicate underlying algorithms to users [35,39].</p>
      <p>
        Another concern is the user experience of opaque algorithmic
systems. Without any form of transparency, users may trust and
understand these systems less [11,24]. Even in low-stakes systems
like the Netflix recommender, users still struggle to understand
how to control and influence internal algorithms [
        <xref ref-type="bibr" rid="ref2">6</xref>
        ]. These
problems surrounding user experience, trust especially, become
more pronounces in high stakes scenarios such as the medical
field where elements of user experience like trust are essential to a
program’s use.
      </p>
      <p>Furthermore, academics and industry practitioners are
discovering other significant issues in deploying these systems.
Intelligent systems powered by machine learning can learn and
embody societal biases. Systems may therefore treat users
differently based on characteristics of users’ speech and writing
[31,37] or even based upon characteristics that are protected under
law [2]. In a particularly egregious example, an intelligent system
used to help inform parole decisions was found to discriminate
against people of color [2].</p>
      <p>
        Despite these challenges of bias and user experience, many
critics have coalesced around a concept they believe could address
these challenges: transparency. The insight underlying
transparency is that an algorithm should reveal itself to users.
There are many important potential benefits for algorithmic
transparency. Transparency enables important oversight by
system designers. Without transparency it may be unclear whether
an algorithm is optimizing the intended behavior [39], or whether
an algorithm has negative, unintended consequences (e.g. filter
bubbles in social media; [26]). These arguments have led some
researchers to argue that machine learning must be ‘interpretable
by design’ [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], and that transparency is even essential for the
adoption of intelligent systems, such as in cases of medical
diagnoses [40]. Transparency has taken on the role of a cure-all
for machine learnings woes.
      </p>
      <p>However, problems remain. Transparency is currently
illdefined [12]. Transparency is purported to address machine
learning problems such as bias [25], while simultaneously
improving the user experience [18,21]. This paper argues that
achieving both goals may be impossible with a single
implementation. An implementation of transparency that allows
someone to infer system bias will likely overwhelm users and lead
to less usage—which in turn will lead to developers refusing to
implement transparency. Transparency should be disaggregated
into two separate classes: explainability and auditability.
Explainability is concerned with building interfaces that promote
accurate mental models of system operation leading to a better
user experience. Auditability is concerned with allowing users or
third-party groups to audit a deployed algorithmic system for bias
and other problems. Separating these aspects of transparency
allows us to build systems with improved user experiences while
maintaining high standards of fairness and unbiased outcomes.</p>
    </sec>
    <sec id="sec-3">
      <title>Why Do We Need Transparency? 2 2.1</title>
    </sec>
    <sec id="sec-4">
      <title>Poor User Experiences in Intelligent Systems</title>
      <p>A wealth of prior work has explored issues surrounding
algorithm transparency in the commercial deployments of systems
for social media and news curation. Social media feeds are often
curated by algorithms that may be invisible to users (e.g.,
Facebook. Twitter, LinkedIn). Work on algorithmic folk theories
shows that making the designs more transparent or seamful,
allowed users to better understand and work within the system
[14].</p>
      <p>
        Addressing the user experience in intelligent systems has now
become a pressing concern for mainstream usability practitioners.
The Nielsen Norman group recently completed a diary study
examining the user experience of normal people with systems
such as Facebook, Instagram, Netflix, and Google News [
        <xref ref-type="bibr" rid="ref2">6</xref>
        ].
Mirroring the work on Facebook folk theories, users found it
unclear which aspects of their own behavior the intelligent
systems used as inputs. Users were also frustrated by the lack of
control over the output. Overall, users struggled to form correct
mental models of system operation which led to poor user
experiences.
      </p>
      <p>Other work shows the importance of transparency for building
trust in algorithmic systems, an important part of the user
experience. Users who receive explanations better understand and
trust complex algorithmic systems [24]. In the presence of
disagreement between the system and the user, transparency can
improve user perceptions of trust and system accuracy [11,23,34].
But in addition to improving user experience, advocates point to
transparency as a counter to more pernicious problems such as
algorithmic bias.
2.2</p>
    </sec>
    <sec id="sec-5">
      <title>Revealing Bias</title>
      <p>Intelligent systems and predictive analytics have been shown
to learn and perpetuate societal biases. One clear example of this
is COMPAS, an algorithm used widely within the United States to
predict risk of recidivism. In 2016 ProPublica published an article
noting that the COMPAS system was more likely to predict higher
risk scores for people of color than other populations, even when
the ground truth was similar [2]. The COMPAS system had been
in use for over 5 years in some locations before these biases were
publicized [13].</p>
      <p>Other work shows how interfaces can discriminate based on
ways of speaking and writing. YouTube captions have been
shown to be less accurate for speakers with a variety of accents
[37]. Common voice interfaces can struggle with specific ways of
speaking [31]. These problems likely arise from how algorithms
were trained on a non-diverse set of voices (i.e., ‘distributional
drift’), and then deployed broadly to all people. Even textual
methods are not immune to embodying societal biases. Word
embeddings have been shown to harbor biases related to gender.
For example, one of the roles most closely related to ‘she’ within
the learned word embeddings is “homemaker”; in contrast, an
occupation closely related to “he” is “boss” [5].</p>
      <p>The fear is that the embodiment of these societal biases within
machine learning systems will perpetuate them. For example,
biased recidivism algorithms will exacerbate existing inequalities,
creating a cycle where those who are not currently privileged will
have even less opportunity in the future. An example of this is
shown in the posting of job ads online. Men saw significantly
more job ads for senior positions compared to women, when
searching online [10]. In other cases, African-American names in
Google search are more likely to display ads for criminal records,
which has been noted as a possible risk for job applicants [36].</p>
      <p>It is not simple to fix these problems. Algorithmic bias
problems are everywhere; but fixing them requires fitting complex
research and auditing practices into iterative agile workflows [32].
This combination requires new tools and extensive organizational
buy-in [9]. Even with these processes and tools, not all biases will
be found and fixed before a system is deployed.</p>
      <p>Transparency has been invoked as a solution to bias.
Bestselling books such as Weapons of Math Destruction call for
increased transparency as a counter to algorithmic bias [25]. Even
the call for papers for this workshop notes that ‘algorithmic
processes are opaque’ and that this can hide issues of algorithmic
bias [20]. The idea is that transparency can expose the inner
working of an algorithm, allowing users to see whether or not the
system is biased. This allows third parties to have the ability to
audit the algorithmic systems they are using. However, showing
complete algorithmic transparency may have negative impacts on
the user experience.
3</p>
    </sec>
    <sec id="sec-6">
      <title>Transparency Troubles</title>
      <p>Although transparency is an active research area in both
machine learning and HCI communities, we believe that a major
barrier to current conceptualizations of transparency is the
potential negative effects on user experience. Even though a goal
of much transparency research is to improve the user experience
by building trust, studies are continually showing that
transparency has mixed effects on the user experience with
intelligent systems.</p>
      <p>One system built by our research team clearly reveals
problems with our current concept of transparency. The E-meter is
an “intelligent” system with an algorithm that assesses the
positivity and negativity of a users’ writing emotional writing in
real time [33]. Users were asked to write about personal emotional
experiences and the system interpreted their writing to evaluate
how each user felt about their experiences. The E-meter was
transparent; it highlighted the words used by the machine learning
model conveyed their corresponding emotional weights through a
color gradient. The results were unexpected. Users of the
transparent system actually felt the system was less accurate
overall [34]. Why was this? In some cases, seeing inevitable
system errors undermined user confidence, and in other cases,
users overrode correct system models that conflicted with their
own (inaccurate) beliefs.</p>
      <p>Further tests on the E-meter system showed other problems
with transparency. Users with a non-transparent version of the
Emeter thought that the system performed more accurately [35]. On
the other hand, users with transparency seemed to find it
distracting. Users of the transparent system were also prone to
focus errors exposed by the transparency, even when the overall
mood prediction was correct. Clearly, distracting users and
leading them to believe the system is more errorful does not create
a positive user experience.</p>
      <p>Furthermore, users may not want complete transparency for
other reasons. Providing such information may be distracting due
to the overhead in processing that transparency requires [7].
Transparency negatively affects the user experience in less
accurate systems [23]. Short explanations of what a system is
doing can improve trust but full transparency can result in less
trust in intelligent systems [21].</p>
      <p>Together these studies provide strong evidence that exhaustive
transparency may undermine the user experience. It may distract
users, provide them with too much information, and provoke
unnecessary doubt in the system. Transparency is trying to do too
much. We cannot exhaustively convey the inner workings of
many algorithms, nor is that what users want. However, without
making these complete inner-workings transparent, how can we
audit these systems for unfairness and bias?</p>
      <p>As we have shown in previous work, diagnosing and fixing
algorithmic bias is not a simple task, even for the creators of a
system [9]. These creators have access to the complete code, data,
and inner workings of the system; even with this access, fixing
algorithmic bias is a challenge. How much harder will it then be
for third parties and users to diagnose algorithmic bias through a
transparent interface which does not display all of this
information? We cannot reasonably expect that our current
operationalization of transparency by explanation will allow third
parties to diagnose bias in deployed systems.</p>
      <p>In summary, these two goals of transparency conflict. We
cannot simultaneously improve the user experience while
providing a mechanism for diagnosing algorithmic bias. Providing
enough information to diagnose algorithmic bias will overwhelm
users and lead to poor experiences. On the other hand, scaffolding
user mental models with selective transparency will not provide
enough information to audit these systems for fairness. In order
for transparency to be successful, we need to clarify our aims. We
must separate transparency into two related concepts:
explainability and auditability.
4</p>
    </sec>
    <sec id="sec-7">
      <title>Two Facets of Transparency</title>
      <p>The first facet, explainability, has a single goal: to improve the
user experience. Many problems with intelligent systems occur
because users lack proper mental models of how the system
operates [14] and helping users form an accurate mental model
improves satisfaction [22]. Therefore, the goal of explainability is
to facilitate an ‘accurate enough’ mental model formation to
enable correct action within the system. Attempting to go beyond
helping users form heuristics may lead to worse user experience
[35]. We need to give users heuristics and approximate
understandings so that they can feel that they are in control of the
interface.</p>
      <p>The key to explainability is to reveal only the information
needed by users [12]. This separates it from many current
conceptualizations of transparency that aim for completeness.
Explanations that aim for completeness may induce poor user
experiences because they are too complex [19] or conflict with
users’ mental models [30,35]. In addition, explaining only the
needed elements conforms better to the extensive bodies of social
science research that study explanation. Explanations should
follow Grices’s maxims [17], i.e. to only explain as much as is
needed and no more. Explanation should be occasioned [16], it
should present itself when needed and disappear when not.
Exhaustive transparency does conform with HCI experimental
results or these social science theories; which is why it is essential
that we study explainability.</p>
      <p>Explainability can happen through a variety of means. For
example, we can use natural language to explain results. For
example, Facebook has a feature labeled ‘Why am I seeing this?’
on ads that provides a natural language explanation of the user
profile factors that led to the targeted ad. These explanations can
also involve data and visualization intended to fill in gaps in the
user’s mental models [12]. The range of explanation types is
large, from simple natural language to explorable explanations.
This is necessary given the many domains in which explanations
are needed. Explanations must be tailored to the domain; doctors
have very different needs than mobile fitness coach users. For
example, doctors are making high-stakes decisions and are likely
to be very invested in each decision; therefore, the explanations
for doctors should be more complete and contain more
information. Such lengthy explanations may not be successful in
more casual settings such as an intelligent mobile fitness coach
where users may be less motivated to process a lengthy
explanation. Again, explanations are to improve the use of the
system and the user experience, not to provide the user the ability
to ensure the system is fair and free from bias.</p>
      <p>But how can transparency satisfy its second goal of ensuring
fair algorithms? Explainability is insufficient to meet this
requirement. It is not possible to ensure that an intelligent system
is fair on the basis of the natural language explanations it
provides. How then, can we determine whether algorithms are fair
and free from bias?</p>
      <p>In addition to explainability, the second facet of transparency
is auditability of deployed systems. We define auditable as the
ability for users or third parties to validate and test the deployed
system by providing their own data for the system to predict on.
While some systems are currently auditable, it is mostly
adversarial; auditors must use methods such as sock-puppet
auditing to determine whether a system is biased [29]. For an
example of auditability, in Facebook, users are beholden to seeing
advertisements targeted to their profile information. An auditable
version of Facebook advertisements would have the ability to
supply any profile data and receive back what targeted
advertisements the supplied data would generate. A current
example of systems that are easily auditable is current facial
recognition APIs created by cloud providers; these are
programmable and thus supplying data and checking for bias can
be done by independent researchers [28].</p>
      <p>Other definitions of auditability rely on seeing the code itself
[8], but this may not be necessary. Relying on seeing the code
itself complicates the audit process considerably due to source
code being highly valued intellectual property. Rather we should
pursue audits that allow the user or a third party to generate their
own conclusions about the fairness of the algorithm, rather than
relying on the explanations it generates. We do not need to know
how the underlying algorithm works to ensure that it is generating
fair predictions for all possible subgroups. Under many criterions
of fairness such as independence and separation, all we need to
know are the predicted output and the data [3]. Knowledge about
the inner-workings of the algorithm is not required to ensure
fairness. The expectation is not that every user has the skill or
desire to audit these algorithms but rather that auditability is
possible, in case it should be needed.</p>
      <p>Given space constraints, we do not attempt to prescribe here
exactly how auditability should be implemented. According to our
definition, it could be as simple as an exposed public API
endpoint that takes parameters and returns a prediction. While an
API endpoint is the simplest implementation for developers, there
is no reason that a user interface to supply data and view
predictions could not be created. For instance, the E-meter we
talked of earlier exhaustively exposed its predictions and data to
users allowing them to edit and explore what text results in
different predictions. These both fit the definition of auditability
by allowing the user to provide known data as input and receive a
prediction. While an API endpoint is a simple solution, further
research should explore what form auditability should take in
interactive programs.
6</p>
    </sec>
    <sec id="sec-8">
      <title>Conclusion</title>
      <p>Algorithmic transparency is purported to improve the user
experience and simultaneously help diagnose algorithmic bias.
We argue that these goals cannot be accomplished simultaneously
with the same implementation. Exposing enough information to
diagnose algorithmic bias overwhelms users and leads to a poor
user experience. We therefore distinguish two aspects of
transparency: explainability and auditability. Explainability aims
to improve the user experience through making users aware of
inputs and reasons for the system predictions; this is necessarily
incomplete, providing just enough information to allow users to
form simple mental system models. Auditability ensures that third
parties and users can test a system’s predictions for fairness and
bias by providing their own data for predictions. Distinguishing
these two aspects of transparency provides a way forward for
industry implementations of usable and safe algorithmic systems.</p>
    </sec>
    <sec id="sec-9">
      <title>ACKNOWLEDGMENTS</title>
      <p>We would like to thank Victoria Hollis, Ryan Compton, and
Lee Taber for their feedback on this project. We would also like
to thank the anonymous reviewers for their insightful comments
that helped refine this work.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Ashraf</given-names>
            <surname>Abdul</surname>
          </string-name>
          , Jo Vermeulen, Danding Wang,
          <string-name>
            <surname>Brian</surname>
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Lim</surname>
            , and
            <given-names>Mohan</given-names>
          </string-name>
          <string-name>
            <surname>Kankanhalli</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Trends and Trajectories for Explainable, Accountable and Intelligible Systems: An HCI Research Agenda</article-title>
          .
          <source>In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems - CHI '18</source>
          ,
          <fpage>1</fpage>
          -
          <lpage>18</lpage>
          . https://doi.org/10.1145/3173574.3174156 [2]
          <string-name>
            <given-names>Julia</given-names>
            <surname>Angwin</surname>
          </string-name>
          and
          <string-name>
            <given-names>Jeff</given-names>
            <surname>Larson</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Machine Bias</article-title>
          .
          <source>ProPublica. Retrieved October 27</source>
          ,
          <year>2017</year>
          from https://www.propublica.org/article/machine-biasrisk
          <article-title>-assessments-in-criminal-</article-title>
          <string-name>
            <surname>sentencing</surname>
            [3]
            <given-names>Solon</given-names>
          </string-name>
          <string-name>
            <surname>Barocas</surname>
            , Moritz Hardt, and
            <given-names>Arvind</given-names>
          </string-name>
          <string-name>
            <surname>Narayanan</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Fairness and machine learning</article-title>
          .
          <source>Fairness and machine learning</source>
          .
          <source>Retrieved January 10</source>
          ,
          <year>2019</year>
          from https://fairmlbook.org/ [4]
          <string-name>
            <given-names>Reuben</given-names>
            <surname>Binns</surname>
          </string-name>
          , Max Van Kleek,
          <string-name>
            <given-names>Michael</given-names>
            <surname>Veale</surname>
          </string-name>
          , Ulrik Lyngs,
          <string-name>
            <given-names>Jun</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>and Nigel</given-names>
            <surname>Shadbolt</surname>
          </string-name>
          .
          <year>2018</year>
          . “
          <article-title>It's Reducing a Human Being to a Percentage”; Perceptions of Justice in Algorithmic Decisions</article-title>
          .
          <source>Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems - CHI '18</source>
          :
          <fpage>1</fpage>
          -
          <lpage>14</lpage>
          . https://doi.org/10.1145/3173574.3173951 [5]
          <string-name>
            <given-names>Tolga</given-names>
            <surname>Bolukbasi</surname>
          </string-name>
          ,
          <string-name>
            <surname>Kai-Wei</surname>
            <given-names>Chang</given-names>
          </string-name>
          , James Y. Zou, Venkatesh Saligrama, and
          <string-name>
            <surname>Adam</surname>
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Kalai</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Man is to computer programmer as woman is to homemaker? Debiasing word embeddings</article-title>
          .
          <source>In Advances in Neural Information Processing Systems</source>
          ,
          <volume>4349</volume>
          -
          <fpage>4357</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>Raluca</given-names>
            <surname>Budiu</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Can Users Control and Understand a UI Driven by Machine Learning? Nielsen Norman Group</article-title>
          .
          <source>Retrieved January 10</source>
          ,
          <year>2019</year>
          from https://www.nngroup.com/articles/machine-learning-ux/ [7]
          <string-name>
            <given-names>Andrea</given-names>
            <surname>Bunt</surname>
          </string-name>
          , Matthew Lount, and
          <string-name>
            <given-names>Catherine</given-names>
            <surname>Lauzon</surname>
          </string-name>
          .
          <year>2012</year>
          .
          <article-title>Are explanations always important?: a study of deployed, low-cost intelligent interactive systems</article-title>
          .
          <source>In Proceedings of the 2012 ACM international conference on Intelligent User Interfaces</source>
          ,
          <fpage>169</fpage>
          -
          <lpage>178</lpage>
          . Retrieved April 25,
          <year>2017</year>
          from http://dl.acm.org/citation.cfm?id=
          <volume>2166996</volume>
          [8]
          <string-name>
            <given-names>Jenna</given-names>
            <surname>Burrell</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>How the machine 'thinks': Understanding opacity in machine learning algorithms</article-title>
          .
          <source>Big Data &amp; Society</source>
          <volume>3</volume>
          ,
          <issue>1</issue>
          :
          <fpage>2053951715622512</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          https://doi.org/10.1177/2053951715622512 Henriette Cramer, Jean Garcia-Gathright, Aaron Springer, and
          <string-name>
            <given-names>Sravana</given-names>
            <surname>Reddy</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Assessing and Addressing Algorithmic Bias in Practice.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <source>Interactions</source>
          <volume>25</volume>
          ,
          <issue>6</issue>
          :
          <fpage>58</fpage>
          -
          <lpage>63</lpage>
          . https://doi.org/10.1145/3278156 Amit Datta, Michael Carl Tschantz, and
          <string-name>
            <given-names>Anupam</given-names>
            <surname>Datta</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <source>Automated Experiments on Ad Privacy Settings. Proceedings on Privacy Enhancing Technologies</source>
          <year>2015</year>
          ,
          <volume>1</volume>
          :
          <fpage>92</fpage>
          -
          <lpage>112</lpage>
          . https://doi.org/10.1515/popets-2015-0007
          <string-name>
            <surname>Mary</surname>
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Dzindolet</surname>
            ,
            <given-names>Scott A.</given-names>
          </string-name>
          <string-name>
            <surname>Peterson</surname>
          </string-name>
          , Regina A.
          <string-name>
            <surname>Pomranky</surname>
          </string-name>
          , Linda G.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <surname>Pierce</surname>
            ,
            <given-names>and Hall P.</given-names>
          </string-name>
          <string-name>
            <surname>Beck</surname>
          </string-name>
          .
          <year>2003</year>
          .
          <article-title>The Role of Trust in Automation Reliance</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <given-names>Int. J.</given-names>
            <surname>Hum</surname>
          </string-name>
          .-Comput. Stud.
          <volume>58</volume>
          ,
          <issue>6</issue>
          :
          <fpage>697</fpage>
          -
          <lpage>718</lpage>
          . https://doi.org/10.1016/S1071-
          <volume>5819</volume>
          (
          <issue>03</issue>
          )
          <fpage>00038</fpage>
          -
          <lpage>7</lpage>
          Malin Eiband, Hanna Schneider,
          <string-name>
            <given-names>Mark</given-names>
            <surname>Bilandzic</surname>
          </string-name>
          , Julian Fazekas-Con,
          <string-name>
            <given-names>Mareike</given-names>
            <surname>Haug</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Heinrich</given-names>
            <surname>Hussmann</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Bringing Transparency Design into Practice</article-title>
          .
          <source>In 23rd International Conference on Intelligent User Interfaces (IUI '18)</source>
          ,
          <fpage>211</fpage>
          -
          <lpage>223</lpage>
          . https://doi.org/10.1145/3172944.3172961 Electronic Privacy Information Center.
          <year>2018</year>
          .
          <article-title>EPIC - Algorithms in the Criminal Justice System</article-title>
          .
          <source>Retrieved November 5</source>
          ,
          <year>2018</year>
          from https://epic.org/algorithmic-transparency/crim-justice/ Motahhare Eslami, Karrie Karahalios, Christian Sandvig, Kristen Vaccaro, Aimee Rickman, Kevin Hamilton, and
          <string-name>
            <given-names>Alex</given-names>
            <surname>Kirlik</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>First I like it, then I hide it: Folk Theories of Social Feeds</article-title>
          .
          <source>In Proceedings of the 2016 cHI conference on human factors in computing systems</source>
          ,
          <volume>2371</volume>
          -
          <fpage>2382</fpage>
          . Retrieved April 25,
          <year>2017</year>
          from http://dl.acm.org/citation.cfm?id=2858494 Motahhare Eslami, Aimee Rickman, Kristen Vaccaro, Amirhossein Aleyasen, Andy Vuong, Karrie Karahalios, Kevin Hamilton, and
          <string-name>
            <given-names>Christian</given-names>
            <surname>Sandvig</surname>
          </string-name>
          .
          <year>2015</year>
          . “
          <string-name>
            <given-names>I</given-names>
            <surname>Always Assumed That I Wasn'T Really</surname>
          </string-name>
          That Close to [Her]
          <article-title>”: Reasoning About Invisible Algorithms in News Feeds</article-title>
          .
          <source>In Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems (CHI '15)</source>
          ,
          <fpage>153</fpage>
          -
          <lpage>162</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          https://doi.org/10.1145/2702123.2702556
          <string-name>
            <given-names>Harold</given-names>
            <surname>Garfinkel</surname>
          </string-name>
          .
          <year>1991</year>
          .
          <article-title>Studies in Ethnomethodology</article-title>
          . Wiley.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <given-names>H. P</given-names>
            <surname>Grice</surname>
          </string-name>
          .
          <year>1975</year>
          .
          <article-title>Logic and conversation</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <given-names>Chloe</given-names>
            <surname>Gui</surname>
          </string-name>
          and
          <string-name>
            <given-names>Victoria</given-names>
            <surname>Chan</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Machine learning in medicine</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <source>University of Western Ontario Medical Journal</source>
          <volume>86</volume>
          ,
          <issue>2</issue>
          :
          <fpage>76</fpage>
          -
          <lpage>78</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          https://doi.org/10.5206/uwomj.v86i2.2060
          <string-name>
            <surname>Jonathan L. Herlocker</surname>
          </string-name>
          , Joseph A.
          <string-name>
            <surname>Konstan</surname>
            ,
            <given-names>and John</given-names>
          </string-name>
          <string-name>
            <surname>Riedl</surname>
          </string-name>
          .
          <year>2000</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <article-title>Explaining collaborative filtering recommendations</article-title>
          .
          <source>In Proceedings of the 2000 ACM conference on Computer supported cooperative work - CSCW '00</source>
          ,
          <fpage>241</fpage>
          -
          <lpage>250</lpage>
          . https://doi.org/10.1145/358916.358995
          <string-name>
            <given-names>IUI</given-names>
            <surname>ATEC</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>IUI ATEC Call for Papers</article-title>
          .
          <source>Retrieved January 10</source>
          ,
          <year>2019</year>
          from https://iuiatec.files.wordpress.com/
          <year>2018</year>
          /09/iui-atec-2019
          <string-name>
            <surname>-</surname>
          </string-name>
          call-forpapers.
          <source>pdf René F. Kizilcec</source>
          .
          <year>2016</year>
          . How Much Information?:
          <source>Effects of Transparency on Trust in an Algorithmic Interface</source>
          .
          <fpage>2390</fpage>
          -
          <lpage>2395</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          https://doi.org/10.1145/2858036.2858402
          <string-name>
            <given-names>Todd</given-names>
            <surname>Kulesza</surname>
          </string-name>
          , Simone Stumpf, Margaret Burnett, and
          <string-name>
            <given-names>Irwin</given-names>
            <surname>Kwan</surname>
          </string-name>
          .
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <string-name>
            <given-names>Tell</given-names>
            <surname>Me</surname>
          </string-name>
          <article-title>More?: The Effects of Mental Model Soundness on Personalizing an Intelligent Agent</article-title>
          .
          <source>In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI '12)</source>
          ,
          <fpage>1</fpage>
          -
          <lpage>10</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          https://doi.org/10.1145/2207676.2207678 Brian
          <string-name>
            <given-names>Y.</given-names>
            <surname>Lim and Anind</surname>
          </string-name>
          <string-name>
            <given-names>K.</given-names>
            <surname>Dey</surname>
          </string-name>
          .
          <year>2011</year>
          .
          <article-title>Investigating intelligibility for uncertain context-aware applications</article-title>
          .
          <source>In Proceedings of the 13th international conference on Ubiquitous computing</source>
          ,
          <fpage>415</fpage>
          -
          <lpage>424</lpage>
          . Retrieved April 25,
          <year>2017</year>
          from http://dl.acm.org/citation.cfm?id=2030168 Brian
          <string-name>
            <given-names>Y.</given-names>
            <surname>Lim</surname>
          </string-name>
          ,
          <string-name>
            <surname>Anind K. Dey</surname>
            , and
            <given-names>Daniel</given-names>
          </string-name>
          <string-name>
            <surname>Avrahami</surname>
          </string-name>
          .
          <year>2009</year>
          .
          <article-title>Why and why not explanations improve the intelligibility of context-aware intelligent systems</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <source>In Proceedings of the 27th international conference on Human factors in computing systems - CHI 09</source>
          ,
          <year>2119</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          https://doi.org/10.1145/1518701.1519023
          <string-name>
            <surname>Cathy O'Neil</surname>
          </string-name>
          .
          <year>2016</year>
          . Weapons of Math Destruction:
          <article-title>How Big Data Increases Inequality and Threatens Democracy</article-title>
          . Crown.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          <string-name>
            <given-names>Eli</given-names>
            <surname>Pariser</surname>
          </string-name>
          .
          <year>2011</year>
          .
          <article-title>The Filter Bubble: What The Internet Is Hiding From You</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          <string-name>
            <surname>PricewaterhouseCoopers.</surname>
          </string-name>
          <year>2018</year>
          . Consumer Intelligence Series:
          <article-title>Prepare for the voice revolution</article-title>
          .
          <source>PwC. Retrieved October 30</source>
          ,
          <year>2018</year>
          from https://www.pwc.com/us/en/services/consulting/library/consumerintelligence-series/voice-assistants.
          <source>html Inioluwa Deborah Raji and Joy Buolamwini. Actionable Auditing: Investigating the Impact of Publicly Naming Biased Performance Results of Commercial AI Products</source>
          .
          <volume>7</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          <string-name>
            <given-names>Christian</given-names>
            <surname>Sandvig</surname>
          </string-name>
          , Kevin Hamilton, Karrie Karahalios, and
          <string-name>
            <given-names>Cedric</given-names>
            <surname>Langbort</surname>
          </string-name>
          .
          <source>Auditing Algorithms: Research Methods for Detecting Discrimination on Internet Platforms</source>
          .
          <volume>23</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          <string-name>
            <given-names>James</given-names>
            <surname>Schaffer</surname>
          </string-name>
          , Prasanna Giridhar, Debra Jones, Tobias Höllerer, Tarek Abdelzaher, and
          <string-name>
            <surname>John O'Donovan</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <article-title>Getting the Message?: A Study of Explanation Interfaces for Microblog Data Analysis</article-title>
          .
          <source>In Proceedings of the 20th International Conference on Intelligent User Interfaces - IUI '15</source>
          ,
          <fpage>345</fpage>
          -
          <lpage>356</lpage>
          . https://doi.org/10.1145/2678025.2701406 Aaron Springer and Henriette Cramer.
          <year>2018</year>
          .
          <article-title>“Play PRBLMS”: Identifying and Correcting Less Accessible Content in Voice Interfaces</article-title>
          .
          <source>In Proceedings [32] [33] [34] [35] [36] [37] [38] [39] [40] of the 2018 CHI Conference on Human Factors in Computing Systems (CHI '18)</source>
          ,
          <volume>296</volume>
          :
          <fpage>1</fpage>
          -
          <lpage>296</lpage>
          :
          <fpage>13</fpage>
          . https://doi.org/10.1145/3173574.3173870 Aaron Springer, Jean Garcia-Gathright, and
          <string-name>
            <given-names>Henriette</given-names>
            <surname>Cramer</surname>
          </string-name>
          .
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          <source>In 2018 AAAI Spring Symposium Series.</source>
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          Aaron Springer, Victoria Hollis, and
          <string-name>
            <given-names>Steve</given-names>
            <surname>Whittaker</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Dice in the Black Box: User Experiences with an Inscrutable Algorithm</article-title>
          .
          <source>Retrieved April 24</source>
          ,
          <year>2017</year>
          from https://aaai.org/ocs/index.php/SSS/SSS17/paper/view/15372 Aaron Springer and Steve Whittaker.
          <year>2018</year>
          .
          <article-title>What are You Hiding? Algorithmic Transparency and User Perceptions</article-title>
          .
          <source>In 2018 AAAI Spring Symposium Series.</source>
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          <source>Aaron Springer and Steve Whittaker</source>
          .
          <year>2019</year>
          .
          <article-title>Progressive Disclosure: Designing for Effective Transparency</article-title>
          .
          <source>In Proceedings of the 24th International Conference on Intelligent User Interfaces - IUI '19.</source>
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          <string-name>
            <given-names>Latanya</given-names>
            <surname>Sweeney</surname>
          </string-name>
          .
          <year>2013</year>
          .
          <article-title>Discrimination in Online Ad Delivery</article-title>
          .
          <source>Queue</source>
          <volume>11</volume>
          ,
          <issue>3</issue>
          :
          <fpage>10</fpage>
          :
          <fpage>10</fpage>
          -
          <lpage>10</lpage>
          :
          <fpage>29</fpage>
          . https://doi.org/10.1145/2460276.2460278
          <string-name>
            <given-names>Rachael</given-names>
            <surname>Tatman</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Gender and Dialect Bias in YouTube's Automatic Captions</article-title>
          .
          <source>EACL</source>
          <year>2017</year>
          :
          <volume>53</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          <string-name>
            <given-names>Jeffrey</given-names>
            <surname>Warshaw</surname>
          </string-name>
          , Nina Taft, and
          <string-name>
            <given-names>Allison</given-names>
            <surname>Woodruff</surname>
          </string-name>
          .
          <year>2016</year>
          . Intuitions, Analytics, and Killing Ants:
          <article-title>Inference Literacy of High School-educated Adults in the US</article-title>
          .
          <volume>16</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          <string-name>
            <given-names>Daniel S.</given-names>
            <surname>Weld</surname>
          </string-name>
          and
          <string-name>
            <given-names>Gagan</given-names>
            <surname>Bansal</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>The Challenge of Crafting Intelligible Intelligence</article-title>
          . arXiv:
          <year>1803</year>
          .04263 [cs].
          <source>Retrieved September 20</source>
          ,
          <year>2018</year>
          from http://arxiv.org/abs/
          <year>1803</year>
          .04263
          <string-name>
            <given-names>Jenna</given-names>
            <surname>Wiens</surname>
          </string-name>
          and
          <string-name>
            <given-names>Erica S.</given-names>
            <surname>Shenoy</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Machine Learning for Healthcare: On the Verge of a Major Shift in Healthcare Epidemiology</article-title>
          .
          <source>Clinical Infectious Diseases</source>
          <volume>66</volume>
          ,
          <issue>1</issue>
          :
          <fpage>149</fpage>
          -
          <lpage>153</lpage>
          . https://doi.org/10.1093/cid/cix731
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>