<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>The Empathy Gap: Why AI Can Forecast Behavior But Cannot Assess Trustworthiness</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Jason R. D'Cruz</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>William Kidder</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Kush R. Varshney</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>IBM Research - Thomas J. Watson Research Center</institution>
          ,
          <addr-line>1101 Kitchawan Road, Yorktown Heights, NY, 10598</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Status Labs</institution>
          ,
          <addr-line>701 Tillery Street, Austin, TX, 78702</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>University at Albany, State University of New York</institution>
          ,
          <addr-line>1400 Washington Avenue, Albany, NY, 12222</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>In previous work we have sought to characterize “trustworthy AI” (Varshney 2022, Knowles et al. 2022). In this work, we examine the case of AI systems that appear to render verdicts about our (human) trustworthiness, and we inquire into the conditions under which we can trust AI systems to trust us appropriately. We argue that the inability to take on another's perspective (henceforth, “empathy deficit”) can both explain and justify our distrust of AI in domains in which AI is tasked with forecasting the likelihood of human (un)trustworthiness. Examples include the use of AI to make forecasts for parole and bail eligibility, academic honesty, and creditworthiness. Humans have an interest in ensuring that judgments of our trustworthiness are based on some degree of empathic understanding of our reasons and unique circumstances. The inability of AI to adopt our subjective perspective calls into question our trust in AI systems's assessments of human trustworthiness. In 2016, a default was registered on my credit file by my bank after I missed three payments of my student overdraft. During this time, I was in an abusive and controlling relationship. I had no money of my own. Shortly before my son was born in the same month, the default was registered. I was forced to move hundreds of miles away to a new city. I wasn't able to divert post. My only source of income was statutory maternity pay, which was</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;empathy</kwd>
        <kwd>moral reasoning</kwd>
        <kwd>trustworthiness</kwd>
        <kwd>artificial intelligence</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>given straight to my partner. When I left him and began to rebuild my life, I was
made aware of the default and immediately repaid it at £60 a month.</p>
      <p>In her compassionate reply, Tims points out that a credit score is a “blunt instrument” that
registers the default “without the circumstances”. She describes GH’s predicament as “heart
rending”.</p>
      <p>
        Where the credit scoring algorithm registered only the missed payments, Tims was able to
apprehend the reasons for the missed payments [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Granted, a more sophisticated machine
learning system fed with more data may eventually discern that whether a person meets financial
commitments when they are mired in an abusive relationship is not predictive of whether they
will meet financial commitments after the abuse has ceased. But what the algorithm will not
register is the fact that being bullied and abused is a morally excusing condition for failing to
meet a financial commitment. That is, the algorithm will not register that it is inappropriate
to adopt the “reactive attitudes” [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] of blame and moral approbation toward a person in such
circumstances because the missed payments do not reveal a lack of conscientiousness on GH’s
part. It is therefore not fair to assess GH as untrustworthy because of them.
      </p>
      <p>
        Most machine learning (ML) tasks are “anti-causal”, inferring causes (labels) from efects
(observations) [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. A machine learning system may discern that, say, a Tottenham resident is
more likely to repay loans when Spurs are winning than when they are losing. But the fact that
Spurs are losing is not an excusing condition for missing a payment.
      </p>
      <p>For the purposes of accurate prediction and eficient allocation of resources, it does not
matter whether the explanation for GH’s default is morally excusing or not. But from GH’s
perspective, a diminishment of her credit score in non-excusing circumstances is unassailable
even if unwelcome, while a diminishment in her credit score in excusing circumstances is
patently unjust. Had GH’s letter been about Spurs losing rather than about domestic abuse,
we can imagine Anna Tims responding with cool dismissal rather than solidarity and helpful
advice.</p>
      <p>
        There is no reason to think that the class of considerations that matter for judgments of
trustworthiness by empathic human observers is coextensive with the class of features that
will increase an ML system’s accuracy in forecasting behavior. The task of forecasting and the
task of assessing trustworthiness are fundamentally diferent, and accordingly require diferent
capacities. Even if decision subjects are granted [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]’s proposal for the ‘right to be an exception’
to an ML system, the ML system itself cannot be the one that makes an empathic decision
regarding the trustworthiness of the individuals that have excusing circumstances.
      </p>
      <p>Anna Tims did something that is routine for humans, but that no extant ML system, no matter
how sophisticated, has the capacity to do. She was able to empathize with GH’s position, putting
herself in GH’s shoes and inviting her readers to do the same. This allowed her to understand
why GH was not able to pay her bills, and to arrive at the judgment that the circumstances were
morally excusing. Her empathic response allowed GH to be recognized as an individual with a
unique perspective and set of experiences. More than mere advice, Tims could ofer validation.</p>
      <p>
        Even ML systems that ascend Pearl’s ladder of causation to the level of counterfactual
imagination do not thereby display a capacity for empathy [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Interventions and counterfactual
imagination may allow systems to better understand why GH did not pay her bills, but do not
allow them to reason about moral excuses.
      </p>
      <p>We naturally interpret decisions that preclude us from opportunities because of our perceived
riskiness as derogations of our character. We want judgments about whether we will be trusted
(with a loan, with bail, with work-from-home accommodations, etc.) to be sensitive to whether
we deserve to be trusted or not. Such judgments necessarily take into account reasons in context,
features that are intuitively apprehended by other humans via empathy.</p>
      <p>We can of course envision an improved algorithm fed with more data and optimized with
considerations of social justice in mind. In an updated schema, victims of a documented history
of domestic abuse could be aforded an amnesty from derogatory marks on a credit report. But
notice that the judgment that the algorithm ought to be updated in this way must be made
by an entity with normative competence. We maintain that, in the context of assessments of
trustworthiness, such competence includes being able to take up the perspective of the person
being assessed.</p>
    </sec>
    <sec id="sec-2">
      <title>2. The Role of Empathy in Calibrating Trustworthiness</title>
      <p>Knowing a person’s track record is essential in gauging their trustworthiness. But a track
record still needs to be interpreted, especially when departures from a pattern are not easily
legible. This will typically require understanding and assessing the reasons behind the relevant
actions. Empathy allows us to uncover reasons that justify attitudes of trust or distrust. It is
therefore a tool to gather evidence of (un)trustworthiness that cannot be accessed by merely
assessing a track record. Deprived of this tool we are at a marked disadvantage in calibrating
trust. When others lack this tool, we have grounds for skepticism about their capacity to assess
our trustworthiness.</p>
      <p>
        The kind of evidence that is aforded by empathy, crucial to making judgments of
(un)trustworthiness, is conspicuously inaccessible to purely data- and pattern-driven “system 1” AI
systems [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. Even knowledge- and reasoning-driven “system 2” AI systems and hybrid
neurosymbolic AI systems are not at a point that can interpret moral reasons in context. (Moral sense
reasoning is more dificult than common sense reasoning, which AI systems are still far from
mastering.)
      </p>
      <p>It is true that attitudes of trust and distrust are continuously updated with feedback about
the trustworthiness of one’s trustees, which is often a matter of whether they behave as you
expect them to. But a trustee’s (un)trustworthiness is not always manifest in their performance
or non-performance of the actions we expect of them. There is an enormous amount of room
for interpretation.</p>
      <p>Consider: I trust you to pick me up at the airport one morning but you are nowhere to be
found. I later learn that the reason you did not show up is that your partner needed to be rushed
to the hospital after sufering an anaphylactic reaction. It would be impossible for you to pick
me up without endangering her health, and there was no time to warn me that you cannot
come. You have not acted as I was counting on you to act, but you have not shown yourself to
be untrustworthy in any measure. Quite the contrary: I would not be able to trust you with
anything important in the future were you to abandon your partner in a moment of acute need
just to meet your commitment to pick me up at the airport. This mundane story illustrates an
important lesson about trustworthiness. Whether a person is trustworthy is not fully legible
in the bare facts of their action, that is, in whether they do what you expect of them. Rather,
determination of trustworthiness requires interpreting information about the context and their
reasons.</p>
      <p>Consider now the alternate possibility that you show up at the airport, but you would have
forgotten were it not the case that a radio story that morning mentioned airports and your
memory was jogged. In this case you have acted in the way that I expected of you, but in a very
close possible world you would not have. Moreover, the explanation for the closeness of that
possible world is that you took insuficient care to ensure that you would not forget. Were I to
learn that you showed up in part because you got lucky, I would not be inclined to judge that
you are trustworthy or to trust you more deeply in the future.</p>
      <p>That trustors shoulder an inescapable interpretive burden is a feature of any plausible account
of trust and trustworthiness. On Jones’ account, a person is “three-place trustworthy” only if she
is competent with respect to the domain of trust and “would take the fact that A is counting on
her [...] as a compelling reason for acting if counted upon” [9, p. 70–71]. Note that a “compelling
reason” need not be an overriding reason. Compelling reasons may be outweighed by other
reasons that are even more compelling. As Jones puts it, “There is a necessary vagueness about
what it is to take a reason to be compelling, and there can be disagreement over whether an
agent has in fact done this” [9, p. 71]. There is always the risk that a person is unfairly judged
untrustworthy as a result of unreasonable expectations. There is also the risk that they disguise
their untrustworthiness behind specious claims that they face pressing countervailing reasons.</p>
      <p>
        On Hawley’s commitment account, trustworthiness requires a person to ensure that their
commitments do not outstrip their actions [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. As Hawley puts it, “This requires judiciousness
in acquiring commitments as well as doggedness in fulfilling commitments already acquired,
independent of others’ expectations. Trustworthy people must sometimes disappoint up-front
by refusing new commitments, rather than violate trust later on: this is the moral ‘power of
no’.” But even the power of ‘no’ has its limits. A person can be blameless in failing to foresee
circumstances in which meeting one commitment will mean falling short with respect to another,
as in the case in which one’s commitment to caring for one’s partner in an emergency outweighs
one’s commitment to pick up a friend at the airport. A broken commitment does not impugn
trustworthiness ipso facto.
      </p>
      <p>Meeting your commitments and being responsive to the dependency of those who count on
you are crucial and closely related aspects to being a trustworthy person. We remain neutral on
which of the two is more fundamental. More important for our argument is that appraisals of a
person as trustworthy or untrustworthy require appreciating the full range of competing reasons
that trustees face. Trustors need to be able to fit their trustee’s actions into an interpretable story
of trustworthiness or untrustworthiness. This requires of trustors that they project themselves
into the deliberative position of their trustees, something that only conscious beings with
normative competence (for the proximate future at least, only humans) are able to do.</p>
      <p>
        Granted, there are domains in which track record alone is suficient evidence of an AI system’s
competence to perform the task entrusted to it, such as detecting a melanoma from a dermoscopic
image or determining the best move in a game of chess. We focus on domains in which AI is
tasked with a particularly delicate and morally freighted commission: making forecasts that
humans will naturally interpret as assessments of their trustworthiness [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ]. Earning trust
in one’s ability to assess trustworthiness requires credible demonstration of the capacity to
slow down and take on the perspective of the target of the assessment. This is a capacity that
cannot be fully compensated by computational virtuosity or predictive accuracy, but requires a
highly advanced level of knowledge and reasoning. In these domains, AI’s empathy deficit both
explains and justifies distrust of AI.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. The Dishonor and Self-Fulfillment of Distrust</title>
      <p>From the perspective of the deployer, AI systems are useful tools for making more accurate
forecasts, managing risk and allocating resources most eficiently. But the stakes are very
diferent for the individual whose trustworthiness is being assessed.</p>
      <p>
        Being categorized as “too risky” is inevitably experienced as a judgment of one’s character.
To be denied opportunities because one is deemed “a bad bet” is deeply dishonoring. Distrust
without warrant risks insulting, demoralizing, and disempowering, planting the seeds of
behavior and character revision that does warrant distrust. Particularly in settings in which
trusting behavior is the norm and distrusting behavior is therefore conspicuous, the risk of
marginalization and dishonor represents significant, even dramatic, losses for the party who is
wrongly distrusted and for the relationship as well. Being cut of from opportunity in a setting
in which others are given the benefit of the doubt can open deep wounds that are slow to heal
[
        <xref ref-type="bibr" rid="ref11">11</xref>
        ].
      </p>
      <p>That we care very much about how we are perceived is an enduring feature of human nature.
Adam Smith, the founding father of economics, thought that the desire for standing in the eyes
of others was one of the most basic of human inclinations.</p>
      <p>Nature, when she formed man for society, endowed him with an original desire
to please, and an original aversion to ofend his brethren. She taught him to feel
pleasure in their favorable, and pain in their unfavorable regard. She rendered their
approbation most flattering and most agreeable to him for its own sake; and their
disapprobation most mortifying and most ofensive.</p>
      <p>
        It is easy to underestimate the degree to which a person who is the object of distrust is
aware that they are so regarded. Slepian and Ames explore the efects of such awareness
empirically [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. Individuals have an intuitive sense of whether or not others perceive them as
trustworthy, and tend to behave in a way that is consistent with how they imagine themselves
to be perceived. They find that face-based judgments of strangers predict trustworthy behavior
in a game involving trust and deception. But this link is mediated by subjects’ expectations
of how other people perceive them and in particular their expectation of whether they would
be trusted by others. As they put it, subjects “seemed to have an awareness of how people
would judge them, and they internalized these expectations and behaved in accordance with
them” [12, p. 287]. This work extends previous work in social psychology on “self-fulfilling
prophecies” [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] that focus on individual contexts (e.g. whether a student will conform to a
teacher’s high expectations). The meta-perceptions that Slepian and Ames find are not derived
from a particular interaction but rather likely “from a range of contexts across a lifetime of
treatment” [12, p. 287].
      </p>
      <p>Just as trustworthiness is cultivated and reinforced by trust, so also untrustworthiness is
cultivated and reinforced by distrust. A lamentable consequence of this recursive pattern of
trust- and distrust-responsiveness is that knowledge that a person is widely distrusted, whether
or not such distrust is merited, provides (pro tanto) evidence that they are untrustworthy.
Conversely, a distrusted person who has a grasp of the interpretive biasing that is a signature
of distrustful attitudes has (pro tanto) reason to believe that entirely innocent actions are likely
to be interpreted as indicating untrustworthiness. This person will not trust others to trust him.</p>
      <p>The mere anticipation of such mis-recognition diminishes the motivation to be responsive to
trust. All of this is the perfect recipe for a self-reinforcing and pernicious equilibrium. Doubtless
there is a logic to distrusting those who are broadly distrusted, just as there is a logic to giving
up on winning the trust of one’s distrustful fellows.</p>
      <p>We have moral reasons to disrupt the aforementioned pernicious equilibrium because it
darkens our prospects for living in harmony, equality, solidarity, and moral community.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Empathy’s Liabilities and Distinctive Value</title>
      <p>
        One implication of the view we advance is that we may be able to cultivate greater trust in
AI by improving a system’s capacity to display empathy (for example, by creating chatbots
that appear to share our emotions) [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. However, in systems that assess trustworthiness, this
deceptive strategy has serious practical, moral, and political liabilities. Empathy itself, and not
just its appearance, provides grounds for trusting others’ assessments of our trustworthiness.
Cultivating trust in the absence of its grounds is deceptive and risky. Liao and Sundar describe
how the trustworthiness of a technology can be communicated through “trustworthiness cues”
that are embedded in interface features [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. Misleading cues about an AI’s underlying lack of
empathic capacity risks overtrust.
      </p>
      <p>
        But empathy’s critics have argued that even empathy itself is risky. As Bloom puts the point,
empathy’s “spotlight nature renders it innumerate and myopic: It doesn’t resonate properly
to the efects of our actions on groups of people, and it is insensitive to statistical data and
estimated costs and benefits” [ 16, p. 31]. Empathy is also susceptible to intergroup biases: We
tend to be more capable of empathizing with those who share our group identities and less
capable of empathizing with those who do not [
        <xref ref-type="bibr" rid="ref17 ref18">17, 18</xref>
        ].
      </p>
      <p>
        Furthermore, nothing we say here rules out the possibility that algorithmic approaches to
prediction will often fare better than human judgment in forecasting untrustworthy behavior.
Indeed, [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] finds that a simple model that uses just two inputs — age, and past court dates
missed — fares better than almost all human bail judges in predicting flight risk.
      </p>
      <p>Yet, despite empathy’s liabilities and the forecasting capabilities of algorithms, empathy’s
distinctive value justifies the inclusion of an empathic agent in the loop of decision making
when it comes to judgments of trustworthiness. The distinctive value of empathy is its ability
to recognize morally excusing circumstances such as those involved in GH’s case. This value
is directly tied to the “spotlight nature” of empathy that troubles Bloom. It is true that in
many cases of moral judgment we do not want such a spotlight to guide our decisions, for
example when determining the most efective way to distribute disaster relief. But when the
question at hand is one of assessing the trustworthiness of an individual, a spotlight is what is
needed. We want the nuances of an individual case to be considered. We want morally excusing
circumstances to be recognized. Empathy is ideally suited to this task precisely because of its
spotlight nature, its ability to allow us to project ourselves into an individual perspective and
appreciate the moral significance of unique circumstances.</p>
      <p>Yet, given that ML systems will often outperform human beings when it comes to accurate
forecasting in these domains, there is a tradeof in requiring the involvement of empathy in
assessments of trustworthiness. We acknowledge empathy’s susceptibility to bias and do not
dispute AI’s forecasting potential, but we maintain that empathy is uniquely suited to enable the
recognition of morally excusing conditions in a way that is inaccessible to current AI systems.
Insofar as the capacity for this recognition is involved in a just assessment of our trustworthiness,
then empathy will be involved in that assessment, and AI, no matter how accurate its forecasts,
is currently unable to make the same sort of assessment.</p>
      <p>Thus, the questions of forecasting accuracy and assessment of trustworthiness, while often
entangled, can be distinguished. Just assessments of trustworthiness should involve more than
forecasting based on track record. They should involve reasoning in a way such that morally
excusing conditions can be recognized, and empathy enables this sort of recognition.</p>
      <p>Of course, a “system 1” machine learning system equipped with enough facts could perhaps
become better at identifying morally excusing conditions, but only if doing so led to better
forecasts. It might turn out that those who experience abuse are more likely to meet financial
commitments than similar individuals who have not experienced abuse. Or it might turn out
that they are not. If the latter is the case, then the system will not consider the abuse as relevant
to the forecast. But being dominated by an abuser is an excusing condition for failing to meet a
ifnancial commitment; it is a context with moral weight that requires normative competence
to grasp. The problem with extant AI systems, even neuro-symbolic ones, is that they lack a
means of determining conditions as morally salient. On the other hand, empathy allows us to
appreciate moral salience by taking on an individual’s perspective and contextualizing their
actions.</p>
      <p>In cases in which the question is precisely about the nuances of an individual’s situation
and perspective, merely gaining more facts about our behavior is not enough for normative
competence when assessing trustworthiness. We want to be understood as more than a collection
of these facts when the question is our trustworthiness, our character. In these cases, a deeper
understanding of the reasons underlying our behavior is required, and empathy furnishes this
deeper understanding.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>In this paper we distinguish the task of forecasting behavior from the task of assessing human
trustworthiness. The former is particularly well-suited to machine learning systems while the
latter requires capacities that such systems lack. Machine learning systems make predictions
about human actions in the same way that they make predictions about natural phenomena
such as floods, earthquakes, and the weather. At times we take up this external perspective on
ourselves, too, asking, “How am I likely to behave?” rather than, “What should I do?”</p>
      <p>In contrast, when we assess an individual’s trustworthiness, we assess not only their actions
but also their reasons. This includes considerations of whether it is fair to count past lapses
against them or whether the presence of excusing conditions means that they are “owed the
benefit of the doubt”. Such evaluations require an understanding of why they acted in the way
that they did. Empathy furnishes us with the ability to occupy the perspective of another human
and understand their reasons. In empathizing we can ask questions such as: Is it reasonable
to expect of this person that they act diferently than they did? Is this action an expression of
their moral character or merely a product of the circumstance? What was it that prevented
them from acting conscientiously and what changes in themselves or in the environment might
make a diference?</p>
      <p>Machine learning systems that are fed with vast troves of data can excel at the task of
forecasting behavior and will likely improve. But forecasting behavior is not suficient for assessing
human trustworthiness, which has an essential normative component. Such assessments require
understanding a person’s reasons given the particular circumstances they face. This we learn
through empathy. Even the most advanced neuro-symbolic AI systems of today that combine
ML with reasoning ability face an empathy deficit that will be dificult to overcome. Empathy
presents a challenging frontier for AI research.</p>
      <p>
        None of this implies that there is not a useful role to AI to play in cultivating and supporting
human trustworthiness. Machine learning systems can be deployed to make forecasts about the
eficacy of interventions aimed at helping people to do what they are trusted to do (for example,
the eficacy of giving taxi vouchers to people dependent on unreliable public transportation
or text alerts about upcoming court dates for people living in chaotic environments). Such
a reorientation would have us abandon the idea that AI systems should serve as arbiters of
human trustworthiness and to think of AI systems as tools to assist humans in cultivating and
supporting social inclusion. Responsiveness to trust is an “ecological capacity” that depends both
on context and on character [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ]. AI is a powerful tool for helping us to create the conditions
for trustworthiness.
      </p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>This research is funded by the SUNY-IBM AI Research Alliance under grant number AI2102. The
authors declare that the research was conducted in the absence of any commercial or financial
relationships that could be construed as a potential conflict of interest.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>K. R.</given-names>
            <surname>Varshney</surname>
          </string-name>
          , Trustworthy Machine Learning, Independently Published, Chappaqua,
          <string-name>
            <surname>NY</surname>
          </string-name>
          , USA,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>B.</given-names>
            <surname>Knowles</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. D'Cruz</surname>
            ,
            <given-names>J. T.</given-names>
          </string-name>
          <string-name>
            <surname>Richards</surname>
            ,
            <given-names>K. R.</given-names>
          </string-name>
          <string-name>
            <surname>Varshney</surname>
          </string-name>
          ,
          <article-title>Humble machines: Attending to the underappreciated costs of misplaced distrust</article-title>
          ,
          <source>in: ACM Conference on Equity and Access in Algorithms</source>
          , Mechanisms, and
          <string-name>
            <surname>Optimization</surname>
            , Arlington,
            <given-names>VA</given-names>
          </string-name>
          , USA,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>T.</given-names>
            <surname>McCoy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Pavlick</surname>
          </string-name>
          , T. Linzen,
          <article-title>Right for the wrong reasons: Diagnosing syntactic heuristics in natural language inference, in: Proceedings of the Annual Meeting of the Association for Computational Linguistics</article-title>
          , Florence, Italy,
          <year>2019</year>
          , pp.
          <fpage>3428</fpage>
          -
          <lpage>3448</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>P. F.</given-names>
            <surname>Strawson</surname>
          </string-name>
          , Freedom and Resentment and
          <string-name>
            <given-names>Other</given-names>
            <surname>Essays</surname>
          </string-name>
          , Routledge, London, UK,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>N.</given-names>
            <surname>Kilbertus</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Parascandolo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Schölkopf</surname>
          </string-name>
          ,
          <article-title>Generalization in anti-causal learning</article-title>
          , arXiv:
          <year>1812</year>
          .00524,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>S. H.</given-names>
            <surname>Cen</surname>
          </string-name>
          ,
          <article-title>The Right to be an Exception in Data-Driven Decision-Making</article-title>
          ,
          <source>Technical Report</source>
          , Massachusetts Institute of Technology,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J.</given-names>
            <surname>Pearl</surname>
          </string-name>
          ,
          <article-title>The seven tools of causal inference, with reflections on machine learning</article-title>
          ,
          <source>Communications of the ACM</source>
          <volume>62</volume>
          (
          <year>2019</year>
          )
          <fpage>54</fpage>
          -
          <lpage>60</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>G.</given-names>
            <surname>Booch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Fabiano</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Horesh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kate</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Lenchner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Linck</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Loreggia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Murgesan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Mattei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Rossi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Srivastava</surname>
          </string-name>
          ,
          <article-title>Thinking fast and slow in AI</article-title>
          ,
          <source>in: Proceedings of the AAAI Conference on Artificial Intelligence</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>15042</fpage>
          -
          <lpage>15046</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>K.</given-names>
            <surname>Jones</surname>
          </string-name>
          , Trustworthiness, Ethics
          <volume>123</volume>
          (
          <year>2012</year>
          )
          <fpage>61</fpage>
          -
          <lpage>85</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>K.</given-names>
            <surname>Hawley</surname>
          </string-name>
          , How to Be Trustworthy, Oxford University Press, New York, NY, USA,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>J. D'Cruz</surname>
          </string-name>
          , Humble trust,
          <source>Philosophical Studies</source>
          <volume>176</volume>
          (
          <year>2019</year>
          )
          <fpage>933</fpage>
          -
          <lpage>953</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>M. L. Slepian</surname>
            ,
            <given-names>D. R.</given-names>
          </string-name>
          <string-name>
            <surname>Ames</surname>
          </string-name>
          ,
          <article-title>Internalized impressions: The link between apparent facial trustworthiness and deceptive behavior is mediated by targets' expectations of how they will be judged</article-title>
          ,
          <source>Psychological Science</source>
          <volume>27</volume>
          (
          <year>2016</year>
          )
          <fpage>282</fpage>
          -
          <lpage>288</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>R.</given-names>
            <surname>Rosenthal</surname>
          </string-name>
          ,
          <article-title>Interpersonal expectancy efects: A 30-year perspective</article-title>
          ,
          <source>Current Directions in Psychological Science</source>
          <volume>3</volume>
          (
          <year>1994</year>
          )
          <fpage>176</fpage>
          -
          <lpage>179</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>J.</given-names>
            <surname>Howick</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Morley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Floridi</surname>
          </string-name>
          ,
          <article-title>An empathy imitation game: Empathy Turing test for careand chat-bots</article-title>
          ,
          <source>Minds and Machines</source>
          <volume>31</volume>
          (
          <year>2021</year>
          )
          <fpage>457</fpage>
          -
          <lpage>461</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>Q. V.</given-names>
            <surname>Liao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. S.</given-names>
            <surname>Sundar</surname>
          </string-name>
          ,
          <article-title>Designing for responsible trust in AI systems: A communication perspective</article-title>
          ,
          <source>in: Proceedings of the ACM Conference on Fairness, Accountability, and Transparency</source>
          , Seoul, Korea,
          <year>2022</year>
          , pp.
          <fpage>1257</fpage>
          -
          <lpage>1268</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>P.</given-names>
            <surname>Bloom</surname>
          </string-name>
          , Against Empathy:
          <article-title>The Case for Rational Compassion</article-title>
          , HarperCollins Publishers, New York, NY, USA,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>M.</given-names>
            <surname>Cikara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. G.</given-names>
            <surname>Bruneau</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. R.</given-names>
            <surname>Saxe</surname>
          </string-name>
          ,
          <article-title>Us and them: Intergroup failures of empathy</article-title>
          ,
          <source>Current Directions in Psychological Science</source>
          <volume>20</volume>
          (
          <year>2011</year>
          )
          <fpage>149</fpage>
          -
          <lpage>153</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>M.</given-names>
            <surname>Cikara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Bruneau</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. J. Van Bavel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Saxe</surname>
          </string-name>
          ,
          <article-title>Their pain gives us pleasure: How intergroup dynamics shape empathic failures and counter-empathic responses</article-title>
          ,
          <source>Journal of Experimental Social Psychology</source>
          <volume>55</volume>
          (
          <year>2014</year>
          )
          <fpage>110</fpage>
          -
          <lpage>125</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>J.</given-names>
            <surname>Jung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Concannon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Shrof</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Goel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. G.</given-names>
            <surname>Goldstein</surname>
          </string-name>
          ,
          <article-title>Simple rules for complex decisions</article-title>
          ,
          <source>arXiv:1702.04690</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>V.</given-names>
            <surname>McGeer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Pettit</surname>
          </string-name>
          ,
          <article-title>The empowering theory of trust</article-title>
          , in: P. Faulkner, T. Simpson (Eds.),
          <source>The Philosophy of Trust</source>
          , Oxford University Press, Oxford, UK,
          <year>2017</year>
          , pp.
          <fpage>14</fpage>
          -
          <lpage>34</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>