<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Sands of Time: A Reliabilistic Account of Justified Credence in the Trustworthiness of AI Systems</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Andrea Ferrario</string-name>
          <email>aferrario@ethz.ch</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Extended Abstract</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>ETH</institution>
          ,
          <addr-line>Zurich</addr-line>
          ,
          <country country="CH">Switzerland</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>EWAF'23: European Workshop on Algorithmic Fairness</institution>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Mobiliar Lab for Analytics at ETH</institution>
          ,
          <addr-line>Zurich</addr-line>
          ,
          <country country="CH">Switzerland</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>We address an open problem in the epistemology of artificial intelligence (AI), namely, the justification of the epistemic attitudes we have towards the trustworthiness of AI systems. We start from a key consideration: the trustworthiness of an AI is a time-relative property of the system, with two distinct facets. One is the actual trustworthiness of the AI, and the other is the perceived trustworthiness of the system as assessed by its users while interacting with it. We show that credences, namely, beliefs we hold with a degree of confidence, are the appropriate attitude for capturing the facets of trustworthiness of an AI over time. Then, we introduce a reliabilistic account providing justification to the credence in the trustworthiness of AI, which we derive from Tang's probabilistic theory of justified credence. Our account stipulates that a credence in the trustworthiness of an AI system is justified if and only if it is caused by an assessment process that tends to result in a high proportion of credences for which the actual and perceived trustworthiness of the AI are calibrated. Our approach informs research on human-AI interactions and trustworthy AI by providing actionable recommendations on how to measure the reliability of the process through which users perceive the trustworthiness of the system and its calibration to the actual levels of trustworthiness of the AI. It also allows investigating the relation between reliability and the appropriate reliance on the system.</p>
      </abstract>
      <kwd-group>
        <kwd>artificial intelligence</kwd>
        <kwd>trustworthiness</kwd>
        <kwd>trustworthy AI</kwd>
        <kwd>epistemology</kwd>
        <kwd>justification</kwd>
        <kwd>reliabilism</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        that the treatment [suggested by the AI] is likely to benefit the patient” [ 4, p. 331]. Ethical, as
the deployment of trustworthy AI systems mitigate the risk of unethical decisions afecting the
lives of many people [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
      <p>
        Recent studies on the epistemological perspective of trusting AI systems have focused on the
grounds on which we can state that our beliefs regarding the trustworthiness of AI predictions
are justified [
        <xref ref-type="bibr" rid="ref11 ref4">11, 4</xref>
        ]. To this end, authors discuss the account of “computational reliabilism”
(CR) for AI systems, which states that we are justified in believing that the predictions of an
AI system are trustworthy because they are generated by a reliable process, i.e., the AI system
itself [
        <xref ref-type="bibr" rid="ref11 ref12 ref4">12, 11, 4</xref>
        ]. The reliability of the AI is described as the tendency to generate predictions
that are worthy of our trust [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. Then, to avoid circularity, a clear definition of trustworthiness
is needed. However, the current implementation of CR to AI systems falls short of providing
it. As a result, at the time of writing, it is still not clear what we hold when stating that we
believe that an AI system—or any of its components—is worthy of trust. Relatedly, the reasons
why a belief could be the appropriate propositional attitude to describe our stance towards this
“trustworthiness” need further investigation. These research gaps afect the epistemology of AI
and the endeavors in human-AI interactions that investigate how to entrench, e.g., “calibrate,”
trust in the trustworthiness of these systems [
        <xref ref-type="bibr" rid="ref3 ref4 ref9">4, 3, 9</xref>
        ].
      </p>
      <p>Therefore, our goal is to introduce an account of justification of the epistemic attitudes
describing our stance towards the trustworthiness of AI systems. Our plan proposes three steps:
(1) reviewing related work on the provision of justifications of the beliefs on the trustworthiness
of AI, (2) answering the question “what is the trustworthiness of AI?,” and (3) arguing that
credences1 are more appropriate than beliefs to describe our stance towards the trustworthiness
of these systems [15]. This allows us introducing our account of justified credence in the
trustworthiness of AI following Tang’s probabilistic theory of reliability [14].</p>
      <p>
        In step (1), we propose to review related work on CR with a focus on its recent implementation
for AI systems [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], showing that this account has a problem with trustworthiness. In step (2),
we propose to take a time-relative perspective on AI systems, which we see as artefacts with a
life cycle that alternates between design and deployment phases. Doing so, we can show that, at
each step of the AI life cycle and moment of time, the trustworthiness of the AI comprises two,
measurable facets. The first—called “actual trustworthiness”—is the objective trustworthiness
of the AI that is measured during the design process of the system and is generally unknown
after its deployment. The second—called “perceived trustworthiness”—provides a point-in-time
perspective on how a deployed system is subjectively perceived as being worthy trust by its
diferent users [ 16]. It is measured by the users of the system instead. Finally, in step (3), we
propose to talk about credences rather than beliefs in the trustworthiness of an AI [15]. In
fact, we can show that a probabilistic account of vindicated credences [14] allows encoding the
diferent facets of the trustworthiness of an AI. In particular, their vindication calibrates the
degree of confidence of the credence, i.e., the level of perceived trustworthiness, to the actual
trustworthiness of the AI. Finally, we can introduce our account of justified credence in the
(actual) trustworthiness of AI systems. It is derived from Tang’s “(Probability)” reliabilistic
theory of justified credence [ 14] and it states that a credence in the trustworthiness of an AI
1A (binary) belief is the attitude we have “whenever we take something to be the case or regard it as true” [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ].
A credence is the epistemic attitude towards  that expresses a degree of confidence in  [14, 15].
system is justified if and only if it is the result of a reliable assessment process, i.e., a process to
assess the trustworthiness of AI that tends to result in a high proportion of vindicated credences.
      </p>
      <p>
        The advantages of our approach are manifold. First, it overcomes the limitations afecting the
existing account of CR for AI systems by appropriately encoding both facets of trustworthiness
in a reliabilist account of justified credence. Then, it allows introducing a measurable definition
of calibration between the perceived and actual trustworthiness of an AI. Finally, it allows
measuring the reliability of the assessment process of an AI explicitly, generalizing existing
reliability scoring of credence-forming processes [17, 18, 19, 20]. This, in turn, suggests the
design of empirical studies to investigate the relation between the reliability of the assessment
process, the diferent manipulations of the trustworthiness of the AI over time, the provision
of methodologies to support users’ assessments—similar to the “reliability indicators” in CR
[
        <xref ref-type="bibr" rid="ref11 ref4">11, 4</xref>
        ]—and the appropriate reliance on the AI’s predictions [21].
[14] W. H. Tang, Reliability theories of justified credence, Mind 125 (2016) 63–94.
[15] E. G. Jackson, The relationship between belief and credence, Philosophy Compass 15
(2020) e12668.
[16] Q. V. Liao, S. S. Sundar, Designing for responsible trust in AI systems: A communication
perspective, arXiv preprint arXiv:2204.13828 (2022).
[17] B. Lam, Calibrated probabilities and the epistemology of disagreement, Synthese 190
(2013) 1079–1098.
[18] R. Pettigrew, Epistemic utility and norms for credences, Philosophy Compass 8 (2013)
897–908.
[19] J. Dunn, Reliability for degrees of belief, Philosophical Studies 172 (2015) 1929–1952.
[20] R. Pettigrew, What is justified credence?, Episteme 18 (2021) 16–30.
[21] M. Schemmer, P. Hemmer, N. Kühl, C. Benz, G. Satzger, Should I follow AI-based
advice? Measuring appropriate reliance in human-ai decision-making, arXiv preprint
arXiv:2204.06916 (2022).
      </p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>L.</given-names>
            <surname>Floridi</surname>
          </string-name>
          ,
          <article-title>Establishing the rules for building trustworthy AI</article-title>
          ,
          <source>Nature Machine Intelligence</source>
          <volume>1</volume>
          (
          <year>2019</year>
          )
          <fpage>261</fpage>
          -
          <lpage>262</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>T.</given-names>
            <surname>Grote</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Berens</surname>
          </string-name>
          ,
          <article-title>On the ethics of algorithmic decision-making in healthcare</article-title>
          ,
          <source>Journal of Medical Ethics</source>
          <volume>46</volume>
          (
          <year>2020</year>
          )
          <fpage>205</fpage>
          -
          <lpage>211</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>A.</given-names>
            <surname>Jacovi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Marasović</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Miller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Goldberg</surname>
          </string-name>
          ,
          <article-title>Formalizing trust in artificial intelligence: Prerequisites, causes and goals of human trust in AI</article-title>
          ,
          <source>Proceedings of the 2021 ACM Conference on Fairness, Accountability</source>
          , and
          <string-name>
            <surname>Transparency</surname>
          </string-name>
          (
          <year>2021</year>
          )
          <fpage>624</fpage>
          -
          <lpage>635</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Durán</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. R.</given-names>
            <surname>Jongsma</surname>
          </string-name>
          ,
          <article-title>Who is afraid of black box algorithms? On the epistemological and ethical basis of trust in medical ai</article-title>
          ,
          <source>Journal of Medical Ethics</source>
          <volume>47</volume>
          (
          <year>2021</year>
          )
          <fpage>329</fpage>
          -
          <lpage>335</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>A.</given-names>
            <surname>Jobin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ienca</surname>
          </string-name>
          ,
          <string-name>
            <surname>E. Vayena,</surname>
          </string-name>
          <article-title>The global landscape of AI ethics guidelines</article-title>
          ,
          <source>Nature Machine Intelligence</source>
          <volume>1</volume>
          (
          <year>2019</year>
          )
          <fpage>389</fpage>
          -
          <lpage>399</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>B.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Qi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Di</surname>
          </string-name>
          , J. Liu,
          <string-name>
            <given-names>J.</given-names>
            <surname>Pei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Yi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <surname>Trustworthy</surname>
            <given-names>AI</given-names>
          </string-name>
          :
          <article-title>From principles to practices</article-title>
          ,
          <source>arXiv preprint arXiv:2110.01167</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>E.</given-names>
            <surname>Petersen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Potdevin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Mohammadi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Zidowitz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Breyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Nowotka</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Henn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Pechmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Leucker</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rostalski</surname>
          </string-name>
          , et al.,
          <article-title>Responsible and regulatory conform machine learning for medicine: A survey of challenges and solutions</article-title>
          , IEEE Access (
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>J. D.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. A.</given-names>
            <surname>See</surname>
          </string-name>
          ,
          <article-title>Trust in automation: Designing for appropriate reliance</article-title>
          ,
          <source>Human Factors</source>
          <volume>46</volume>
          (
          <year>2004</year>
          )
          <fpage>50</fpage>
          -
          <lpage>80</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>A.</given-names>
            <surname>Ferrario</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Loi</surname>
          </string-name>
          ,
          <article-title>How explainability contributes to trust in AI</article-title>
          , in: 2022 ACM Conference on Fairness, Accountability, and
          <string-name>
            <surname>Transparency</surname>
          </string-name>
          ,
          <year>2022</year>
          , pp.
          <fpage>1457</fpage>
          -
          <lpage>1466</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>M.</given-names>
            <surname>Loi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ferrario</surname>
          </string-name>
          , E. Viganò,
          <article-title>How much do you trust me? A logico-mathematical analysis of the concept of the intensity of trust</article-title>
          ,
          <source>Synthese</source>
          <volume>201</volume>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>J. M. Durán</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Formanek</surname>
          </string-name>
          ,
          <article-title>Grounds for trust: Essential epistemic opacity and computational reliabilism</article-title>
          ,
          <source>Minds and Machines</source>
          <volume>28</volume>
          (
          <year>2018</year>
          )
          <fpage>645</fpage>
          -
          <lpage>666</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>J. M. Durán</surname>
          </string-name>
          ,
          <article-title>Explaining simulated phenomena: A defense of the epistemic power of computer simulations (</article-title>
          <year>2014</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>E.</given-names>
            <surname>Schwitzgebel</surname>
          </string-name>
          , Belief, in: E. N.
          <string-name>
            <surname>Zalta</surname>
          </string-name>
          (Ed.),
          <source>The Stanford Encyclopedia of Philosophy</source>
          , Winter 2021 ed., Metaphysics Research Lab, Stanford University,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>