<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Evaluating CBR Explanation Capabilities: Survey and Next Steps</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Lawrence Gates</string-name>
          <email>gatesla@indiana.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>David Leake</string-name>
          <email>leake@indiana.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Luddy School of Informatics, Computing, and Engineering, Indiana University Bloomington IN 47408</institution>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The promise of case-based reasoning (CBR) methods for explaining intelligent systems has prompted much interest, including interest in explaining the CBR process itself. However, there is an impediment to understanding and advancing explanations of the CBR process: There has been little assessment of how well they perform. This paper develops strategies for filling this gap. We first highlight selected work on evaluating systems in explainable artificial intelligence, and then present a selective survey of work on explainable CBR systems to identify those that have been assessed with human subjects studies. From this we develop a proposed set of dimensions for characterizing explanation components of CBR systems. These dimensions provide a framework that can be used to guide evaluation, as well as a vehicle for identifying new research questions. We close with proposed next steps for research and for community initiatives.</p>
      </abstract>
      <kwd-group>
        <kwd>Explainable Case-based Reasoning</kwd>
        <kwd>Explanation</kwd>
        <kwd>Evaluation</kwd>
        <kwd>Human Subjects</kwd>
        <kwd>Survey</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Explanation of intelligent systems is a major area of AI research. Progress has
been driven both by AI research initiatives, most notably DARPA’s Explainable
Artificial Intelligence (XAI) Program [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], and by practical needs arising from
the European Union directive on the “right to explanation” for conclusions from
decision-making systems [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. The CBR community has long seen
interpretability as a key advantage of CBR (e.g., [
        <xref ref-type="bibr" rid="ref25 ref27">25, 27</xref>
        ]). There is substantial evidence for
human use of case-based reasoning [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ], and experience with CBR applications
suggests that people find prior examples a natural vehicle for explaining
decisions. Numerous approaches have been proposed for using CBR to generate
explanations for black box systems (e.g., [
        <xref ref-type="bibr" rid="ref20 ref21 ref36">20, 21, 36</xref>
        ]), including “twinning” CBR
and black box systems and using cases to explain the black box system results
(see Keane and Kenny [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ] for a survey). There are also strong arguments for
focusing directly on the construction of interpretable systems rather than
attempting to augment noninterpretable ones with explanatory capabilities [
        <xref ref-type="bibr" rid="ref44">44</xref>
        ]—
another opportunity for CBR. Workshops on explainable CBR (XCBR) [
        <xref ref-type="bibr" rid="ref2 ref6">2, 6</xref>
        ]
Copyright © 2021 for this paper by its authors. Use permitted under
Creative Commons License Attribution 4.0 International (CC BY 4.0).
have highlighted opportunities both for applying CBR to explain other systems
and for explaining the behavior of CBR systems.
      </p>
      <p>
        This paper focuses on needs for explaining CBR system behavior. Researchers
have observed that simply presenting the most similar case in a case base may not
be sufficient for optimal explanation (e.g., [
        <xref ref-type="bibr" rid="ref11 ref31 ref32 ref43">11, 31, 32, 43</xref>
        ]). A challenge for CBR
is assessing the explanations that are generated. The obvious approach—and the
only definitive method—for assessing the usefulness of explanations to end users
is human subjects studies. However, such studies require substantial resources
and few have been done for CBR explanation systems. Keane and Kenny refer
to this gap as “[t]he embarrassment of user testing.” In their extensive survey
of twinned CBR-ANN systems, they “found less than a handful (i.e., &lt; 5) that
performed any adequate user testing of the proposal that cases improved the
interpretability of models” [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ]. Bel´en D´ıaz Agudo’s ICCBR-19 keynote talk on
explanation listed human subjects studies of human-oriented explanations as one
of four key explanation challenges for the CBR community.1
      </p>
      <p>This paper aims to solidify understanding of the state of the art in assessing
explanation of CBR systems. First, it presents general background on the nature
of explanation and human subjects evaluation in XAI. Second, it presents a
survey of human-subjects evaluations of explanation capabilities in CBR, to
provide a reference for the methods and capabilities that have been assessed.
Third, it categorizes this work, to identify the space to be explored and current
gaps. Finally, based on the categorization and observations of prior research, it
proposes five steps—both short and longer term—for advancing the evaluation
of explanation of CBR systems and their explanation capabilities.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Aims of Explanation</title>
      <p>
        Numerous studies have examined what makes a good explanation and how to
explain in XCBR. Kenny et al. [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ] propose a “tricorn” user model for XAI,
with explanations relating to a user’s domain model, model of the system, and
model of how particular explanation strategies (e.g., graphical presentation of
information) relate evidence to a conclusion. Schoenborn et al. [
        <xref ref-type="bibr" rid="ref45">45</xref>
        ] survey
explainable CBR and Sormo, Cassens and Aamodt [
        <xref ref-type="bibr" rid="ref47">47</xref>
        ] survey explanation in CBR
with an emphasis on criteria for good explanations. In general, explanations can
serve a rich range of needs, depending on the user’s knowledge state, goals, and
tasks [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ]. Here we focus on two fundamental goals. One is transparency,
understanding how the system produced its answer. This may be useful both to foster
trust (e.g., [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]) and to diagnose problems to guide system repair or refinement
of system knowledge or processes. The other is justification, understanding why
the system’s answer is a good answer [
        <xref ref-type="bibr" rid="ref47">47</xref>
        ].
1 https://iccbr2019.com/wp-content/uploads/2019/09/Keynote-BelenDiazAgudo.pdf
      </p>
    </sec>
    <sec id="sec-3">
      <title>Assessing Explanation Systems</title>
      <p>
        Given the goal of serving a human user, the gold standard for system
assessments must be the results of human subjects studies. Human subjects studies
can be designed directly to assess and compare explanation quality of
alternative systems (e.g., [
        <xref ref-type="bibr" rid="ref10 ref30 ref37 ref38">10, 30, 37, 38</xref>
        ]). However, such assessments are challenging to
design and may be time-consuming and expensive, especially given that many
aspects of system processing (e.g., situation assessment, similarity assessment,
or adaptation) might need to be explained. As a result, few XCBR efforts use
human subject evaluations and the space of possible factors affecting
explanation effectiveness is sparsely covered. Some studies rely on informal evaluations,
seeking reactions of a small set of subjects, and many system designs simply rely
on the system-builder’s intuitions. A goal of this paper is to provide a foundation
for choosing aspects of a system to evaluate and to help identify the aspects that
are most and least understood, to illuminate areas for future study.
4
      </p>
    </sec>
    <sec id="sec-4">
      <title>Categorizing Explanations of CBR Systems</title>
      <p>
        This section presents a categorization of research on explaining CBR systems.
We begin by introducing the dimensions used and then place research
according to those dimensions. The dimensions primarily consider the aspects of the
CBR process being explained and the knowledge (derived from the CBR
knowledge containers) that is the focus of those explanations. Of particular interest is
identifying systems that have been evaluated with human subjects studies and
opportunities for future human subjects studies. Schoenborn et al. [
        <xref ref-type="bibr" rid="ref45">45</xref>
        ] present
a complementary survey, categorizing XCBR systems according to four types
of aspects: definitional (explanation goal and kinds such as why-explanations or
how-explanations), model-based or model-agnostic explanation, and
(overlapping with this survey) the medium of presentation. We note that XCBR is not
limited to explanation of CBR systems, including, for example, much interesting
work on CBR to support explanation in non-CBR systems (such as [
        <xref ref-type="bibr" rid="ref20 ref41">20, 41</xref>
        ]).
However, that work is outside the scope of this paper.
      </p>
      <p>
        Explanations of the conclusions of a CBR system may focus on the solution
provided by the system or the results of any of the intermediate steps in the
CBR cycle—retrieve, reuse, revise, or retain [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]—with an account of how those
results contribute to the solution. For example, CBR system behavior could be
explained by the retrieved case, or a retrieved case plus how it was adapted. In
principle, the retrieval could be explained both by similarity criteria and by how
previous retention decisions affected the choice of cases to retain.
      </p>
      <p>
        For any of the steps, the generation of a particular result depends on the
knowledge brought to bear. Knowledge from any of the CBR knowledge
containers [
        <xref ref-type="bibr" rid="ref42">42</xref>
        ] could be relevant to an explanation. In addition, given the well known
tight coupling of knowledge containers, there may be a choice in determining
which type of knowledge to highlight for a user (e.g., indexing or similarity
knowledge accounting for retrieval of a case, if case relevance may not be
obvious to the user, or case adaptation knowledge, if case relevance is intuitive but
the adapted solution may not be). Each of these choices provides a dimension
for categorizing explanations of CBR systems.
4.1
      </p>
      <sec id="sec-4-1">
        <title>Categorization Table</title>
        <p>The table in Figure 1 categorizes the explanation approaches taken by sixteen
projects. Research evaluated with human subjects studies is highlighted in bold.
Rows are divided into four sections. The first section categorizes the research by
Explanation Medium, the presentation of the explanation. The medium may be
in graphical/visual or textual (explanations in internal system symbolic
representations are considered to be in textual form). The second section categorizes
the research by the Target of Explanation. Targets of the explanations may be
the system’s non-adapted final result, the reason for the retrieval of the case
used, the result when the retrieved case was adapted, or aspects of the ongoing
system process.</p>
        <p>The third section of rows categorizes the research by the primary Knowledge
Focus of the explanation. A holistic focus provides an entire retrieved case to
the user; the user determines how the case relates to explanatory needs. A key
features focus provides information about the case features salient for retrieval.
An index knowledge focus explains retrievals in terms of memory organization,
the knowledge used to generate indices from cases, and/or knowledge underlying
how those indices may be revised or transformed during the retrieval process.
An adaptation knowledge focus explains the generation of a solution in terms of
the application of knowledge such as adaptation rules, adaptation cases, or other
knowledge from the adaptation knowledge container. A similarity focus explains
retrieval in terms of the similarity measure and/or analysis of case features in
terms of similarity criteria. In principle, a system explaining CBR system
behavior could have multiple knowledge focuses. However, in practice, most research
has focused on single focus types. The final two sections of rows categorize the
research by Explanation Recipient and the Process or Result Focus. The recipient
of the explanation may be a human or an AI system (as in systems that explain
internally to guide processing). The focus of explanation is the final the system
process used to obtain it (for transparency) or system result (for justification).
4.2</p>
      </sec>
      <sec id="sec-4-2">
        <title>The Systems Categorized</title>
        <p>
          Armengol, Ontan´on and Plaza [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] present a method for explaining similarity
to justify solutions in classification tasks. The approach was applied to provide
justification to other agents in a multi-agent system. This system focuses on
explaining the similarity of the non-adapted final result in text (symbolic) format,
for AI systems.
        </p>
        <p>
          Burke and Kass [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] present a system that explains the retrieval result for
video clips with “bridging text” accounting for the relevant indices, focusing on
index knowledge, for presentation to a human user.
        </p>
        <p>
          Cunningham, Doyle and Loughrey [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ] performed, to our knowledge, the
first human subjects study on prior cases as explanations. Their study, using
        </p>
        <sec id="sec-4-2-1">
          <title>Explanation</title>
        </sec>
        <sec id="sec-4-2-2">
          <title>Medium</title>
        </sec>
        <sec id="sec-4-2-3">
          <title>Target for</title>
        </sec>
        <sec id="sec-4-2-4">
          <title>Explanation</title>
        </sec>
        <sec id="sec-4-2-5">
          <title>Knowledge</title>
        </sec>
        <sec id="sec-4-2-6">
          <title>Focus</title>
        </sec>
        <sec id="sec-4-2-7">
          <title>Explanation</title>
        </sec>
        <sec id="sec-4-2-8">
          <title>Recipient</title>
        </sec>
        <sec id="sec-4-2-9">
          <title>Process or</title>
        </sec>
        <sec id="sec-4-2-10">
          <title>Result Focus</title>
        </sec>
        <sec id="sec-4-2-11">
          <title>Visualization</title>
        </sec>
        <sec id="sec-4-2-12">
          <title>Text</title>
        </sec>
        <sec id="sec-4-2-13">
          <title>Non-Adapted</title>
        </sec>
        <sec id="sec-4-2-14">
          <title>Final Result</title>
        </sec>
        <sec id="sec-4-2-15">
          <title>Retrieval</title>
        </sec>
        <sec id="sec-4-2-16">
          <title>Result</title>
        </sec>
        <sec id="sec-4-2-17">
          <title>Adaptation</title>
        </sec>
        <sec id="sec-4-2-18">
          <title>Result</title>
        </sec>
        <sec id="sec-4-2-19">
          <title>Ongoing</title>
        </sec>
        <sec id="sec-4-2-20">
          <title>Process</title>
        </sec>
        <sec id="sec-4-2-21">
          <title>Case Only</title>
          <p>(Holistic)</p>
        </sec>
        <sec id="sec-4-2-22">
          <title>Key Features</title>
        </sec>
        <sec id="sec-4-2-23">
          <title>Index</title>
        </sec>
        <sec id="sec-4-2-24">
          <title>Knowledge</title>
        </sec>
        <sec id="sec-4-2-25">
          <title>Adaptation</title>
        </sec>
        <sec id="sec-4-2-26">
          <title>Knowledge</title>
        </sec>
        <sec id="sec-4-2-27">
          <title>Similarity</title>
        </sec>
        <sec id="sec-4-2-28">
          <title>Human User</title>
        </sec>
        <sec id="sec-4-2-29">
          <title>AI System</title>
        </sec>
        <sec id="sec-4-2-30">
          <title>System Process</title>
        </sec>
        <sec id="sec-4-2-31">
          <title>System Result</title>
          <p>C
ll,,trzgaaae04PnnnnóoAO&amp;m ,rsskae96KuB&amp; li,,rLyyggaee03hunnuhnooD&amp;m lil,,sygaae06ChnnnuhoD&amp;mW ,ssa90K l,re83Kdnoo l,tLyaae19m ,Lkaee91 il,trLaeee08b ii,,trrssgaaaae04Cun&amp;wMW iii,,frcxSFaaaae04nhnß&amp;mmM ,rrcySe04hM ,rrcySe01hM l,.taaae15hdummM li,,tyggaee05CnuhnnnuoDN&amp;m l,.tgae97nO</p>
          <p> 
     
 </p>
          <p>     

 
 
 </p>
          <p>  






 
 






   
     </p>
          <p>
       

 
 
 
 </p>
          <p>  
</p>
          <p>
  



37 subjects, compared the convincingness of no explanation, presentation of the
case on which a CBR system’s conclusion was based, and rule-based explanation
for predictions of blood alcohol levels. The system’s target is the non-adapted
final result in text format, with holistic knowledge focus, to a human user.</p>
          <p>
            Doyle, Cunningham and Walsh [
            <xref ref-type="bibr" rid="ref12">12</xref>
            ] study explanation via a selected prior
case (not necessarily the nearest case), justification text and a confidence level,
in the domain of hospital admission and discharge decisions. The justification
text highlighted features that support or undermine relevance of the explanation
case. The study had two evaluation stages, a formative evaluation and a second
stage of two studies, one with 106 domain experts judging confidence in the
system and the appropriateness of the explanation case, and the second with 14
domain experts judging whether the nearest neighbor or explanation case better
explained the system decision. The system’s target is the non-adapted final result
in text format, with holistic knowledge focus, to a human user.
          </p>
          <p>
            Kass [
            <xref ref-type="bibr" rid="ref19">19</xref>
            ] presents a case adaptation system that provides textual descriptions
of adaptation paths that could be presented to explain adaptations to a user.
The system’s target is the adaptation result in text format, with knowledge focus
of adaptation knowledge, for a human user.
          </p>
          <p>
            Kolodner [
            <xref ref-type="bibr" rid="ref22">22</xref>
            ] presents explanation of retrieval by a question transformation
trace leading to the retrieval result in a text format, with the knowledge focus
index knowledge to a human user.
          </p>
          <p>
            Lamy et al. [
            <xref ref-type="bibr" rid="ref23">23</xref>
            ] focus on visualization of differences and similarities between
a query case and similar cases, in the context of a CBR system to diagnose breast
cancer. Their system was evaluated by 11 users who selected a treatment based
on the data presented by the system and indicated their confidence level in the
choice. The target is the non-adapted final result in the form of a visualization,
with a holistic knowledge focus, to a human user.
          </p>
          <p>
            Leake [
            <xref ref-type="bibr" rid="ref24">24</xref>
            ] presents a set of criteria for evaluating CBR-generated
explanations of anomalous events in stories for various goals (e.g., prediction of an event
or prevention of an event). The system’s target is the non-adapted final result in
text format, with a knowledge focus of holistic, for an AI system.
          </p>
          <p>
            Lieber et al. [
            <xref ref-type="bibr" rid="ref29">29</xref>
            ] present an approach to explaining adaptations, in the
context of breast cancer treatment, by presenting adaptation paths with annotations
of each step. The system’s target is the adaptation result in text format, with
knowledge focus on adaptation knowledge, to a human user.
          </p>
          <p>
            Massie, Craw and Wiratunga [
            <xref ref-type="bibr" rid="ref30">30</xref>
            ] present a visualization tool aimed at
presenting feature similarity between a new problem and retrieved case, in the
domain of drug design. The work is assessed by asking two domain experts to
assess similarity and express their confidence in the answer the solution provided.
The target for explanation is the retrieval result in the form of a visualization,
with a holistic knowledge focus and explanation to a human user.
          </p>
          <p>
            Maximini, Freßman, and Schaaf [
            <xref ref-type="bibr" rid="ref31">31</xref>
            ] present a system including a range of
explanation capabilities for an intellectual property domain. The system includes
two explanation components. The first explains the set of cases retrieved in
terms of attributes and supports user navigation through the candidates, as well
as presenting the most decisive attributes to narrow the retrieval set. When
no candidates are retrieved, it can also help the user understand which query
features could not be matched, to aid query refinement. The other supports
examination of the result, with color coding of features by similarity as well as
a graphical visualization and textual explanation of feature connections. The
system’s target is the non-adapted final result in text format, with a knowledge
focus of the similarity for a human user.
          </p>
          <p>
            McSherry [
            <xref ref-type="bibr" rid="ref32">32</xref>
            ] presents an explanation method that provides the features of
the nearest cases that support or oppose a prediction. This system focuses on
explaining the non-adapted final result in form of case text, with knowledge focus
on key features to a human user.
          </p>
          <p>
            McSherry [
            <xref ref-type="bibr" rid="ref33">33</xref>
            ] presents an explanation method for conversational CBR for
diagnosis of computer faults. The system provides explanations of the relevance
of questions. This system focuses on explaining the ongoing process in the form
of text, with knowledge focus on index generation, to a human user.
          </p>
          <p>
            Muhammad et al. [
            <xref ref-type="bibr" rid="ref34">34</xref>
            ] present an approach to explaining recommendations
based on positive and negative features compared to user preferences, in a hotel
recommendation domain. The recommender personalizes explanations by
highlighting the features the given user seeks. This system focuses on explaining the
non-adapted final result in form of text, with a knowledge focus on key features
to a human user.
          </p>
          <p>
            Nugent, Cunningham and Doyle [
            <xref ref-type="bibr" rid="ref37">37</xref>
            ] extend their original study [
            <xref ref-type="bibr" rid="ref10">10</xref>
            ] for an
explanation component providing the retrieved case, text explaining the difference
between the query and explanation case, and a confidence value. If confidence is
below a threshold, an additional counter case is provided, with a description of
how feature values in both cases impact classification. The study used 12 human
subjects. The system target is the non-adapted final result in text format, with
a holistic knowledge focus and explanation for a human user.
          </p>
          <p>
            Ong et al. [
            <xref ref-type="bibr" rid="ref39">39</xref>
            ] present a system that explains similarity by presenting
similarity scores for matching features, in a cancer diagnosis domain. This system
focuses on explaining the non-adapted final result in form of text, with knowledge
focus on the case as a whole, so holistic, to a human user.
5
          </p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Next Steps</title>
      <p>Examination of research in the survey suggests five steps for advancing
evaluation of explanation of CBR systems. The first, building on the dimensions of the
categorization, is to fill gaps in coverage of evaluation dimensions in the
categorizations. The second and third involve community effort to make evaluations
more practical to perform and results more comparable. This requires both
advancing automated evaluations, to enable gathering initial evaluation data more
rapidly, and establishing standard datasets and standardized evaluation designs,
in order to make studies more useful for understanding the relationships between
different explanation focuses and how they contribute to the benefit of
explanation. The fourth and fifth are to move beyond the current space of explanations,
focused on local explanation of a single solution, to build explanations to help
users better understand how systems will perform in the future and to explain
system failures. Understanding the full scope of explanation of CBR systems will
require experiments that assess the full system, including effects of component
interactions.
5.1</p>
      <sec id="sec-5-1">
        <title>Using the Categorization Scheme to Suggest Evaluation Targets</title>
        <p>
          The categorization table of the previous section shows the distribution of
human subjects studies for explaining CBR. We observe that existing evaluations
have primarily focused on case-oriented explanations—explaining by presenting
cases and highlighted features to the user. This has been important for assessing
the fundamental CBR premise that cases are a natural mode of explanation.
However, in domains unfamiliar to the user, explanatory support for assessing
similarity may be crucial. This received some recognition [
          <xref ref-type="bibr" rid="ref30 ref4">4, 30</xref>
          ], but so far, to
our knowledge, only one method has been tested with human subjects, and in
a small-scale assessment (by two experts) [
          <xref ref-type="bibr" rid="ref30">30</xref>
          ]. Adaptation-focused explanation
has also received little attention.
        </p>
        <p>
          The survey also suggests the potential for studies of how different modes of
information presentation affect efficacy of explanations. Systematic evaluation
will require understanding:
– Alternative methods and tradeoffs for conveying to the user information
about system knowledge in each of the CBR knowledge containers [
          <xref ref-type="bibr" rid="ref42">42</xref>
          ]: the
vocabulary, similarity measure, case base, and adaptation knowledge.
– How best to explain the processes underlying each step of the CBR cycle:
retrieve, reuse, revise, and retain [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ].
– The influence of information needs [
          <xref ref-type="bibr" rid="ref24 ref47">24, 47</xref>
          ]: how to select parts of that process
to highlight to particular users.
        </p>
        <p>
          The ideal result would be a toolkit of best practices for explaining classes of CBR
systems based on the tasks they perform. For example, explaining a “compare
and contrast” task, as in American legal reasoning [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ], will necessarily differ from
explaining anomalies [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ]. Expert systems research on generic tasks [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] suggests
the potential to identify a comparatively limited set of task classes. This in
turn would increase the practicality and usefulness of human subjects studies by
encouraging studies with wide task applicability. A community effort to develop
such a collection could leverage both XCBR research and the application of
XCBR in domains for which explanation is important.
5.2
        </p>
      </sec>
      <sec id="sec-5-2">
        <title>Advancing Automatic Evaluation</title>
        <p>
          There is a longstanding tradition of testing conversational CBR systems with
simulated users (e.g., [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ]). Research on case-based explanation generation has
examined criteria for evaluating the goodness of explanations generated [
          <xref ref-type="bibr" rid="ref18 ref24 ref46">18, 24, 46</xref>
          ].
This prior work raises an interesting question: Would it be possible to develop
automated evaluation schemes to test aspects of the explanations generated by
CBR systems? Even if such methods did not have complete fidelity to human
subjects studies, automatic testing could be a valuable intermediate step.
        </p>
        <p>
          Work on case-based explanation for story understanding systems has
developed criteria for evaluating explanations according to their ability to resolve
anomalies in stories, address system questions, and satisfy other explanation
goals [
          <xref ref-type="bibr" rid="ref24 ref40">24, 40</xref>
          ]. Applying such methods to a broader class of tasks would require
analysis of the information needs associated with those tasks and how they can
be satisfied. Providing a general solution for this would be an enormous task.
However, it could be more feasible in testbed microdomains—were the
community to embrace developing such resources.
5.3
        </p>
      </sec>
      <sec id="sec-5-3">
        <title>Developing Standardized Evaluation Resources</title>
        <p>
          The ability to compare and replicate human subjects studies is impeded by the
lack of agreed-upon testbeds for such evaluations. Different research groups use
different data sets, making it more difficult to assess the relative contributions
of explanations of different parts of the CBR process and how those components
interact. From the five human subject studies reported in this paper’s
categorization, two (done by the same research group) used the blood-alcohol domain [
          <xref ref-type="bibr" rid="ref10 ref37">10,
37</xref>
          ], one used the tablet formulation domain [
          <xref ref-type="bibr" rid="ref30">30</xref>
          ], one used Bronchiolitis medical
data [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ], and one used breast cancer data [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ]. Of the 4 unique domains, only
the breast cancer data [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ] is publicly available. None of these data sets involve
CBR systems including case adaptation.
        </p>
        <p>
          The development of the needed data sets—and agreement to use them—must
be a community endeavor. The CBR community has a history of developing
shared resources and advancing the state of the art with competitions, as in
the Computer Cooking Competition series (e.g., [
          <xref ref-type="bibr" rid="ref35">35</xref>
          ]). This series included
various forms of human evaluations, both by human judges and by the conference
attendees as a whole. An explanation challenge, based on standard tasks and
data sets and informally evaluated by humans as part of the competition, could
“jump start” the development of standard resources. In addition, the design of
the competition could help establish evaluation protocols to form the basis for
more formal human subjects studies.
5.4
        </p>
      </sec>
      <sec id="sec-5-4">
        <title>Explaining the prior cases themselves</title>
        <p>
          Explanations based on presenting prior cases assume the quality of the retrieved
case can be trusted—that no justification is needed for why the case should
be believed. However, especially if cases have been automatically generated by
adaptation processes, understanding and accepting a case as an explanation
depends on understanding the presented case and its relevance. That may depend
on understanding how it was derived—its provenance [
          <xref ref-type="bibr" rid="ref28">28</xref>
          ]. The use of provenance
for explaining CBR is an interesting area for future research.
5.5
        </p>
      </sec>
      <sec id="sec-5-5">
        <title>Broadening CBR Explanations to Explaining Failures</title>
        <p>Existing studies of explaining CBR focus primarily on accounting for why the
system reasoned as it did. Users may use that information to determine the
applicability of a proposed solution. This is sufficient to determine trust in the
current solution. However, when users determine that the system went astray, the
explanation components generally have no capability to explain why the CBR
process failed. In the absence of such an explanation it is harder to determine
trust in the system’s future processing.</p>
        <p>
          Some CBR research has applied metareasoning to explain and repair
reasoning failures within the CBR process for internal system learning (e.g., [
          <xref ref-type="bibr" rid="ref15 ref9">9, 15</xref>
          ]).
This points to another topic for CBR system explanation: accounting for
expectation failures in how the system operates, to help users refine their predictive
model of the CBR system overall.
        </p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Conclusion</title>
      <p>This paper has presented an initial categorization of research on explaining CBR
systems, highlighting dimensions for which human subjects evaluations have
been done—and for which they are still needed. Based on this categorization
it has proposed XCBR research opportunities including increased attention to
explaining similarity and adaptation. It has also advocated new initiatives on
automatic evaluation, building community evaluation resources, explaining case
provenance, and explaining system failures.
7</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgments</title>
      <p>We acknowledge support from the Department of the Navy, Office of Naval
Research (Award N00014-19-1-2655), and the US Department of Defense (Contract
W52P1J2093009).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Aamodt</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Plaza</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          :
          <article-title>Case-based reasoning: Foundational issues, methodological variations, and system approaches</article-title>
          .
          <source>AI Communications</source>
          <volume>7</volume>
          (
          <issue>1</issue>
          ),
          <fpage>39</fpage>
          -
          <lpage>52</lpage>
          (
          <year>1994</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Aha</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Agudo</surname>
            ,
            <given-names>B.D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Garcia</surname>
            ,
            <given-names>J.R</given-names>
          </string-name>
          . (eds.): Proceedings of XCBR-2018:
          <article-title>First workshop on case-based reasoning for the explanation of intelligent systems</article-title>
          .
          <source>ICCBR</source>
          (
          <year>2018</year>
          ), http://iccbr18.com/wp-content/uploads/ICCBR-2018
          <article-title>-V3</article-title>
          .pdf
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Aha</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Breslow</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Refining conversational case libraries</article-title>
          .
          <source>In: Proceedings of the Second International Conference on Case-Based Reasoning</source>
          . pp.
          <fpage>267</fpage>
          -
          <lpage>278</lpage>
          . Springer Verlag, Berlin (
          <year>1997</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Armengol</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          , Ontan´on,
          <string-name>
            <given-names>S.</given-names>
            ,
            <surname>Plaza</surname>
          </string-name>
          , E.:
          <article-title>Explaining similarity in CBR</article-title>
          .
          <source>In: ECCBR 2004 Workshop Proceedings</source>
          . pp.
          <fpage>155</fpage>
          -
          <lpage>164</lpage>
          (
          <year>2004</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Ashley</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rissland</surname>
          </string-name>
          , E.:
          <article-title>Compare and contrast, a test of expertise</article-title>
          .
          <source>In: Proceedings of the Sixth Annual National Conference on Artificial Intelligence</source>
          . pp.
          <fpage>273</fpage>
          -
          <lpage>284</lpage>
          . AAAI, Morgan Kaufmann, San Mateo, CA (
          <year>1987</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <given-names>B.</given-names>
            <surname>Diaz Agudo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.R.G.</given-names>
            ,
            <surname>Watson</surname>
          </string-name>
          , I. (eds.): Proceedings of XCBR-2019:
          <article-title>Second workshop on case-based reasoning for the explanation of intelligent systems</article-title>
          .
          <source>ICCBR</source>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Burke</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kass</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Retrieving stories for case-based teaching</article-title>
          . In: Leake,
          <string-name>
            <surname>D</surname>
          </string-name>
          . (ed.)
          <source>Case-Based Reasoning: Experiences</source>
          , Lessons, and Future Directions, pp.
          <fpage>93</fpage>
          -
          <lpage>109</lpage>
          . AAAI Press, Menlo Park, CA (
          <year>1996</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Chandrasekaran</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Generic tasks in knowledge-based reasoning: High level building blocks for expert system design</article-title>
          .
          <source>IEEE Expert</source>
          <volume>1</volume>
          (
          <issue>3</issue>
          ),
          <fpage>23</fpage>
          -
          <lpage>30</lpage>
          (
          <year>1986</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Cox</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ram</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Introspective multistrategy learning: On the construction of learning strategies</article-title>
          .
          <source>Artificial Intelligence</source>
          <volume>112</volume>
          (
          <issue>1-2</issue>
          ),
          <fpage>1</fpage>
          -
          <lpage>55</lpage>
          (
          <year>1999</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Cunningham</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Doyle</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Loughrey</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>An evaluation of the usefulness of casebased explanation</article-title>
          .
          <source>In: Case-Based Reasoning Research and Development: Proceedings of the Fifth International Conference on Case-Based Reasoning, ICCBR-03</source>
          . pp.
          <fpage>122</fpage>
          -
          <lpage>130</lpage>
          . Springer-Verlag, Berlin (
          <year>2003</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Doyle</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cunningham</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bridge</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rahman</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          :
          <article-title>Explanation oriented retrieval</article-title>
          . In: Funk,
          <string-name>
            <surname>P.</surname>
          </string-name>
          , Gonz´alez Calero, P.A. (eds.)
          <source>Advances in Case-Based Reasoning</source>
          . pp.
          <fpage>157</fpage>
          -
          <lpage>168</lpage>
          . Springer, Berlin, Heidelberg (
          <year>2004</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Doyle</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cunningham</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Walsh</surname>
            ,
            <given-names>P.:</given-names>
          </string-name>
          <article-title>An evaluation of the usefulness of explanation in a case-based reasoning system for decision support in bronchiolitis treatment</article-title>
          .
          <source>Computational Intelligence</source>
          <volume>22</volume>
          (
          <issue>3-4</issue>
          ),
          <fpage>269</fpage>
          -
          <lpage>281</lpage>
          (
          <year>2006</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Dua</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Graff</surname>
            ,
            <given-names>C.:</given-names>
          </string-name>
          <article-title>UCI machine learning repository (</article-title>
          <year>2017</year>
          ), http://archive.ics.uci.edu/ml
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Floyd</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Aha</surname>
          </string-name>
          , D.:
          <article-title>Special issue on case-based reasoning</article-title>
          .
          <source>AI</source>
          Communications
          <volume>30</volume>
          (
          <issue>3-4</issue>
          ),
          <fpage>281</fpage>
          -
          <lpage>294</lpage>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Fox</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Leake</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Introspective reasoning for index refinement in case-based reasoning</article-title>
          .
          <source>The Journal of Experimental and Theoretical Artificial Intelligence</source>
          <volume>13</volume>
          (
          <issue>1</issue>
          ),
          <fpage>63</fpage>
          -
          <lpage>88</lpage>
          (
          <year>2001</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Goodman</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Flaxman</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>European Union regulations on algorithmic decisionmaking and a “right to explanation”</article-title>
          .
          <source>AI</source>
          Magazine
          <volume>38</volume>
          (
          <issue>3</issue>
          ),
          <fpage>50</fpage>
          -
          <lpage>57</lpage>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Gunning</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Aha</surname>
            ,
            <given-names>D.W.:</given-names>
          </string-name>
          <article-title>DARPA's explainable artificial intelligence program</article-title>
          .
          <source>AI</source>
          Magazine
          <volume>40</volume>
          (
          <issue>2</issue>
          ),
          <fpage>44</fpage>
          -
          <lpage>58</lpage>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Johs</surname>
            ,
            <given-names>A.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lutts</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Weber</surname>
            ,
            <given-names>R.O.</given-names>
          </string-name>
          :
          <article-title>Measuring explanation quality in XCBR</article-title>
          . In: Minor, M. (ed.)
          <source>Workshop proceedings ICCBR 2018</source>
          . pp.
          <fpage>75</fpage>
          -
          <lpage>83</lpage>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Kass</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Developing Creative Hypotheses by Adapting Explanations</article-title>
          .
          <source>Ph.D. thesis</source>
          , Yale University (
          <year>1990</year>
          ), northwestern University Institute for the
          <source>Learning Sciences, Technical Report 6</source>
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Keane</surname>
            ,
            <given-names>M.T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kenny</surname>
            ,
            <given-names>E.M.:</given-names>
          </string-name>
          <article-title>How case-based reasoning explains neural networks: A theoretical analysis of XAI using post-hoc explanation-by-example from a survey of ANN-CBR twin-systems</article-title>
          .
          <source>In: Case-Based Reasoning Research and Development: ICCBR-19</source>
          . pp.
          <fpage>155</fpage>
          -
          <lpage>171</lpage>
          . Springer, Berlin (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>Kenny</surname>
            ,
            <given-names>E.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ford</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Quinn</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Keane</surname>
          </string-name>
          , M.T.:
          <article-title>Explaining black-box classifiers using post-hoc explanations-by-example: The effect of explanations and error-rates in XAI user studies</article-title>
          .
          <source>Artificial Intelligence</source>
          <volume>294</volume>
          ,
          <issue>103459</issue>
          (
          <year>2021</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22. Kolodner, J.L.:
          <article-title>Reconstructive memory: A computer model</article-title>
          .
          <source>Cognitive Science</source>
          <volume>7</volume>
          (
          <issue>4</issue>
          ),
          <fpage>281</fpage>
          -
          <lpage>328</lpage>
          (
          <year>1983</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <surname>Lamy</surname>
            ,
            <given-names>J.B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sekar</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Guezennec</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bouaud</surname>
          </string-name>
          , J., S´eroussi, B.:
          <article-title>Explainable artificial intelligence for breast cancer: A visual case-based reasoning approach</article-title>
          .
          <source>Artificial intelligence in medicine 94</source>
          ,
          <fpage>42</fpage>
          -
          <lpage>53</lpage>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          24.
          <string-name>
            <surname>Leake</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Goal-based explanation evaluation</article-title>
          .
          <source>Cognitive Science</source>
          <volume>15</volume>
          (
          <issue>4</issue>
          ),
          <fpage>509</fpage>
          -
          <lpage>545</lpage>
          (
          <year>1991</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          25.
          <string-name>
            <surname>Leake</surname>
            ,
            <given-names>D.:</given-names>
          </string-name>
          <article-title>CBR in context: The present and future</article-title>
          . In: Leake,
          <string-name>
            <surname>D</surname>
          </string-name>
          . (ed.)
          <source>Case-Based Reasoning: Experiences</source>
          , Lessons, and Future Directions, pp.
          <fpage>3</fpage>
          -
          <lpage>30</lpage>
          . AAAI Press, Menlo Park, CA (
          <year>1996</year>
          ), http://www.cs.indiana.edu/˜leake/papers/a-96-01.html
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          26.
          <string-name>
            <surname>Leake</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Cognition as case-based reasoning</article-title>
          . In: Bechtel,
          <string-name>
            <given-names>W.</given-names>
            ,
            <surname>Graham</surname>
          </string-name>
          ,
          <string-name>
            <surname>G</surname>
          </string-name>
          . (eds.) A Companion to Cognitive Science, pp.
          <fpage>465</fpage>
          -
          <lpage>476</lpage>
          . Blackwell, Oxford (
          <year>1998</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          27.
          <string-name>
            <surname>Leake</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>McSherry</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Introduction to the special issue on explanation in casebased reasoning</article-title>
          .
          <source>Artificial Intelligence Review</source>
          <volume>24</volume>
          (
          <issue>2</issue>
          ),
          <fpage>103</fpage>
          -
          <lpage>108</lpage>
          (
          <year>2005</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          28.
          <string-name>
            <surname>Leake</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Whitehead</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Case provenance: The value of remembering case sources</article-title>
          .
          <source>In: Case-Based Reasoning Research and Development: ICCBR-07</source>
          . pp.
          <fpage>194</fpage>
          -
          <lpage>208</lpage>
          . Springer-Verlag, Berlin (
          <year>2007</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          29.
          <string-name>
            <surname>Lieber</surname>
          </string-name>
          , J.,
          <string-name>
            <surname>d'Aquin</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Badra</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Napoli</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Modeling adaptation of breast cancer treatment decision protocols in the KASIMIR project</article-title>
          .
          <source>Applied Intelligence</source>
          <volume>28</volume>
          (
          <issue>3</issue>
          ),
          <fpage>261</fpage>
          -
          <lpage>274</lpage>
          (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          30.
          <string-name>
            <surname>Massie</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Craw</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wiratunga</surname>
          </string-name>
          , N.:
          <article-title>A visualisation tool to explain case-base reasoning solutions for tablet formulation</article-title>
          .
          <source>In: Proceedings of the 24th SGAI International Conference on Innovative Techniques and Applications of Artificial Intelligence</source>
          . Springer-Verlag, Berlin (
          <year>2004</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          31.
          <string-name>
            <surname>Maximini</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Freßmann</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Schaaf</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Explanation service for complex CBR applications</article-title>
          .
          <source>In: Advances in Case-Based Reasoning</source>
          , ECCBR. pp.
          <fpage>302</fpage>
          -
          <lpage>316</lpage>
          . Springer, Berlin (
          <year>2004</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          32.
          <string-name>
            <surname>McSherry</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Explaining the pros and cons of conclusions in CBR</article-title>
          .
          <source>In: Advances in Case-Based Reasoning</source>
          . pp.
          <fpage>317</fpage>
          -
          <lpage>330</lpage>
          . Springer, Berlin (
          <year>2004</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          33.
          <string-name>
            <surname>McSherry</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Interactive case-based reasoning in sequential diagnosis</article-title>
          .
          <source>Applied Intelligence</source>
          <volume>14</volume>
          (
          <issue>1</issue>
          ),
          <fpage>65</fpage>
          -
          <lpage>76</lpage>
          (
          <year>2001</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          34.
          <string-name>
            <surname>Muhammad</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lawlor</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rafter</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Smyth</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Great explanations: Opinionated explanations for recommendations</article-title>
          .
          <source>In: Case-Based Reasoning Research and Development</source>
          . pp.
          <fpage>244</fpage>
          -
          <lpage>258</lpage>
          . Springer, Berlin (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          35.
          <string-name>
            <surname>Najjar</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          , Wilson, D.: Computer cooking contest workshop at the twentyfifth
          <source>international conference on case-based reasoning (ICCBR</source>
          <year>2017</year>
          ). In: ICCBR 2017 Workshops, Doctoral Consortium, and
          <string-name>
            <surname>Competitions</surname>
          </string-name>
          (
          <year>2017</year>
          ), http://ceurws.org/Vol-2028/XXCCC17 preface.pdf
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          36.
          <string-name>
            <surname>Nugent</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cunningham</surname>
            ,
            <given-names>P.:</given-names>
          </string-name>
          <article-title>A case-based recommender for black-box systems</article-title>
          .
          <source>Artificial Intelligence Review</source>
          <volume>24</volume>
          (
          <issue>2</issue>
          ),
          <fpage>163</fpage>
          -
          <lpage>178</lpage>
          (
          <year>2005</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          37.
          <string-name>
            <surname>Nugent</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cunningham</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Doyle</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>The best way to instil confidence is by being right</article-title>
          .
          <source>In: International Conference on Case-Based Reasoning</source>
          . pp.
          <fpage>368</fpage>
          -
          <lpage>381</lpage>
          . Springer (
          <year>2005</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          38.
          <string-name>
            <surname>Nugent</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Doyle</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cunningham</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Gaining insight through case-based explanation</article-title>
          .
          <source>Journal of Intelligent Information Systems</source>
          <volume>32</volume>
          (
          <issue>3</issue>
          ),
          <fpage>267</fpage>
          -
          <lpage>295</lpage>
          (
          <year>Jun 2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          39.
          <string-name>
            <surname>Ong</surname>
            ,
            <given-names>L.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shepherd</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tong</surname>
            ,
            <given-names>L.C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Seow-Choen</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ho</surname>
            ,
            <given-names>Y.H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tang</surname>
            ,
            <given-names>C.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ho</surname>
            ,
            <given-names>Y.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tan</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>The colorectal cancer recurrence support (CARES) system</article-title>
          .
          <source>Artificial Intelligence in Medicine</source>
          <volume>11</volume>
          (
          <issue>3</issue>
          ),
          <fpage>175</fpage>
          -
          <lpage>188</lpage>
          (
          <year>1997</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref40">
        <mixed-citation>
          40.
          <string-name>
            <surname>Ram</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>AQUA: Asking questions and understanding answers</article-title>
          .
          <source>In: Proceedings of the Sixth Annual National Conference on Artificial Intelligence</source>
          . pp.
          <fpage>312</fpage>
          -
          <lpage>316</lpage>
          . Morgan Kaufmann, Seattle, WA (
          <year>July 1987</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref41">
        <mixed-citation>
          41. Recio-Garc´ıa, J.A.,
          <string-name>
            <given-names>D</given-names>
            <surname>´</surname>
          </string-name>
          ıaz-Agudo,
          <string-name>
            <given-names>B.</given-names>
            ,
            <surname>Pino-Castilla</surname>
          </string-name>
          ,
          <string-name>
            <surname>V.</surname>
          </string-name>
          :
          <article-title>CBR-LIME: A case-based reasoning approach to provide specific local interpretable model-agnostic explanations</article-title>
          .
          <source>In: Case-Based Reasoning Research and Development: ICCBR-20</source>
          . pp.
          <fpage>179</fpage>
          -
          <lpage>194</lpage>
          . Springer, Online (
          <year>2020</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref42">
        <mixed-citation>
          42.
          <string-name>
            <surname>Richter</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          : Introduction. In: Lenz,
          <string-name>
            <surname>M.</surname>
          </string-name>
          , Bartsch-Sp¨orl,
          <string-name>
            <given-names>B.</given-names>
            ,
            <surname>Burkhard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.D.</given-names>
            ,
            <surname>Wess</surname>
          </string-name>
          , S. (eds.) CBR Technology: From Foundations to Applications, chap. 1, pp.
          <fpage>1</fpage>
          -
          <lpage>15</lpage>
          . Springer, Berlin (
          <year>1998</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref43">
        <mixed-citation>
          43.
          <string-name>
            <surname>Roth-Berghofer</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Explanations and case-based reasoning: Foundational issues</article-title>
          .
          <source>In: Advances in Case-Based Reasoning</source>
          . pp.
          <fpage>389</fpage>
          -
          <lpage>403</lpage>
          . Springer, Berlin (
          <year>2004</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref44">
        <mixed-citation>
          44.
          <string-name>
            <surname>Rudin</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead</article-title>
          .
          <source>Nature Machine Intelligence</source>
          <volume>1</volume>
          ,
          <fpage>206</fpage>
          -
          <lpage>215</lpage>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref45">
        <mixed-citation>
          45.
          <string-name>
            <surname>Schoenborn</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Weber</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Aha</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cassens</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          , Althoff,
          <string-name>
            <surname>K.</surname>
          </string-name>
          :
          <article-title>Explainable casebased reasoning: A survey</article-title>
          .
          <source>In: Proceedings of the AAAI Explainable Agency in Artificial Intelligence Workshop</source>
          (
          <year>2020</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref46">
        <mixed-citation>
          46.
          <string-name>
            <surname>Sizov</surname>
            , G., O¨ztu¨rk,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bach</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>Evaluation of explanations extracted from textual reports</article-title>
          .
          <source>In: Proceedings of the Twenty-Ninth International Florida Artificial Intelligence Research Society Conference, FLAIRS-16</source>
          . pp.
          <fpage>425</fpage>
          -
          <lpage>429</lpage>
          . AAAI Press (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref47">
        <mixed-citation>
          47.
          <string-name>
            <surname>Sormo</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cassens</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Aamodt</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Explanation in case-based reasoningperspectives and goals</article-title>
          .
          <source>Artificial Intelligence Review</source>
          <volume>24</volume>
          (
          <issue>2</issue>
          ),
          <fpage>109</fpage>
          -
          <lpage>143</lpage>
          (
          <year>2005</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>