Evaluating CBR Explanation Capabilities: Survey and Next Steps Lawrence Gates and David Leake Luddy School of Informatics, Computing, and Engineering, Indiana University Bloomington IN 47408, USA gatesla@indiana.edu, leake@indiana.edu Abstract. The promise of case-based reasoning (CBR) methods for ex- plaining intelligent systems has prompted much interest, including inter- est in explaining the CBR process itself. However, there is an impediment to understanding and advancing explanations of the CBR process: There has been little assessment of how well they perform. This paper develops strategies for filling this gap. We first highlight selected work on eval- uating systems in explainable artificial intelligence, and then present a selective survey of work on explainable CBR systems to identify those that have been assessed with human subjects studies. From this we de- velop a proposed set of dimensions for characterizing explanation com- ponents of CBR systems. These dimensions provide a framework that can be used to guide evaluation, as well as a vehicle for identifying new research questions. We close with proposed next steps for research and for community initiatives. Keywords: Explainable Case-based Reasoning, Explanation, Evalua- tion, Human Subjects, Survey 1 Introduction Explanation of intelligent systems is a major area of AI research. Progress has been driven both by AI research initiatives, most notably DARPA’s Explainable Artificial Intelligence (XAI) Program [17], and by practical needs arising from the European Union directive on the “right to explanation” for conclusions from decision-making systems [16]. The CBR community has long seen interpretabil- ity as a key advantage of CBR (e.g., [25, 27]). There is substantial evidence for human use of case-based reasoning [26], and experience with CBR applications suggests that people find prior examples a natural vehicle for explaining de- cisions. Numerous approaches have been proposed for using CBR to generate explanations for black box systems (e.g., [20, 21, 36]), including “twinning” CBR and black box systems and using cases to explain the black box system results (see Keane and Kenny [20] for a survey). There are also strong arguments for focusing directly on the construction of interpretable systems rather than at- tempting to augment noninterpretable ones with explanatory capabilities [44]— another opportunity for CBR. Workshops on explainable CBR (XCBR) [2, 6] Copyright © 2021 for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0). have highlighted opportunities both for applying CBR to explain other systems and for explaining the behavior of CBR systems. This paper focuses on needs for explaining CBR system behavior. Researchers have observed that simply presenting the most similar case in a case base may not be sufficient for optimal explanation (e.g., [11, 31, 32, 43]). A challenge for CBR is assessing the explanations that are generated. The obvious approach—and the only definitive method—for assessing the usefulness of explanations to end users is human subjects studies. However, such studies require substantial resources and few have been done for CBR explanation systems. Keane and Kenny refer to this gap as “[t]he embarrassment of user testing.” In their extensive survey of twinned CBR-ANN systems, they “found less than a handful (i.e., < 5) that performed any adequate user testing of the proposal that cases improved the interpretability of models” [20]. Belén Dı́az Agudo’s ICCBR-19 keynote talk on explanation listed human subjects studies of human-oriented explanations as one of four key explanation challenges for the CBR community.1 This paper aims to solidify understanding of the state of the art in assessing explanation of CBR systems. First, it presents general background on the nature of explanation and human subjects evaluation in XAI. Second, it presents a survey of human-subjects evaluations of explanation capabilities in CBR, to provide a reference for the methods and capabilities that have been assessed. Third, it categorizes this work, to identify the space to be explored and current gaps. Finally, based on the categorization and observations of prior research, it proposes five steps—both short and longer term—for advancing the evaluation of explanation of CBR systems and their explanation capabilities. 2 Aims of Explanation Numerous studies have examined what makes a good explanation and how to explain in XCBR. Kenny et al. [21] propose a “tricorn” user model for XAI, with explanations relating to a user’s domain model, model of the system, and model of how particular explanation strategies (e.g., graphical presentation of information) relate evidence to a conclusion. Schoenborn et al. [45] survey ex- plainable CBR and Sormo, Cassens and Aamodt [47] survey explanation in CBR with an emphasis on criteria for good explanations. In general, explanations can serve a rich range of needs, depending on the user’s knowledge state, goals, and tasks [24]. Here we focus on two fundamental goals. One is transparency, under- standing how the system produced its answer. This may be useful both to foster trust (e.g., [14]) and to diagnose problems to guide system repair or refinement of system knowledge or processes. The other is justification, understanding why the system’s answer is a good answer [47]. 1 https://iccbr2019.com/wp-content/uploads/2019/09/Keynote-BelenDiazAgudo.pdf 3 Assessing Explanation Systems Given the goal of serving a human user, the gold standard for system assess- ments must be the results of human subjects studies. Human subjects studies can be designed directly to assess and compare explanation quality of alterna- tive systems (e.g., [10, 30, 37, 38]). However, such assessments are challenging to design and may be time-consuming and expensive, especially given that many aspects of system processing (e.g., situation assessment, similarity assessment, or adaptation) might need to be explained. As a result, few XCBR efforts use human subject evaluations and the space of possible factors affecting explana- tion effectiveness is sparsely covered. Some studies rely on informal evaluations, seeking reactions of a small set of subjects, and many system designs simply rely on the system-builder’s intuitions. A goal of this paper is to provide a foundation for choosing aspects of a system to evaluate and to help identify the aspects that are most and least understood, to illuminate areas for future study. 4 Categorizing Explanations of CBR Systems This section presents a categorization of research on explaining CBR systems. We begin by introducing the dimensions used and then place research accord- ing to those dimensions. The dimensions primarily consider the aspects of the CBR process being explained and the knowledge (derived from the CBR knowl- edge containers) that is the focus of those explanations. Of particular interest is identifying systems that have been evaluated with human subjects studies and opportunities for future human subjects studies. Schoenborn et al. [45] present a complementary survey, categorizing XCBR systems according to four types of aspects: definitional (explanation goal and kinds such as why-explanations or how-explanations), model-based or model-agnostic explanation, and (overlap- ping with this survey) the medium of presentation. We note that XCBR is not limited to explanation of CBR systems, including, for example, much interesting work on CBR to support explanation in non-CBR systems (such as [20, 41]). However, that work is outside the scope of this paper. Explanations of the conclusions of a CBR system may focus on the solution provided by the system or the results of any of the intermediate steps in the CBR cycle—retrieve, reuse, revise, or retain [1]—with an account of how those results contribute to the solution. For example, CBR system behavior could be explained by the retrieved case, or a retrieved case plus how it was adapted. In principle, the retrieval could be explained both by similarity criteria and by how previous retention decisions affected the choice of cases to retain. For any of the steps, the generation of a particular result depends on the knowledge brought to bear. Knowledge from any of the CBR knowledge contain- ers [42] could be relevant to an explanation. In addition, given the well known tight coupling of knowledge containers, there may be a choice in determining which type of knowledge to highlight for a user (e.g., indexing or similarity knowledge accounting for retrieval of a case, if case relevance may not be obvi- ous to the user, or case adaptation knowledge, if case relevance is intuitive but the adapted solution may not be). Each of these choices provides a dimension for categorizing explanations of CBR systems. 4.1 Categorization Table The table in Figure 1 categorizes the explanation approaches taken by sixteen projects. Research evaluated with human subjects studies is highlighted in bold. Rows are divided into four sections. The first section categorizes the research by Explanation Medium, the presentation of the explanation. The medium may be in graphical/visual or textual (explanations in internal system symbolic repre- sentations are considered to be in textual form). The second section categorizes the research by the Target of Explanation. Targets of the explanations may be the system’s non-adapted final result, the reason for the retrieval of the case used, the result when the retrieved case was adapted, or aspects of the ongoing system process. The third section of rows categorizes the research by the primary Knowledge Focus of the explanation. A holistic focus provides an entire retrieved case to the user; the user determines how the case relates to explanatory needs. A key features focus provides information about the case features salient for retrieval. An index knowledge focus explains retrievals in terms of memory organization, the knowledge used to generate indices from cases, and/or knowledge underlying how those indices may be revised or transformed during the retrieval process. An adaptation knowledge focus explains the generation of a solution in terms of the application of knowledge such as adaptation rules, adaptation cases, or other knowledge from the adaptation knowledge container. A similarity focus explains retrieval in terms of the similarity measure and/or analysis of case features in terms of similarity criteria. In principle, a system explaining CBR system behav- ior could have multiple knowledge focuses. However, in practice, most research has focused on single focus types. The final two sections of rows categorize the research by Explanation Recipient and the Process or Result Focus. The recipient of the explanation may be a human or an AI system (as in systems that explain internally to guide processing). The focus of explanation is the final the system process used to obtain it (for transparency) or system result (for justification). 4.2 The Systems Categorized Armengol, Ontanón and Plaza [4] present a method for explaining similarity to justify solutions in classification tasks. The approach was applied to provide justification to other agents in a multi-agent system. This system focuses on ex- plaining the similarity of the non-adapted final result in text (symbolic) format, for AI systems. Burke and Kass [7] present a system that explains the retrieval result for video clips with “bridging text” accounting for the relevant indices, focusing on index knowledge, for presentation to a human user. Cunningham, Doyle and Loughrey [10] performed, to our knowledge, the first human subjects study on prior cases as explanations. Their study, using Cunningham, Doyle & Loughrey, 03 Nugent, Cunningham & Doyle, 05 Doyle, Cunningham & Walsh, 06 Maximini, Freßman & Schaaf, 04 Armengol, Ontanón & Plaza, 04 Massie, Craw & Wiratunga, 04 Muhammad et al., 15 Burke & Kass, 96 Lieber et al, 08 Lamy et al, 19 McSherry, 04 McSherry, 01 Ong et al., 97 Kolodner, 83 Leake, 91 Kass, 90 a Explanation Visualization   Medium Text               Non-Adapted           Final Result Retrieval    Target for Result Explanation Adaptation   Result Ongoing  Process Case Only       (Holistic) Key Features     Knowledge Index    Focus Knowledge Adaptation   Knowledge Similarity   Explanation Human User               Recipient AI System   Process or System Process      Result Focus System Result            Fig. 1. A Categorization of Explanations of Case-Based Reasoning. 37 subjects, compared the convincingness of no explanation, presentation of the case on which a CBR system’s conclusion was based, and rule-based explanation for predictions of blood alcohol levels. The system’s target is the non-adapted final result in text format, with holistic knowledge focus, to a human user. Doyle, Cunningham and Walsh [12] study explanation via a selected prior case (not necessarily the nearest case), justification text and a confidence level, in the domain of hospital admission and discharge decisions. The justification text highlighted features that support or undermine relevance of the explanation case. The study had two evaluation stages, a formative evaluation and a second stage of two studies, one with 106 domain experts judging confidence in the system and the appropriateness of the explanation case, and the second with 14 domain experts judging whether the nearest neighbor or explanation case better explained the system decision. The system’s target is the non-adapted final result in text format, with holistic knowledge focus, to a human user. Kass [19] presents a case adaptation system that provides textual descriptions of adaptation paths that could be presented to explain adaptations to a user. The system’s target is the adaptation result in text format, with knowledge focus of adaptation knowledge, for a human user. Kolodner [22] presents explanation of retrieval by a question transformation trace leading to the retrieval result in a text format, with the knowledge focus index knowledge to a human user. Lamy et al. [23] focus on visualization of differences and similarities between a query case and similar cases, in the context of a CBR system to diagnose breast cancer. Their system was evaluated by 11 users who selected a treatment based on the data presented by the system and indicated their confidence level in the choice. The target is the non-adapted final result in the form of a visualization, with a holistic knowledge focus, to a human user. Leake [24] presents a set of criteria for evaluating CBR-generated explana- tions of anomalous events in stories for various goals (e.g., prediction of an event or prevention of an event). The system’s target is the non-adapted final result in text format, with a knowledge focus of holistic, for an AI system. Lieber et al. [29] present an approach to explaining adaptations, in the con- text of breast cancer treatment, by presenting adaptation paths with annotations of each step. The system’s target is the adaptation result in text format, with knowledge focus on adaptation knowledge, to a human user. Massie, Craw and Wiratunga [30] present a visualization tool aimed at pre- senting feature similarity between a new problem and retrieved case, in the domain of drug design. The work is assessed by asking two domain experts to assess similarity and express their confidence in the answer the solution provided. The target for explanation is the retrieval result in the form of a visualization, with a holistic knowledge focus and explanation to a human user. Maximini, Freßman, and Schaaf [31] present a system including a range of explanation capabilities for an intellectual property domain. The system includes two explanation components. The first explains the set of cases retrieved in terms of attributes and supports user navigation through the candidates, as well as presenting the most decisive attributes to narrow the retrieval set. When no candidates are retrieved, it can also help the user understand which query features could not be matched, to aid query refinement. The other supports examination of the result, with color coding of features by similarity as well as a graphical visualization and textual explanation of feature connections. The system’s target is the non-adapted final result in text format, with a knowledge focus of the similarity for a human user. McSherry [32] presents an explanation method that provides the features of the nearest cases that support or oppose a prediction. This system focuses on explaining the non-adapted final result in form of case text, with knowledge focus on key features to a human user. McSherry [33] presents an explanation method for conversational CBR for diagnosis of computer faults. The system provides explanations of the relevance of questions. This system focuses on explaining the ongoing process in the form of text, with knowledge focus on index generation, to a human user. Muhammad et al. [34] present an approach to explaining recommendations based on positive and negative features compared to user preferences, in a hotel recommendation domain. The recommender personalizes explanations by high- lighting the features the given user seeks. This system focuses on explaining the non-adapted final result in form of text, with a knowledge focus on key features to a human user. Nugent, Cunningham and Doyle [37] extend their original study [10] for an ex- planation component providing the retrieved case, text explaining the difference between the query and explanation case, and a confidence value. If confidence is below a threshold, an additional counter case is provided, with a description of how feature values in both cases impact classification. The study used 12 human subjects. The system target is the non-adapted final result in text format, with a holistic knowledge focus and explanation for a human user. Ong et al. [39] present a system that explains similarity by presenting sim- ilarity scores for matching features, in a cancer diagnosis domain. This system focuses on explaining the non-adapted final result in form of text, with knowledge focus on the case as a whole, so holistic, to a human user. 5 Next Steps Examination of research in the survey suggests five steps for advancing evalua- tion of explanation of CBR systems. The first, building on the dimensions of the categorization, is to fill gaps in coverage of evaluation dimensions in the cate- gorizations. The second and third involve community effort to make evaluations more practical to perform and results more comparable. This requires both ad- vancing automated evaluations, to enable gathering initial evaluation data more rapidly, and establishing standard datasets and standardized evaluation designs, in order to make studies more useful for understanding the relationships between different explanation focuses and how they contribute to the benefit of explana- tion. The fourth and fifth are to move beyond the current space of explanations, focused on local explanation of a single solution, to build explanations to help users better understand how systems will perform in the future and to explain system failures. Understanding the full scope of explanation of CBR systems will require experiments that assess the full system, including effects of component interactions. 5.1 Using the Categorization Scheme to Suggest Evaluation Targets The categorization table of the previous section shows the distribution of hu- man subjects studies for explaining CBR. We observe that existing evaluations have primarily focused on case-oriented explanations—explaining by presenting cases and highlighted features to the user. This has been important for assessing the fundamental CBR premise that cases are a natural mode of explanation. However, in domains unfamiliar to the user, explanatory support for assessing similarity may be crucial. This received some recognition [4, 30], but so far, to our knowledge, only one method has been tested with human subjects, and in a small-scale assessment (by two experts) [30]. Adaptation-focused explanation has also received little attention. The survey also suggests the potential for studies of how different modes of information presentation affect efficacy of explanations. Systematic evaluation will require understanding: – Alternative methods and tradeoffs for conveying to the user information about system knowledge in each of the CBR knowledge containers [42]: the vocabulary, similarity measure, case base, and adaptation knowledge. – How best to explain the processes underlying each step of the CBR cycle: retrieve, reuse, revise, and retain [1]. – The influence of information needs [24, 47]: how to select parts of that process to highlight to particular users. The ideal result would be a toolkit of best practices for explaining classes of CBR systems based on the tasks they perform. For example, explaining a “compare and contrast” task, as in American legal reasoning [5], will necessarily differ from explaining anomalies [24]. Expert systems research on generic tasks [8] suggests the potential to identify a comparatively limited set of task classes. This in turn would increase the practicality and usefulness of human subjects studies by encouraging studies with wide task applicability. A community effort to develop such a collection could leverage both XCBR research and the application of XCBR in domains for which explanation is important. 5.2 Advancing Automatic Evaluation There is a longstanding tradition of testing conversational CBR systems with simulated users (e.g., [3]). Research on case-based explanation generation has ex- amined criteria for evaluating the goodness of explanations generated [18, 24, 46]. This prior work raises an interesting question: Would it be possible to develop automated evaluation schemes to test aspects of the explanations generated by CBR systems? Even if such methods did not have complete fidelity to human subjects studies, automatic testing could be a valuable intermediate step. Work on case-based explanation for story understanding systems has devel- oped criteria for evaluating explanations according to their ability to resolve anomalies in stories, address system questions, and satisfy other explanation goals [24, 40]. Applying such methods to a broader class of tasks would require analysis of the information needs associated with those tasks and how they can be satisfied. Providing a general solution for this would be an enormous task. However, it could be more feasible in testbed microdomains—were the commu- nity to embrace developing such resources. 5.3 Developing Standardized Evaluation Resources The ability to compare and replicate human subjects studies is impeded by the lack of agreed-upon testbeds for such evaluations. Different research groups use different data sets, making it more difficult to assess the relative contributions of explanations of different parts of the CBR process and how those components interact. From the five human subject studies reported in this paper’s categoriza- tion, two (done by the same research group) used the blood-alcohol domain [10, 37], one used the tablet formulation domain [30], one used Bronchiolitis medical data [12], and one used breast cancer data [23]. Of the 4 unique domains, only the breast cancer data [13] is publicly available. None of these data sets involve CBR systems including case adaptation. The development of the needed data sets—and agreement to use them—must be a community endeavor. The CBR community has a history of developing shared resources and advancing the state of the art with competitions, as in the Computer Cooking Competition series (e.g., [35]). This series included var- ious forms of human evaluations, both by human judges and by the conference attendees as a whole. An explanation challenge, based on standard tasks and data sets and informally evaluated by humans as part of the competition, could “jump start” the development of standard resources. In addition, the design of the competition could help establish evaluation protocols to form the basis for more formal human subjects studies. 5.4 Explaining the prior cases themselves Explanations based on presenting prior cases assume the quality of the retrieved case can be trusted—that no justification is needed for why the case should be believed. However, especially if cases have been automatically generated by adaptation processes, understanding and accepting a case as an explanation depends on understanding the presented case and its relevance. That may depend on understanding how it was derived—its provenance [28]. The use of provenance for explaining CBR is an interesting area for future research. 5.5 Broadening CBR Explanations to Explaining Failures Existing studies of explaining CBR focus primarily on accounting for why the system reasoned as it did. Users may use that information to determine the applicability of a proposed solution. This is sufficient to determine trust in the current solution. However, when users determine that the system went astray, the explanation components generally have no capability to explain why the CBR process failed. In the absence of such an explanation it is harder to determine trust in the system’s future processing. Some CBR research has applied metareasoning to explain and repair reason- ing failures within the CBR process for internal system learning (e.g., [9, 15]). This points to another topic for CBR system explanation: accounting for expec- tation failures in how the system operates, to help users refine their predictive model of the CBR system overall. 6 Conclusion This paper has presented an initial categorization of research on explaining CBR systems, highlighting dimensions for which human subjects evaluations have been done—and for which they are still needed. Based on this categorization it has proposed XCBR research opportunities including increased attention to explaining similarity and adaptation. It has also advocated new initiatives on automatic evaluation, building community evaluation resources, explaining case provenance, and explaining system failures. 7 Acknowledgments We acknowledge support from the Department of the Navy, Office of Naval Re- search (Award N00014-19-1-2655), and the US Department of Defense (Contract W52P1J2093009). References 1. Aamodt, A., Plaza, E.: Case-based reasoning: Foundational issues, methodological variations, and system approaches. AI Communications 7(1), 39–52 (1994) 2. Aha, D., Agudo, B.D., Garcia, J.R. (eds.): Proceedings of XCBR-2018: First work- shop on case-based reasoning for the explanation of intelligent systems. ICCBR (2018), http://iccbr18.com/wp-content/uploads/ICCBR-2018-V3.pdf 3. Aha, D., Breslow, L.: Refining conversational case libraries. In: Proceedings of the Second International Conference on Case-Based Reasoning. pp. 267–278. Springer Verlag, Berlin (1997) 4. Armengol, E., Ontanón, S., Plaza, E.: Explaining similarity in CBR. In: ECCBR 2004 Workshop Proceedings. pp. 155–164 (2004) 5. Ashley, K., Rissland, E.: Compare and contrast, a test of expertise. In: Proceedings of the Sixth Annual National Conference on Artificial Intelligence. pp. 273–284. AAAI, Morgan Kaufmann, San Mateo, CA (1987) 6. B. Diaz Agudo, J.R.G., Watson, I. (eds.): Proceedings of XCBR-2019: Second workshop on case-based reasoning for the explanation of intelligent systems. IC- CBR (2018) 7. Burke, R., Kass, A.: Retrieving stories for case-based teaching. In: Leake, D. (ed.) Case-Based Reasoning: Experiences, Lessons, and Future Directions, pp. 93–109. AAAI Press, Menlo Park, CA (1996) 8. Chandrasekaran, B.: Generic tasks in knowledge-based reasoning: High level build- ing blocks for expert system design. IEEE Expert 1(3), 23–30 (1986) 9. Cox, M., Ram, A.: Introspective multistrategy learning: On the construction of learning strategies. Artificial Intelligence 112(1-2), 1–55 (1999) 10. Cunningham, P., Doyle, D., Loughrey, J.: An evaluation of the usefulness of case- based explanation. In: Case-Based Reasoning Research and Development: Proceed- ings of the Fifth International Conference on Case-Based Reasoning, ICCBR-03. pp. 122–130. Springer-Verlag, Berlin (2003) 11. Doyle, D., Cunningham, P., Bridge, D., Rahman, Y.: Explanation oriented re- trieval. In: Funk, P., González Calero, P.A. (eds.) Advances in Case-Based Rea- soning. pp. 157–168. Springer, Berlin, Heidelberg (2004) 12. Doyle, D., Cunningham, P., Walsh, P.: An evaluation of the usefulness of explana- tion in a case-based reasoning system for decision support in bronchiolitis treat- ment. Computational Intelligence 22(3-4), 269–281 (2006) 13. Dua, D., Graff, C.: UCI machine learning repository (2017), http://archive.ics.uci.edu/ml 14. Floyd, M., Aha, D.: Special issue on case-based reasoning. AI Communications 30(3-4), 281–294 (2017) 15. Fox, S., Leake, D.: Introspective reasoning for index refinement in case-based rea- soning. The Journal of Experimental and Theoretical Artificial Intelligence 13(1), 63–88 (2001) 16. Goodman, B., Flaxman, S.: European Union regulations on algorithmic decision- making and a “right to explanation”. AI Magazine 38(3), 50–57 (2017) 17. Gunning, D., Aha, D.W.: DARPA’s explainable artificial intelligence program. AI Magazine 40(2), 44–58 (2019) 18. Johs, A.J., Lutts, M., Weber, R.O.: Measuring explanation quality in XCBR. In: Minor, M. (ed.) Workshop proceedings ICCBR 2018. pp. 75–83 (2018) 19. Kass, A.: Developing Creative Hypotheses by Adapting Explanations. Ph.D. thesis, Yale University (1990), northwestern University Institute for the Learning Sciences, Technical Report 6 20. Keane, M.T., Kenny, E.M.: How case-based reasoning explains neural networks: A theoretical analysis of XAI using post-hoc explanation-by-example from a survey of ANN-CBR twin-systems. In: Case-Based Reasoning Research and Development: ICCBR-19. pp. 155–171. Springer, Berlin (2019) 21. Kenny, E.M., Ford, C., Quinn, M., Keane, M.T.: Explaining black-box classifiers using post-hoc explanations-by-example: The effect of explanations and error-rates in XAI user studies. Artificial Intelligence 294, 103459 (2021) 22. Kolodner, J.L.: Reconstructive memory: A computer model. Cognitive Science 7(4), 281–328 (1983) 23. Lamy, J.B., Sekar, B., Guezennec, G., Bouaud, J., Séroussi, B.: Explainable artifi- cial intelligence for breast cancer: A visual case-based reasoning approach. Artificial intelligence in medicine 94, 42–53 (2019) 24. Leake, D.: Goal-based explanation evaluation. Cognitive Science 15(4), 509–545 (1991) 25. Leake, D.: CBR in context: The present and future. In: Leake, D. (ed.) Case-Based Reasoning: Experiences, Lessons, and Future Directions, pp. 3–30. AAAI Press, Menlo Park, CA (1996), http://www.cs.indiana.edu/˜leake/papers/a-96-01.html 26. Leake, D.: Cognition as case-based reasoning. In: Bechtel, W., Graham, G. (eds.) A Companion to Cognitive Science, pp. 465–476. Blackwell, Oxford (1998) 27. Leake, D., McSherry, D.: Introduction to the special issue on explanation in case- based reasoning. Artificial Intelligence Review 24(2), 103–108 (2005) 28. Leake, D., Whitehead, M.: Case provenance: The value of remembering case sources. In: Case-Based Reasoning Research and Development: ICCBR-07. pp. 194–208. Springer-Verlag, Berlin (2007) 29. Lieber, J., d’Aquin, M., Badra, A., Napoli, A.: Modeling adaptation of breast cancer treatment decision protocols in the KASIMIR project. Applied Intelligence 28(3), 261–274 (2008) 30. Massie, S., Craw, S., Wiratunga, N.: A visualisation tool to explain case-base rea- soning solutions for tablet formulation. In: Proceedings of the 24th SGAI Interna- tional Conference on Innovative Techniques and Applications of Artificial Intelli- gence. Springer-Verlag, Berlin (2004) 31. Maximini, R., Freßmann, A., Schaaf, M.: Explanation service for complex CBR ap- plications. In: Advances in Case-Based Reasoning, ECCBR. pp. 302–316. Springer, Berlin (2004) 32. McSherry, D.: Explaining the pros and cons of conclusions in CBR. In: Advances in Case-Based Reasoning. pp. 317–330. Springer, Berlin (2004) 33. McSherry, D.: Interactive case-based reasoning in sequential diagnosis. Applied Intelligence 14(1), 65–76 (2001) 34. Muhammad, K., Lawlor, A., Rafter, R., Smyth, B.: Great explanations: Opinion- ated explanations for recommendations. In: Case-Based Reasoning Research and Development. pp. 244–258. Springer, Berlin (2015) 35. Najjar, N., Wilson, D.: Computer cooking contest workshop at the twenty- fifth international conference on case-based reasoning (ICCBR 2017). In: ICCBR 2017 Workshops, Doctoral Consortium, and Competitions (2017), http://ceur- ws.org/Vol-2028/XXCCC17 preface.pdf 36. Nugent, C., Cunningham, P.: A case-based recommender for black-box systems. Artificial Intelligence Review 24(2), 163–178 (2005) 37. Nugent, C., Cunningham, P., Doyle, D.: The best way to instil confidence is by being right. In: International Conference on Case-Based Reasoning. pp. 368–381. Springer (2005) 38. Nugent, C., Doyle, D., Cunningham, P.: Gaining insight through case-based expla- nation. Journal of Intelligent Information Systems 32(3), 267–295 (Jun 2009) 39. Ong, L.S., Shepherd, B., Tong, L.C., Seow-Choen, F., Ho, Y.H., Tang, C.L., Ho, Y.S., Tan, K.: The colorectal cancer recurrence support (CARES) system. Artificial Intelligence in Medicine 11(3), 175–188 (1997) 40. Ram, A.: AQUA: Asking questions and understanding answers. In: Proceedings of the Sixth Annual National Conference on Artificial Intelligence. pp. 312–316. Morgan Kaufmann, Seattle, WA (July 1987) 41. Recio-Garcı́a, J.A., Dı́az-Agudo, B., Pino-Castilla, V.: CBR-LIME: A case-based reasoning approach to provide specific local interpretable model-agnostic expla- nations. In: Case-Based Reasoning Research and Development: ICCBR-20. pp. 179–194. Springer, Online (2020) 42. Richter, M.: Introduction. In: Lenz, M., Bartsch-Spörl, B., Burkhard, H.D., Wess, S. (eds.) CBR Technology: From Foundations to Applications, chap. 1, pp. 1–15. Springer, Berlin (1998) 43. Roth-Berghofer, T.: Explanations and case-based reasoning: Foundational issues. In: Advances in Case-Based Reasoning. pp. 389–403. Springer, Berlin (2004) 44. Rudin, C.: Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nature Machine Intelligence 1, 206–215 (2019) 45. Schoenborn, J., Weber, R., Aha, D., Cassens, J., Althoff, K.: Explainable case- based reasoning: A survey. In: Proceedings of the AAAI Explainable Agency in Artificial Intelligence Workshop (2020) 46. Sizov, G., Öztürk, P., Bach, K.: Evaluation of explanations extracted from tex- tual reports. In: Proceedings of the Twenty-Ninth International Florida Artificial Intelligence Research Society Conference, FLAIRS-16. pp. 425–429. AAAI Press (2016) 47. Sormo, F., Cassens, J., Aamodt, A.: Explanation in case-based reasoning— perspectives and goals. Artificial Intelligence Review 24(2), 109–143 (2005)