<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Fostering Human-AI interaction: development of a Clinical Decision Support System enhanced by eXplainable AI and Natural Language Processing</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Laura</forename><surname>Bergomi</surname></persName>
							<email>laura.bergomi01@universitadipavia.it</email>
							<affiliation key="aff0">
								<orgName type="department">Department of Electrical, Computer and Biomedical Engineering</orgName>
								<orgName type="institution">University of Pavia</orgName>
								<address>
									<settlement>Pavia</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Fostering Human-AI interaction: development of a Clinical Decision Support System enhanced by eXplainable AI and Natural Language Processing</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">E883CD54B6A8017401614839343B60A6</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T19:37+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Clinical decision making</term>
					<term>Explainable artificial intelligence</term>
					<term>Natural language interaction</term>
					<term>Human-AI collaboration protocol</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Artificial Intelligence (AI) is increasingly integrated into Decision Support Systems (DSS). The explainability of AI-based systems becomes crucial in sensitive and critical domains, such as healthcare, where ethical considerations and reliability are paramount concerns. In the clinical setting, it is important to evaluate how humans and AI can collaborate on cognitive tasks. Collaboration protocols (HAI-CP) allow for the investigation of the usefulness of AI models and their impact on users (both positive and negative). Although research on the application of these methods is blooming, there is little understanding of the impact on clinical decision-making, especially for eXplainable AI (XAI) systems, due to the lack of user studies. Therefore, the goal of this proposal is to develop a clinical DSS enhanced by XAI and Natural Language Processing (NLP): their synergy can add value to the interaction between users and AI, fostering a more linguistically natural, comprehensible, trustworthy, and supporting interfacing, that blends into the existing workflows. This proposal explores potential solutions to tailor natural language explanations and data visualizations to the end-user, improving the comprehensibility of the reasons behind a decision, and increasing the user's confidence in the decision; investigates and tests possible strategies to "get the patient-in-the-loop"; explores uncertainty quantification and counterfactual approaches, and finally assesses the impact on naturalistic (i.e., real-world) decision-making and long-term effects and biases.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Artificial Intelligence (AI) is increasingly integrated into Decision Support Systems (DSS). To understand how useful and usable an AI-based system actually is, it is important to assess its impact on the decision support workflow. For a DSS to meet the basic requirement of Art. 22 para 1 GDPR (which prohibits "decision based solely on automated processing"), it is important to provide that the human-in-the-loop has substantial evaluative power and so the last word on the outcome of a decision <ref type="bibr" target="#b0">[1]</ref>. To ensure that a human can make a thoughtful and reliable decision, the explainability of AI-based systems becomes crucial. In sensitive and critical domains, such as healthcare, where ethical considerations and reliability are paramount concerns, AI-based DSSs should provide high-quality explanations. In this context, eXplainable AI (XAI) has become increasingly important as a counterbalancing force to the widespread adoption of complex black box models, that leave users, and even developers, in the dark as to how results were obtained <ref type="bibr" target="#b1">[2]</ref>. XAI refers to the development of AI systems that can provide clear, understandable, and interpretable explanations. XAI methods can be described based on several categorizations; what remains essential is to adapt and test their explanations in Human-Artificial Intelligence collaboration protocol (HAI-CP). An HAI-CP is "the instance of a process schema that stipulates the use of AI tools by competent practitioners to perform a certain task or do a certain job" <ref type="bibr" target="#b2">[3]</ref>, thus allowing to study the interaction of several parameters involving at least the following dimensions : Affordance (functionalities and task automation), Fit (fitting into the existing work practice), Optimization (learning phase tuning), Output (type of result returned) and Target (the characteristics of the intended user).</p><p>Despite the application of AI methods in healthcare being a highly active field of research <ref type="bibr" target="#b3">[4]</ref>, how AI recommendations affect clinical decision-making is still poorly understood due to the lack of user studies <ref type="bibr" target="#b4">[5]</ref>. This motivates the present proposal to investigate extensively how AI systems can be integrated into medical practice, being an aid that is easily understood, but also naturally interfaceable (enabling natural language interaction, NLI). The latter issue can be supported by using Natural Language Processing (NLP) techniques to generate explanations in natural language, that are tailored to the end-user. The type of end-user has intentionally not been specified, suggesting that it may be not only a medical professional, but also the patient (one of the original contributions of the research). GDPR (Recital 71) remarks that providing information about the existence of automated processes, the logic behind them, and their potential consequences should nevertheless be provided voluntarily as a good practice to ensure fairness and transparency. This means patients should have access to explanations for clinical decisions, allowing them to engage in discussions about their care and understand the reasoning behind treatments or examinations.</p><p>This proposal is part of the Italian project PRIN PNRR 2022 "InXAID -Interaction with eXplainable Artificial Intelligence in (medical) Decision-making" (CUP: H53D23008090001 funded by the European Union -Next Generation EU), whose aim is to explore a model-agnostic perspective on the development and evaluation of AI-based (medical) DSSs. Our focus is on the interaction between an AI system and the human user, i.e., understanding to what extend the advice (and the way it is presented) can influence, be understood, and used by users; the interaction features and effects.</p><p>The rest of the paper is organized as follows. A brief review of related work is provided in Section 1.1; Section 2 presents the goals of this research, followed by the planned approaches and methods to achieve them in Section 3. Expected results of our research are proposed in Section 4, and finally, the conclusions are described in Section 5 offering some questions I would like feedback on.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.1.">Related works</head><p>A human-in-the-loop approach has been advocated as essential for proper evaluation of AI for healthcare <ref type="bibr" target="#b5">[6]</ref>: explanations are ultimately directed to a human (expert) interacting with the AI system and should be optimized for this. Following this direction, Gaube et al. <ref type="bibr" target="#b6">[7]</ref> compare the detection of abnormalities in chest radiology images by physicians with the support of an AI or a second human agent; Tschandl et al. <ref type="bibr" target="#b7">[8]</ref> evaluate, both with and without the use of AI, clinicians skin cancers recognition. Cabitza et al. <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b8">9]</ref> propose experimental results of applying XAI to knee MRI and ECG studies, intending to compare the effectiveness of HAI-CP (testing different orders of presentation (human first vs. AI first) and availability of explanations (yes vs. no)). In <ref type="bibr" target="#b8">[9]</ref> the support of the AI system includes both a proposed diagnosis and a textual explanation to back the former one.</p><p>Some surveys (e.g. <ref type="bibr" target="#b9">[10,</ref><ref type="bibr" target="#b10">11]</ref>) have studied the use of natural language techniques in creating explanations, e.g. the synergy of NLP and XAI methods. Sokol and Flach <ref type="bibr" target="#b11">[12]</ref> argue that natural language explanations give the process a natural feeling, increasing the reliability of explanations and helping to gain acceptance from a wider range of users. Despite these considerations, a small part of XAI's works uses natural language presentation methods <ref type="bibr" target="#b9">[10]</ref>. Works in the literature concerning the involvement of humans in clinical decision-making, such as those cited so far, point to the clinician as the user (i.e., the human-in-the-loop). Only a few examples involve patients, e.g., Donatello et al. <ref type="bibr" target="#b12">[13]</ref> address the challenges of a system that supports patients in following a healthy behavior by presenting an XAI system that supports the monitoring of users' behaviors and persuades them to follow a healthy lifestyle.</p><p>Ultimately, some works (e.g. <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b13">14,</ref><ref type="bibr" target="#b14">15]</ref>) emphasize the need to examine biases in AI and XAI support and focus on AI advice effects.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Research goals</head><p>Current XAI research often overlooks the importance of presentation techniques, leading to challenges for researchers and practitioners in selecting appropriate methods for explainability. The lack of comprehensive studies adds complexity and potential errors to the process <ref type="bibr" target="#b9">[10]</ref>. This research will focus primarily on two main dimensions: the output and the target. In the first case (i.e., the output) the goal is to compare the presentation techniques of XAI methods, studying what type of output the end-user prefers and understands. The output of an XAI system can be multimodal (e.g., classes, confidence scores, category lists, visual or textual explanations, etc.). A better understanding of this output allows users to increase the overall acceptability of the system. In the second case (i.e., the target) the goal is to tailor explanations to the intended end-user in terms of profiling (expertise, experience, role, etc.), content, and form. Going towards dialogues that directly engage the user in the explanation process, we can offer rich and personalized interactions that mimic how humans explain their decisions.</p><p>• What factors determine the choice of how to represent the explanations obtained from AI-based methods? • Does XAI support, both in terms of visual aids and textual explanations, have a significant effect on taking better clinical decisions, and reducing errors?</p><p>Clinicians may lack the computer science expertise to comprehend algorithmic decisionmaking processes, highlighting the need for clear communication. As a result, it becomes imperative for clinicians, as decision-makers, to ensure that decisions made through these processes can be clearly communicated to patients, who may have little technical knowledge.</p><p>• What is the impact of asking decision-makers to explain their AI-based decisions? • What is the best way to "get the patient-in-the-loop"?</p><p>Regarding the assessment of impact on decision-making, we are asking about counterfactual explanations and biases in AI and XAI support:</p><p>• What is the impact of providing decision-makers with similar cases, or providing counterfactual outcomes, or playing the devil's advocate role? • Does the AI advice affect the users' performance and proficiency-building processes, both in the short and in the long run? Is this influence also relevant according to the user's experience, expertise, and role? • Are explanations always beneficial or can they induce paradoxically harmful effects?</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Planned approaches and methods</head><p>This section outlines the main phases of the research (as shown in the Gantt diagram in Figure <ref type="figure">1</ref>), the approaches, and methods thought to be used as a starting point in each phase and from which to develop extensions and insights. At each stage, there is the analysis of State-of-Art methods and the implementation of HAI-CPs to test hypotheses and collect reliable and naturalistic (i.e., real-world) decision-making results. Particular attention will be paid to interpreting and ranking the relative strength of claims about the effectiveness of design solutions and their superiority concerning other possible alternatives <ref type="bibr" target="#b15">[16]</ref>. The inXAID project will ensure that the frameworks, metrics and methods developed, as well as the results of the experiments performed throughout the duration of the research activity, will be disseminated to the appropriate target communities and audiences. A research period abroad is also planned. Research phase 1: Tailoring XAI and NLP explanations. In this phase, the aim is to investigate (i) techniques for deriving the explanation and (ii) presentation to the end-user.</p><p>The main XAI techniques to test are related to feature importance analysis (e.g. SHAP <ref type="bibr" target="#b16">[17]</ref>, T-EBAnO <ref type="bibr" target="#b17">[18]</ref>) and the use of surrogate models (e.g. LIME <ref type="bibr" target="#b18">[19]</ref>, AraucanaXAI <ref type="bibr" target="#b19">[20]</ref>). Comparing different models, we can assess which type of data visualization is preferred by users. Among the existing NLP models, we analyze transformers models for two pivotal reasons: they rely on the attention mechanism and they are exceptionally effective for common natural language understanding (NLU) and natural language generation (NLG) tasks <ref type="bibr" target="#b20">[21]</ref>. A textual explanation may be presented in different ways to the end-user. Some techniques to test include: saliency, visualization of the importance scores, showing input-output word alignment, highlighting words in input text, or displaying extracted relations or word clouds; rewrite the explanations by changing linguistic register. A possible comparison to explore concerns, on the one hand, is the self-explaining approach, which generates the explanation at the same time as the prediction; on the other hand, the post-hoc approach, which requires that an additional operation be performed after the return of the prediction. The final step regards the combination of XAI and NLP techniques; a possible simple solution could be rendering XAI explanations through NLG or using XAI techniques to explain image, text, and graph classification models (e.g., Layerwise relevance propagation <ref type="bibr" target="#b21">[22]</ref>); or incorporating feedback from human users to improve the explanations generated by the model. This could involve allowing users to provide feedback on the explanations or incorporating user preferences into the explanation process. In the example of <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b8">9]</ref>, it is also important to test alternative fitting of AI advice in existing practice (e.g., human first vs. AI first, availability of explanations only in critical cases, AI advice on request).</p><p>Research phase 2: Get the patient-in-the-loop. In this phase, the aim is to investigate different strategies to involve the patient in the clinical decision-making process. Firstly, it is necessary to review the existing work in the literature regarding surveys on patient-AI collaborations. Subsequently the idea is to investigate different ways of including patients, possibly going through step: (i) give clinicians help in explaining their decision (patient-out-theloop); (ii) analyze how patients' behavior changes knowing that the doctor's decision about their health status is based on AI advice (patient aware-of-the-loop); (iii) use conversational chatbots to interact with patients and, thus, collect important information (patient-in-the-loop). In this direction, Bennettot et al. <ref type="bibr" target="#b20">[21]</ref> provide an example of how XAI can benefit healthcare, specifically in monitoring diabetic patients. They discuss a virtual coaching system that offers guidance on healthy behaviors based on patient-reported data. When the system detects undesirable behaviors, it generates tailored explanations for clinicians and patients. This approach enables clinicians to adjust treatment decisions as needed, while patients gain confidence, feel supported, and become more engaged in their care decisions. At this level, also for ethical considerations, the inclusion of a professional figure such as a behavioral psychologist, is essential.</p><p>Research phase 3: Assessment of impact on decision-making. In this phase, the aim is to assess how, how much, and when, an AI-based system influences the cognitive processes involved in clinicians' interpretation of AI support. An explanation, whether true or fictitious, could be correlated by Uncertainty Quantification (UQ) assessment. In this context, we can apply different approaches, also integrated in NLP models <ref type="bibr" target="#b17">[18,</ref><ref type="bibr" target="#b22">23]</ref>. On the other hand, some methods explain the model by providing information on feature-perturbed versions of the analyzed instance. These methods fall into the counterfactual explanations methods <ref type="bibr" target="#b20">[21]</ref>. Some common approaches to test could be <ref type="bibr" target="#b23">[24]</ref>: add or delete information to what users know about the facts; create counterfactuals that imagine how the outcome could have been better or worse; construct explanations by ensuring they identify cause-effect or reason-action relations between events; semi-factual about how the outcome could have been the same "even if" the action had been different. Identifying the cognitive processes involved in physicians' interpretation of AI support also precludes analysis of how these same processes may change over time. It can be useful to realize a mapping of reliance patterns and biases affecting human decision-making in the short and long run (e.g., preference for usability over performance, more attention to false negatives than false positives, deskilling/ upskilling etc.); moreover, summarize promising mitigation strategies and research directions to support human critical thinking (e.g., delay showing the AI's prediction and/or explanations, give arguments for non-predicted outcomes, enable to actively explore the data, etc.).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Expected results</head><p>The main objective of this research proposal is to make a significant contribution towards advancing the methodological leveraging of XAI methods and natural language explanations, fostering a more linguistically natural, comprehensible, trustworthy, and supporting interfacing among AI systems and human users, that fits into the existing work practices. On the clinical front, the project's main objective is to enhance decision-making processes by expanding output solutions and taking into consideration aspects that are currently unexplored. These aspects include the contribution of socio-technical elements, experience, expertise, habits, and environmental and temporal information towards cognitive processes. It is expected that "getting the patient-in-the-loop" of decision-making can be an original strength for AI collaborative systems; so that it can also be supportive to the patient, improving well-being. Preliminary results and contributions to date are related to the test of ALFABETO <ref type="bibr" target="#b24">[25]</ref> (whose aim is to aid clinicians during COVID-19 patients' hospital admission through the application of machine learning approaches exploiting clinical and chest x-ray features) in a clinical survey. In this framework, different predictions and explanations are proposed to clinicians: from an interpretable model (i.e., ALFABETO original Bayesian network) and from a black box model (i.e., Gradient boosting), in this latter case, with the explanations of two different XAI approaches (SHAP and Araucana XAI).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Research challenges and future direction</head><p>The research is still at an early stage, and thus, there are several challenges that we should cope with. In XAI, some works have claimed that explainability may come at the price of losing predictive performance. Studying such possible trade-offs is an important research area, but one that cannot advance until standardized metrics are developed for evaluating the quality of explanations <ref type="bibr" target="#b25">[26]</ref>. Indeed, an open debate in the NLG community is about finding the right way to measure the goodness of generated explanations <ref type="bibr" target="#b10">[11]</ref>. The main issues revolve around whether to rely only on automatic metrics (e.g., ROUGE, BLEU), or instead, how to properly perform human evaluations. That said, although human evaluation remains the gold standard for overall system quality assessment, using it at every stage of the development process would be too costly and slow. Another challenge may be the interactive integration of desired behavior, fairness, correctness, and reliability; as well as verifying and integrating knowledge in each step of decision-making process, instead of XAI producing a single description of a static system <ref type="bibr" target="#b9">[10]</ref>. From these general considerations, and more, arise the following questions, on which I would appreciate feedback by the DC mentors, as I believe they can help improve my research path:</p><p>• What metrics would be better to investigate (automatic or human) to evaluate the goodness of explanations, especially in natural language? • How to engage the user interactively, hoping for fruitful and continuous use of the provided DSS? • What other human actors (besides clinicians and patients) could we include?</p><p>• What aspects should we not forget to consider? (e.g., from an ethical and legal perspective) • What approaches and potential collaborations should we consider to address the challenges of developing a reliable and explainable DSS using clinical data?</p><p>I would like to express my willingness to receive constructive criticism and suggestions on this work and to clarify any points that may need further explanation.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>1 :RP 2 : 3 Figure 1 :</head><label>1231</label><figDesc>Figure 1: Gantt chart of the possible timeline of Research phases (RP) and related activities. * denotes the duration of the inXAID project. Abbreviations: SoA= analysis of State-of-Art methods and approches, TI= Tools Implementation, ES= Evaluation Study with users, RA= Results Analysis, RP= Review Pubblication.</figDesc></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>I would like to express my gratitude to my supervisor, Enea Parimbelli, for his invaluable assistance in developing this project and its implementation in the future. Authors acknowledge funding support provided by the Italian project PRIN PNRR 2022 InXAID -Interaction with eXplainable Artificial Intelligence in (medical) Decision-making. CUP: H53D23008090001 funded by the European Union -Next Generation EU.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">The european legal framework for medical ai</title>
		<author>
			<persName><forename type="first">D</forename><surname>Schneeberger</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-030-57321-8_12</idno>
		<idno>-3-030-57321-8_12</idno>
		<ptr target="org/10.1007/978" />
	</analytic>
	<monogr>
		<title level="m">Machine Learning and Knowledge Extraction</title>
				<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="209" to="226" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">A manifesto on explainability for artificial intelligence in medicine</title>
		<author>
			<persName><forename type="first">C</forename><surname>Combi</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.artmed.2022.102423</idno>
		<idno>doi:</idno>
		<ptr target="https://doi.org/10.1016/j.artmed.2022.102423" />
	</analytic>
	<monogr>
		<title level="j">Artificial Intelligence in Medicine</title>
		<imprint>
			<biblScope unit="volume">133</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Rams, hounds and white boxes: Investigating human-ai collaboration protocols in medical diagnosis</title>
		<author>
			<persName><forename type="first">F</forename><surname>Cabitza</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.artmed.2023.102506</idno>
		<idno>doi:</idno>
		<ptr target="https://doi.org/10.1016/j.artmed.2023.102506" />
	</analytic>
	<monogr>
		<title level="j">Artificial Intelligence in Medicine</title>
		<imprint>
			<biblScope unit="volume">138</biblScope>
			<biblScope unit="page">102506</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">The Coming of Age of Artificial Intelligence in Medicine</title>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">L</forename><surname>Patel</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.artmed.2008.07.017</idno>
	</analytic>
	<monogr>
		<title level="j">Artificial intelligence in medicine</title>
		<imprint>
			<biblScope unit="volume">46</biblScope>
			<biblScope unit="page" from="5" to="17" />
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Quod erat demonstrandum? -towards a typology of the concept of explanation for the design of explainable ai</title>
		<author>
			<persName><forename type="first">F</forename><surname>Cabitza</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.eswa.2022.118888</idno>
		<idno>doi:</idno>
		<ptr target="https://doi.org/10.1016/j.eswa.2022.118888" />
	</analytic>
	<monogr>
		<title level="j">Expert Systems with Applications</title>
		<imprint>
			<biblScope unit="volume">213</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<title level="m" type="main">Interactive machine learning for health informatics: when do we need the human-in-the-loop</title>
		<author>
			<persName><forename type="first">A</forename><surname>Holzinger</surname></persName>
		</author>
		<idno type="DOI">10.1007/s40708-016-0042-6</idno>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="page" from="119" to="131" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Do as ai say: susceptibility in deployment of clinical decision-aids</title>
		<author>
			<persName><forename type="first">S</forename><surname>Gaube</surname></persName>
		</author>
		<idno type="DOI">10.1038/s41746-021-00385-9</idno>
	</analytic>
	<monogr>
		<title level="j">NPJ digital medicine</title>
		<imprint>
			<biblScope unit="volume">4</biblScope>
			<biblScope unit="page">31</biblScope>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Human-computer collaboration for skin cancer recognition</title>
		<author>
			<persName><forename type="first">P</forename><surname>Tschandl</surname></persName>
		</author>
		<idno type="DOI">10.1038/s41591-020-0942-0</idno>
	</analytic>
	<monogr>
		<title level="j">Nature Medicine</title>
		<imprint>
			<biblScope unit="volume">26</biblScope>
			<biblScope unit="page" from="1229" to="1234" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Painting the black box white: Experimental findings from applying XAI to an ECG reading setting</title>
		<author>
			<persName><forename type="first">F</forename><surname>Cabitza</surname></persName>
		</author>
		<idno type="DOI">10.3390/make5010017</idno>
	</analytic>
	<monogr>
		<title level="j">Machine Learning and Knowledge Extraction</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="page" from="269" to="286" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">A survey on xai and natural language explanations</title>
		<author>
			<persName><forename type="first">E</forename><surname>Cambria</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.ipm.2022.103111</idno>
	</analytic>
	<monogr>
		<title level="j">Information Processing &amp; Management</title>
		<imprint>
			<biblScope unit="volume">60</biblScope>
			<biblScope unit="page">103111</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Xnlp: A living survey for xai research in natural language processing</title>
		<author>
			<persName><forename type="first">K</forename><surname>Qian</surname></persName>
		</author>
		<idno type="DOI">10.1145/3397482.3450728</idno>
	</analytic>
	<monogr>
		<title level="m">26th International Conference on Intelligent User Interfaces-Companion</title>
				<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="78" to="80" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Conversational explanations of machine learning predictions through class-contrastive counterfactual statements</title>
		<author>
			<persName><forename type="first">K</forename><surname>Sokol</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Flach</surname></persName>
		</author>
		<idno type="DOI">10.24963/ijcai.2018/836</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence</title>
				<meeting>the Twenty-Seventh International Joint Conference on Artificial Intelligence</meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="5785" to="5786" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Persuasive explanation of reasoning inferences on dietary data</title>
		<author>
			<persName><forename type="first">I</forename><surname>Donadello</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Dragoni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Eccher</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">PROFILES/SEMEX@ISWC</title>
				<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="volume">2465</biblScope>
			<biblScope unit="page" from="46" to="61" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">How cognitive biases affect xai-assisted decision-making: A systematic review</title>
		<author>
			<persName><forename type="first">A</forename><surname>Bertrand</surname></persName>
		</author>
		<idno type="DOI">10.1145/3514094.3534164</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2022 AAAI/ACM Conference on AI, Ethics, and Society</title>
				<meeting>the 2022 AAAI/ACM Conference on AI, Ethics, and Society</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="78" to="91" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Biases affecting human decision making in AI-supported second opinion settings</title>
		<author>
			<persName><forename type="first">F</forename><surname>Cabitza</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-030-26773-5_25</idno>
	</analytic>
	<monogr>
		<title level="m">Modeling Decisions for Artificial Intelligence</title>
				<imprint>
			<publisher>Springer International Publishing</publisher>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="283" to="294" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Evidence-based XAI: An empirical approach to design more effective and explainable decision support systems</title>
		<author>
			<persName><forename type="first">L</forename><surname>Famiglini</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.compbiomed.2024.108042</idno>
	</analytic>
	<monogr>
		<title level="j">Computers in Biology and Medicine</title>
		<imprint>
			<biblScope unit="volume">170</biblScope>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<title level="m" type="main">A unified approach to interpreting model predictions</title>
		<author>
			<persName><forename type="first">S</forename><surname>Lundberg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S.-I</forename><surname>Lee</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.1705.07874</idno>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Trusting deep learning natural-language models via local and global explanations</title>
		<author>
			<persName><forename type="first">F</forename><surname>Ventura</surname></persName>
		</author>
		<idno type="DOI">10.1007/s10115-022-01690-9</idno>
	</analytic>
	<monogr>
		<title level="j">Knowledge and Information Systems</title>
		<imprint>
			<biblScope unit="volume">64</biblScope>
			<biblScope unit="page" from="1863" to="1907" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<title level="m" type="main">Explaining the predictions of any classifier</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">T</forename><surname>Ribeiro</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.1602.04938</idno>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
	<note>why should i trust you?</note>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Why did AI get this one wrong? -tree-based explanations of machine learning model predictions</title>
		<author>
			<persName><forename type="first">E</forename><surname>Parimbelli</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.artmed.2022.102471</idno>
	</analytic>
	<monogr>
		<title level="j">Artificial Intelligence in Medicine</title>
		<imprint>
			<biblScope unit="volume">135</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<monogr>
		<title level="m" type="main">A practical guide on explainable AI techniques applied on biomedical use case applications</title>
		<author>
			<persName><forename type="first">A</forename><surname>Bennetot</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2111.14260</idno>
		<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">On pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation</title>
		<author>
			<persName><forename type="first">S</forename><surname>Bach</surname></persName>
		</author>
		<idno type="DOI">10.1371/journal.pone.0130140</idno>
	</analytic>
	<monogr>
		<title level="j">PloS one</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<monogr>
		<title level="m" type="main">Quantifying uncertainty in natural language explanations of large language models</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">H</forename><surname>Tanneru</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2311.03533</idno>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Counterfactuals in explainable artificial intelligence (xai): Evidence from human reasoning</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">M J</forename><surname>Byrne</surname></persName>
		</author>
		<idno type="DOI">10.24963/ijcai.2019/876</idno>
	</analytic>
	<monogr>
		<title level="j">IJCAI</title>
		<imprint>
			<biblScope unit="volume">19</biblScope>
			<biblScope unit="page" from="6276" to="6282" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Bayesian networks in the management of hospital admissions: A comparison between explainable ai and black box ai during the pandemic</title>
		<author>
			<persName><forename type="first">G</forename><surname>Nicora</surname></persName>
		</author>
		<idno type="DOI">10.3390/jimaging10050117</idno>
	</analytic>
	<monogr>
		<title level="j">Journal of Imaging</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<monogr>
		<title level="m" type="main">A survey of the state of explainable AI for natural language processing</title>
		<author>
			<persName><forename type="first">M</forename><surname>Danilevsky</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2010.00711</idno>
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
