<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Argumentative Dialogue As Basis For Human-AI Collaboration</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Alexander</forename><surname>Berman</surname></persName>
							<email>alexander.berman@gu.se</email>
							<affiliation key="aff0">
								<orgName type="department" key="dep1">Dept. of Philosophy</orgName>
								<orgName type="department" key="dep2">Linguistics and Theory of Science</orgName>
								<orgName type="institution">University of Gothenburg</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">Argumentative Dialogue As Basis For Human-AI Collaboration</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">BE59EB6ED2B67C23E9922EC54069B741</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T17:12+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>human-AI collaboration</term>
					<term>hybrid human-AI intelligence</term>
					<term>conversational explainability</term>
					<term>argumentation theory</term>
					<term>explainable AI</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Argumentation, by which we here mean the ability to give reasons or arguments for a claim, plays a central role in society generally and in collaborative decision-making specifically. However, the role of argumentation in human-AI collaboration and AI-assisted decision-making has received limited attention, despite the widespread interest in "explainable" AI. This paper aims to bridge this gap. First, it is shown that many kinds of AI models are not argumentative in the sense that they do not enable human-AI interfaces to provide reasons or arguments for AI predictions. Second, it is shown that some interpretable AI models encode a knowledge structure that can be harvested for the purpose of supporting argumentative human-AI interaction. Third, a method for extracting such structures from an interpretable model is outlined. Finally, a prototype supporting argumentative dialogue between AI and human user is presented.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Argumentation plays a crucial role in society generally and in collaborative decision-making more specificially <ref type="bibr" target="#b0">[1]</ref>. By requesting and providing support for claims, we justify our beliefs and actions, and evaluate claims made by others. In the context of artificial intelligence (AI) based on machine learning (ML), it is natural to treat predictions made by AI systems as claims. For example, if a statistical model predicts that a certain individual is introverted, we can intuitively understand this prediction as a claim. It is then natural to also ask whether the model can provide arguments for its claim? In current discourses revolving around AI, this question is typically approached as a matter of explainability or interpretability (see e.g. <ref type="bibr" target="#b1">[2]</ref>). A distinction is often made between black-box models whose predictions and inner workings can only be explained by means of inherently unreliable explanation methods <ref type="bibr" target="#b2">[3]</ref>, and interpretable models whose logic can in principle be understood by humans <ref type="bibr" target="#b3">[4]</ref>. However, from the perspective of human-AI collaboration, the notions of explainability and interpretability are not necessarily crucial in and of themselves. In this paper, we instead hypothesize that human-AI collaboration yield more value when the AI systems can engage in argumentation. The main aims of the paper are to briefly discuss the conditions that support argumentative dialogue between a machine learning model and human, and to demonstrate how such a capacity can be conceived theoretically and implemented technically.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Theoretical Framework</head><p>The present work applies Toulmin's <ref type="bibr" target="#b4">[5]</ref> theory of argumentation to ML-based AI. <ref type="foot" target="#foot_0">1</ref> According to this theory, argumentation is an interactive process through which presented claims are challenged and backed. Specifically, a claim (e.g. that Sam is introverted) can be backed with data (e.g. that Sam doesn't like danceable music). Data support claims by highlighting specific facts or circumstances. Furthermore, the backing of claims by data (often implicitly) rests on warrants (e.g. that people that like non-danceable music are generally introverted). While claims and data are specific, warrants are general; their argumentative function is to bridge data and claims by means of e.g. taxonomy (that instances of one category are always instances of another category) or statistics (that instances of one category also tend to be instances of another category). Warrants support conclusions with varying degrees of force, signalled linguistically with a qualifier (e.g., statistically, people that like non-danceable music are more introverted, and Sam doesn't like danceable music; so, presumably Sam is introverted).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Argumentation Affordances</head><p>The extent to which ML-based AI systems support argumentation differs across different kinds of ML models. Although a comprehensive assessment of this matter is beyond the scope of this paper, some brief remarks can be made. First, it can be noted that black-box models such as deep neural networks and random forests do not afford argumentation in any obvious manner. While claims (e.g. classifiying a person as introverted) can straightforwardly be qualified in terms of confidence (e.g. that the prediction is associated with a probability of 67%), data and warrants cannot readily be identified due to the complex inner workings of these models. To some extent, feature-importance based explanation methods such as LIME <ref type="bibr" target="#b6">[7]</ref> and SHAP <ref type="bibr" target="#b7">[8]</ref> can be conceived to identify data, since they highlight the features that were the most important for a particular prediction. For example, if a neural network predicts that a person is introverted based on the person's music preferences (measured as numerical values for features such as danceability and loudness), LIME may highlight danceability as the most important feature for the prediction at hand. From this information, one can construct the datum "On a scale from 0 to 1, Sam's preference for danceable music is 0.34" as a claim-backing datum. But what kind of warrant supports the conclusion from datum to claim? Has the model learned that people with a preference for danceable music of exactly 0.34 are generally introverted? Or has it learned something more general, e.g. that introverts prefer music with a danceability value below a certain threshold? These questions cannot be answered by methods such as LIME or SHAP. In reality, a black box may combine preference for danceable music with preference for loud music and other feature values in non-linear and complicated ways that may be difficult or impossible to express in words. In argumentative terms, no warrant can be generated.</p><p>For a more interpretable option, we can consider linear additive models such as logistic regression. In contrast to black boxes, these models force features to affect output independently of each other, without any interactions. <ref type="foot" target="#foot_1">2</ref> Furthermore, features affect output monotonically; for example, a stronger preference for danceable music always increases the predicted probability that the person is extraverted. Due to these formal properties, warrants such as "statistically, people that like danceable music are more extraverted" will faithfully reflect the actual knowledge learned by the model. Below, we will formally show how to extract data and warrants for claims obtained from linear additive models.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Extracting Arguments From Linear Additive Models</head><p>We assume a linear additive model on the form</p><formula xml:id="formula_0">y ˆ= β 0 + m ∑ i=1 β i X i</formula><p>where β 0 is the intercept/bias, β i are the coefficients (both of which are learned when fitting the model to training data), X is the instance (feature values), i denotes feature (e.g. 1 for energy, 2 for danceability, etc.), and y ˆis the output (predicted value). We also assume that output and features are standardized continuous variables so that 0 corresponds to mean. Sticking to the example domain above, we say that if the prediction is positive (y ˆ&gt; 0), it is claimed that the person described by X is extraverted, or else introverted. Data supporting a positive claim can then be extracted by listing features with a positive value and a positive coefficient, and features with a negative value and a negative coefficient, since in both of these cases the feature contributes to a positive prediction. Conversely, a negative claim is supported by features with a positive value and a negative coefficient, and features with a negative value and a positive coefficient. A datum can be conveyed linguistically with reference to the feature and its polarity; for example, X 2 &gt; 0 can be expressed as "The person likes danceable music".</p><p>As warrants, we construct support-relations between the polarity of a coefficient (e.g. β 2 &lt; 0) and the polarity of the prediction (e.g. y ˆ&lt; 0). This can be expressed linguistically as "Statistically, people that like non-danceable music are more likely to be introverted". The force of each combination of datum and warrant can be defined as the magnitude of the respective addend</p><formula xml:id="formula_1">(|β i X i |).</formula><p>In what follows, we will see how this extraction procedure can form the basis for an argumentative AI communicator.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Prototype</head><p>We briefly describe MindTone 3 , a browser-based game featuring argumentative communication between AI and human. The task of the game is to estimate whether persons are extraverted or introverted based on their music preferences. In each round of the game, the player is shown specific tracks that a person has listened to frequently as well as audio statistics of music heard by the person, such as loudness and energy (see figure <ref type="figure" target="#fig_1">1</ref>). The player is assisted by a chatbot that predicts whether the person is extraverted or introverted using a logistic regression model trained on audio features and personality traits <ref type="bibr" target="#b8">[9]</ref>. The dialogue manager is implemented in a rule-based information-state update approach <ref type="bibr" target="#b9">[10]</ref> inspired by conversation-oriented semantics <ref type="bibr" target="#b10">[11]</ref>. Specifically, when a session starts, the system makes a claim, e.g. "I think this person is introverted" or "If I had to guess, I'd say that this person is extraverted" (where the qualification reflects the model's confidence). When the user challenges a claim, the system backs it up with its strongest datum, e.g. "The person likes high-energy music." When the user indicates not understanding how a datum supports a claim, the system provides a warrant, e.g. "Statistically, people that like high-energy music are more likely to be introverted". Importantly, such warrants faithfully reflect the actual reasoning process of the model and are not post-hoc approximations. The communicative capabilities are examplified by the following dialogue between user (U) and assistant (A):</p><p>A:</p><p>I think this person is introverted.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>U:</head><p>Why?</p><p>A:</p><p>The person likes high-energy music. Music heard by the person has a higher average score for energy than music in general.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>U:</head><p>OK, and does the fact that the person likes non-danceable music support the assessment that the person is introverted? A:</p><p>Yes.</p><p>For natural-language understanding, the system uses GPT-4 <ref type="bibr" target="#b11">[12]</ref> to parse user utterances into a formal language for dialogue moves, while templates are used for natural-language generation. Note that while the prototype supports a particular domain (estimation of personality trait from music preferences), both the method for extracting argument structure and the dialogue system are domain-independent and can therefore be applied to any domain of choice.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Related Work</head><p>The present work can be situated in the context of "conversational explainable AI", i.e. formalization, implementation and evaluation of systems that can explain AI predictions in a natural-language dialogue between system and human (see e.g. <ref type="bibr" target="#b12">[13,</ref><ref type="bibr" target="#b13">14,</ref><ref type="bibr" target="#b14">15,</ref><ref type="bibr" target="#b15">16,</ref><ref type="bibr" target="#b16">17,</ref><ref type="bibr" target="#b17">18,</ref><ref type="bibr" target="#b18">19]</ref>). Typically, previous approaches do not support argumentation of the kind discussed here. For example, the system TalkToModel <ref type="bibr" target="#b17">[18]</ref> uses explanation methods such as LIME and SHAP to explain specific predictions by listing features deemed important; for the current domain, it would amount to a phrase such as "the top 2 most important features are (1) energy (2) danceability". It can be noted that such an explanation provides neither data (whether the person likes how-or low-energy music, etc.) or warrants (how preferences for different aspects of music correlate with extraversion).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>PERSON #4 09:12 SCORE: 1</head><p>This person has listened to...</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>The B Foundation Dirty Girls &gt;10 times</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Lower Definition</head><p>To Satellite</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>&gt;10 times</head><p>The B Foundation Bloo Bucket</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>&gt;10 times</head><p>The B Foundation Your Eyes (Realggae)</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>&gt;10 times</head><p>The B Foundation Be Alright &gt;10 times ...and more generally:</p><p>The bars show properties of the music that the person has listened to, compared to the population at large (dotted blue lines).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Assistant</head><p>I think this person is introverted.</p><p>You OK, why?</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Assistant</head><p>The person likes high-energy music.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>You</head><p>Hmm I don't understand Assistant Statistically, people that like high-energy music are more likely to be introverted. As for dialogue systems with argumentative capabilities, Breitholtz <ref type="bibr" target="#b22">[23]</ref> presented a formal account of how claims can be motivated with enthymemes, which, in Tolmin's framework, corresponds to backing claims with data; Maraev et al. <ref type="bibr" target="#b19">[20]</ref> later implemented a working prototype on this basis. In contrast to previous work, the approach presented in this paper enables arguments for ML predictions, in the form of both data and warrants, to be communicated by the system.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Assessment</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.">Discussion and Future Work</head><p>We have argued that black-box AI systems cannot generate warrants to support their predictions, even in tandem with popular explainability methods such as LIME or SHAP. However, a human user may still identify or produce a warrant to fill the gap. In fact, implicit premises are ubiquitous in human communication and rarely need to be made explicit. <ref type="foot" target="#foot_3">4</ref> Someone says "It's cold in here so let's close the window" and you immediately understand the warrant: closing the window is likely to increase the indoor temperature (assuming that it's colder outside than inside). Since the listener can identify a warrant that makes the speaker's utterance comprehensible, no warrant needs to be verbalized. From this perspective, a lack of warrant-production in human-AI collaboration does not necessarily need to be a problem. But if the purpose of the AI is to enable improved human decision-making <ref type="bibr" target="#b23">[24]</ref>, one cannot assume that an AI always "reasons" in similar ways as humans. Arguably, potential differences in reasoning between AI and human is precisely the reason why explanations and arguments are needed.</p><p>As shown in previous sections, at least some interpretable models afford argumentation. However, to our best knowledge there is no empirical data that supports the hypothesis that argumentation of the kind discussed here benefits human-AI collaboration. (A survey study found that decision-makers prefer interactive explanations in the form of natural language dialogue <ref type="bibr" target="#b24">[25]</ref>, but did not specifically investigate argumentation.) In fact, previous work suggests that explanations can cause human over-reliance on AI <ref type="bibr" target="#b25">[26]</ref> or have no effect on accuracy <ref type="bibr" target="#b26">[27]</ref>. However, as far as we can tell, previously evaluated human-AI interactions have not involved argumentative AI systems. We propose two mechanisms through which argumentation might benefit hybrid human-machine decision-making. First, warrants may enable users to assess whether claims are supported by reasonable generalizations. If the system argues that liking music with high energy makes it more likely that one is introverted, and the user finds this correlation generally questionable, then the user can take this into account when assessing the reliablity of the claim that the generalization is intended to support. Second, warrants can potentially make it easier to combine an AI's assessment with the user's own judgement about the case at hand. If the system supports its claim with a statistical generalization, the user can assess to what extent the generalization seems relevant for the case at hand. Sure, you may reason, people that like music with high energy may be more introverted in general, but in this case you know exactly what kind of high-energy music the person listens to, and you don't associate this music with introversion. To the extent that users successfully assess the relevance of the system's generalizations, decision-making accuracy can be improved compared to a scenario without argumentation. In future work, it would be useful to empirically study how an argumentative AI communicator affects human decision patterns in comparison with a non-argumentative interface, and thereby generate data to either support or contradict our claim that argumentation is crucial not only in human communication, but also in communication between humans and AI.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head></head><label></label><figDesc>that like high-energy music are more likely to be introverted.U:OK, I see. Why do you think this person likes music with high energy?A:</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Screenshot of prototype. To the left, a list of tracks frequently listened to by the person whose personality is currently being assessed, followed by a bar graph visualizing audio statistics of music heard by the person. To the right, a chat window for player-AI interaction.</figDesc></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">For alternative theories of argumentation, see e.g.<ref type="bibr" target="#b5">[6]</ref>.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">We here disregard the possibility to use combinations of variables as features.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">Live demo: https://github.com/alex-berman/argumentative-explainability</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_3">This phenomenon has previously been discussed in terms of e.g. conversational implicature<ref type="bibr" target="#b20">[21]</ref>, presuppositions<ref type="bibr" target="#b21">[22]</ref>, and enthymemes<ref type="bibr" target="#b22">[23]</ref>.</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>This work was supported by the Swedish Research Council (VR) grant 2014-39 for the establishment of the Centre for Linguistic Theory and Studies in Probability (CLASP) at the University of Gothenburg.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title level="m" type="main">The enigma of reason</title>
		<author>
			<persName><forename type="first">H</forename><surname>Mercier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Sperber</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2017">2017</date>
			<publisher>Harvard University Press</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI</title>
		<author>
			<persName><forename type="first">Barredo</forename><surname>Arrieta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Díaz-Rodríguez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename></persName>
		</author>
		<author>
			<persName><forename type="first">Del</forename><surname>Ser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Bennetot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Tabik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Barbado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename></persName>
		</author>
		<ptr target="https://www.sciencedirect.com/science/article/pii/S1566253519308103" />
	</analytic>
	<monogr>
		<title level="j">Information Fusion</title>
		<imprint>
			<biblScope unit="volume">58</biblScope>
			<biblScope unit="page" from="82" to="115" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">To trust or not to trust an explanation: using LEAF to evaluate local linear XAI methods</title>
		<author>
			<persName><forename type="first">E</forename><surname>Amparore</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Perotti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Bajardi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">PeerJ Computer Science</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="page">e479</biblScope>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Interpretable machine learning: Fundamental principles and 10 grand challenges</title>
		<author>
			<persName><forename type="first">C</forename><surname>Rudin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Semenova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Zhong</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Statistic Surveys</title>
		<imprint>
			<biblScope unit="volume">16</biblScope>
			<biblScope unit="page" from="1" to="85" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<title level="m" type="main">The uses of argument</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">E</forename><surname>Toulmin</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2003">2003</date>
			<publisher>Cambridge university press</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<title level="m" type="main">Fundamentals of argumentation theory: A handbook of historical backgrounds and contemporary developments</title>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">H</forename><surname>Van Eemeren</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Grootendorst</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">H</forename><surname>Johnson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Plantin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">A</forename><surname>Willard</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2013">2013</date>
			<publisher>Routledge</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Why Should I Trust You?&quot;: Explaining the Predictions of Any Classifier</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">T</forename><surname>Ribeiro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Singh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Guestrin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</title>
				<meeting>the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="1135" to="1144" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">A unified approach to interpreting model predictions</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Lundberg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">I</forename><surname>Lee</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 31st International Conference on Neural Information Processing Systems. NIPS&apos;17</title>
				<meeting>the 31st International Conference on Neural Information Processing Systems. NIPS&apos;17<address><addrLine>Red Hook, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Curran Associates Inc</publisher>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page" from="4768" to="4777" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Personality Correlates of Music Audio Preferences for Modelling Music Listeners</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">B</forename><surname>Melchiorre</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Schedl</surname></persName>
		</author>
		<idno type="DOI">10.1145/3340631.3394874</idno>
		<ptr target="https://doi-org.ezproxy.ub.gu.se/10.1145/3340631.3394874" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 28th ACM Conference on User Modeling, Adaptation and Personalization</title>
				<meeting>the 28th ACM Conference on User Modeling, Adaptation and Personalization<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="313" to="317" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">Issue-Based Dialogue Management</title>
		<author>
			<persName><forename type="first">S</forename><surname>Larsson</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<title level="m" type="main">The interactive stance: Meaning for conversation</title>
		<author>
			<persName><forename type="first">J</forename><surname>Ginzburg</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2012">2012</date>
			<publisher>Oxford University Press</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title/>
		<author>
			<persName><forename type="first">Achiam</forename><forename type="middle">J</forename><surname>Openai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Adler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Agarwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Ahmad</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Akkaya</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note type="report_type">GPT-4 Technical Report</note>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">iSee: Intelligent Sharing of Explanation Experience by Users for Users</title>
		<author>
			<persName><forename type="first">A</forename><surname>Wijekoon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Wiratunga</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Palihawadana</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Nkisi-Orji</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Corsar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Martin</surname></persName>
		</author>
		<idno type="DOI">10.1145/3581754.3584137</idno>
		<ptr target="https://doi.org/10.1145/3581754.3584137" />
	</analytic>
	<monogr>
		<title level="m">Companion Proceedings of the 28th International Conference on Intelligent User Interfaces. IUI &apos;23 Companion</title>
				<meeting><address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="79" to="82" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Glass-Box: Explaining AI Decisions With Counterfactual Statements Through Conversation With a Voice-enabled Virtual Assistant</title>
		<author>
			<persName><forename type="first">K</forename><surname>Sokol</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">A</forename><surname>Flach</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IJCAI</title>
		<imprint>
			<biblScope unit="page" from="5868" to="5870" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Explaining predictions with enthymematic counterfactuals</title>
		<author>
			<persName><forename type="first">A</forename><surname>Berman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Breitholtz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Howes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">P</forename><surname>Bernardy</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 1st Workshop on Bias, Ethical AI, Explainability and the role of Logic and Logic Programming</title>
				<meeting>the 1st Workshop on Bias, Ethical AI, Explainability and the role of Logic and Logic Programming</meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="volume">22</biblScope>
			<biblScope unit="page" from="95" to="100" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Explainable AI through Rule-based Interactive Conversation</title>
		<author>
			<persName><forename type="first">C</forename><surname>Werner</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Workshop Proceedings of the EDBT/ICDT 2020 Joint Conference</title>
				<meeting><address><addrLine>Copenhagen, Denmark)</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2020-04-02">March 30-April 2, 2020. 2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">What Would You Ask the Machine Learning Model? Identification of User Needs for Model Explanations Based on Human-Model Conversations</title>
		<author>
			<persName><forename type="first">M</forename><surname>Kuźba</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Biecek</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ECML PKDD 2020 Workshops</title>
				<editor>
			<persName><forename type="first">I</forename><surname>Koprinska</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Kamp</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Appice</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Loglisci</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Antonie</forename><forename type="middle">L</forename><surname>Zimmermann</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename></persName>
		</editor>
		<meeting><address><addrLine>Cham</addrLine></address></meeting>
		<imprint>
			<publisher>Springer International Publishing</publisher>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="447" to="459" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Explaining machine learning models with interactive natural language conversations using TalkToModel</title>
		<author>
			<persName><forename type="first">D</forename><surname>Slack</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Krishna</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Lakkaraju</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Singh</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Nature Machine Intelligence</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="issue">8</biblScope>
			<biblScope unit="page" from="873" to="883" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<title level="m" type="main">Mediators: Conversational Agents Explaining NLP Model Behavior</title>
		<author>
			<persName><forename type="first">N</forename><surname>Feldhus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">M</forename><surname>Ravichandran</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Möller</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Why should I turn left? Towards active explainability for spoken dialogue systems</title>
		<author>
			<persName><forename type="first">V</forename><surname>Maraev</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Breitholtz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Howes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">P</forename><surname>Bernardy</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Reasoning and Interaction Conference</title>
				<meeting>the Reasoning and Interaction Conference<address><addrLine>ReInAct</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2021">2021. 2021</date>
			<biblScope unit="page" from="58" to="64" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Logic and conversation</title>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">P</forename><surname>Grice</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Speech acts</title>
				<imprint>
			<publisher>Brill</publisher>
			<date type="published" when="1975">1975</date>
			<biblScope unit="page" from="41" to="58" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Scorekeeping in a language game</title>
		<author>
			<persName><forename type="first">D</forename><surname>Lewis</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of philosophical logic</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="page" from="339" to="359" />
			<date type="published" when="1979">1979</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<monogr>
		<title level="m" type="main">Enthymemes and Topoi in Dialogue: The Use of Common Sense Reasoning in Conversation</title>
		<author>
			<persName><forename type="first">E</forename><surname>Breitholtz</surname></persName>
		</author>
		<ptr target="https://brill.com/view/title/58383" />
		<imprint>
			<date type="published" when="2020">2020</date>
			<publisher>Brill</publisher>
			<pubPlace>Leiden, The Netherlands</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Directions in hybrid intelligence: complementing AI systems with human intelligence</title>
		<author>
			<persName><forename type="first">E</forename><surname>Kamar</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence</title>
				<meeting>the Twenty-Fifth International Joint Conference on Artificial Intelligence</meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="4070" to="4073" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<monogr>
		<title level="m" type="main">Rethinking explainability as a dialogue: A practitioner&apos;s perspective</title>
		<author>
			<persName><forename type="first">H</forename><surname>Lakkaraju</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Slack</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Tan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Singh</surname></persName>
		</author>
		<idno>arXiv:220201875. 2022</idno>
		<imprint/>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">When do XAI methods work? A cost-benefit approach to human-AI collaboration</title>
		<author>
			<persName><forename type="first">H</forename><surname>Vasconcelos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Jörke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Grunde-Mclaughlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Krishna</forename><forename type="middle">R</forename><surname>Gerstenberg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Bernstein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">S</forename></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CHI Workshop on Trust and Reliance in AI-Human Teams</title>
				<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="1" to="15" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Does explainable artificial intelligence improve human decision-making?</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Alufaisan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">R</forename><surname>Marusich</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">Z</forename><surname>Bakdash</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kantarcioglu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the AAAI Conference on Artificial Intelligence</title>
				<meeting>the AAAI Conference on Artificial Intelligence</meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="volume">35</biblScope>
			<biblScope unit="page" from="6618" to="6626" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
