<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Teaching AI to Explain its Decisions Can Affect Class Balance</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Kevin</forename><surname>Volkert</surname></persName>
							<email>volkert@stud.uni-bamberg.de</email>
							<affiliation key="aff0">
								<orgName type="laboratory">Cognitive Systems Group</orgName>
								<orgName type="institution">University of Bamberg</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">Teaching AI to Explain its Decisions Can Affect Class Balance</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">DF2E337B27A34D1D63DDDC07C665F028</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T05:36+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Explainable AI</term>
					<term>Supervised classification</term>
					<term>Machine learning</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>TED has been proposed as a simple framework for using any supervised classification approach to create Explainable AI models <ref type="bibr" target="#b2">[3]</ref>. It requires manual annotation of training data with explanations. The authors of TED report that it yields results as good or better than conventional supervised learning. This paper shows that one possible reason for this is that the extension of training data can introduce bias by changing the class distribution. Experiments on the Iris dataset and a synthetic dataset show that TED performs significantly worse than initially reported when using class weights during training to offset class imbalance in training data.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Explainable Artificial Intelligence as a field of research is becoming more popular. While black-box approaches like deep learning have been very successful, good performance alone is not always enough. As AI becomes more widely used it also becomes more important to understand what it is actually doing. A decision may be good if it is correct, but it is better if it is correct and understandable.</p><p>There are good, old-fashioned AI approaches, like Inductive Logic Programming (ILP), which are inherently white-box and thus interpretable. This can, for example, be used to generate human-readable representations of the decision making of a system. Dare2Del <ref type="bibr" target="#b5">[6]</ref> is an ILP-based companion system which learns rules that are directly interpretable and can be presented to end users as explanations for suggested actions.</p><p>LIME <ref type="bibr" target="#b4">[5]</ref> is a technique proposed with the goal of adding explainability to arbitrary classifiers. This has become a popular idea (Google Scholar lists over 2,400 citations at the time of this writing). For example, this can be used to gain insight into black-box approaches like deep neural networks.</p><p>Another approach has a similar goal, but a very different design: TED (Teaching Explanations for Decisions) is a simple, high-level framework . The basic idea is to use any supervised classification approach to learn not only labels, but also explanations. This approach depends heavily on the quality of the training data. The authors explicitly state the importance of good explanations provided by domain experts <ref type="bibr" target="#b2">[3]</ref>.</p><p>Its high-level, semi-manual nature makes this approach very flexible. The explanations are part of the training data, therefore they can be adapted to a variety of domains and target groups. The authors argue that the requirement to manually create training data with explanations can also help reduce bias, since adding a biased example would be hard to justify <ref type="bibr" target="#b2">[3]</ref>.</p><p>However, bias is not always obvious when considering single examples. Class imbalance is one type of bias that can occur in training data. When classes are not equally distributed, trained classifiers may inadvertently favor more frequent classes when making predictions.</p><p>This paper investigates this problem in the context of TED. In section 2 the design of the TED framework is described and the original evaluation by <ref type="bibr" target="#b2">[3]</ref> is summarized. Section 3 shortly explains how class imbalance can become an issue when using TED. Section 4 presents experiments that make the issue apparent. Finally, section 5 provides a conclusion.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Related Work</head><p>TED, unlike traditional classification approaches, learns class labels and explanations from examples <ref type="bibr" target="#b2">[3]</ref>. Any supervised classification approach can be used with TED. The training data consists of: X a set of features Y a class label for X E an explanation for Y</p><p>The form and content of explanations are not specified and may contain arbitrary values. The only constraint is that they must be enumerable.</p><p>Because explanations are meant to be provided as part of the training data, the application of TED is extremely simple. Instead of learning only labels from examples (X → Y ), TED learns labels and explanations (X → Y E). The Cartesian product instantiation described in <ref type="bibr" target="#b2">[3]</ref> simply encodes labels and explanations as Y E before training. Then the trained model predicts Y E which is decoded into a separate label and explanation.</p><p>The Cartesian product instantiation of TED has been evaluated on two datasets. The Tic-Tac-Toe example is based on the game of tic-tac-toe. The legal non-terminal board positions are labeled with the optimal move to make and an explanation based on a small set of rules for playing the game. For any move, the explanation can be (in order of preference):</p><p>1. win (if the game is won with the move) 2. block (if the move prevents the other player from winning) 3. threat (if the game could be won with the following move) 4. empty (otherwise).</p><p>This results in a dataset of 4,520 examples. There are 36 possible combinations of label and explanation, since each rule can apply to 9 different moves.</p><p>The Loan Repayment example contains 10,000 applications for credit with two labels (good or bad payment performance). Explanations are not included in the original dataset. In <ref type="bibr" target="#b2">[3]</ref> explanations were generated by applying a small rule set on the data. Two rules consisting of three literals each were used to classify the data. The labels of examples that were not consistent with these rules were changed. For positive examples, the rule itself was used as the explanation.</p><p>For negative examples, the explanation was determined by the rule whose first literal matched, combined with which of the other literals that did not match. This resulted in 8 explanations.</p><p>Table <ref type="table" target="#tab_0">1</ref> shows the accuracy in the Tic-Tac-Toe and Loan Repayment examples reported in <ref type="bibr" target="#b2">[3]</ref>. The results appear promising, though they are unexpected. The performance for label prediction is reported to be as good or better than that of the equivalent models without the TED framework. In the evaluation of the Tic-Tac-Toe example in <ref type="bibr" target="#b2">[3]</ref> the authors point out that, "[g]iven the increase in number of classes, one might expect the accuracy to decrease." They also admit that the accuracy in the Loan Repayment example "may be artificially high due to the data generation method" <ref type="bibr" target="#b2">[3]</ref>. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Potential Problems</head><p>Adding explanations to the training data and learning the explanations in addition to the labels results in a larger output space. Each class may have multiple possible explanations. The authors of TED describe this as an implicit 2-level hierarchy in the data that could potentially be exploited <ref type="bibr" target="#b2">[3]</ref>.</p><p>In the Cartesian product instantiation the result is simply a larger number of classes and therefore a smaller number of examples per class compared to the same dataset without added explanations. This means that the underlying classifier in a simple TED instance has to learn more from less. Intuitively there should be a decrease in performance.</p><p>The addition of explanations to an existing dataset can change the effective class balance. Even if the original class labels are left untouched, it is possible to have a different number of explanations for each class. In that case the combined Y E learned by the classifier will not have the same distribution of classes as the original dataset without explanations. This can introduce or change existing bias in the training data.</p><p>It seems plausible that the improved performance which TED shows (in comparison to an equivalent classifier trained without explanations) is at least partially due to this altered class balance. Section 4 attempts to test this idea.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Experiments</head><p>In this section TED is evaluated on two additional datasets: the well-known Iris flower dataset <ref type="bibr" target="#b1">[2]</ref> (included in scikit-learn) and the Proactive Retention dataset (included in the TED source code). The performance of different classifiers is compared in different configurations:</p><p>with/without TED (i. e. with/without explanations) with/without balanced class weights for the underlying classifiers TED is included in IBM AIX360 <ref type="bibr" target="#b0">[1]</ref> which is implemented using scikitlearn <ref type="bibr" target="#b3">[4]</ref>. scikit-learn provides a parameter class weight which can be used to balance classes. When this parameter is set to "balanced", each sample is given a weight that is inversely proportional to the class frequency in the training data. This prevents the classifier from being biased toward a certain class simply because it has the most examples.</p><p>Three classifiers were chosen for these experiments: a Support Vector Machine (SVM) classifier, a Decision Tree classifier, and a DummyClassifier. The DummyClassifier is used with the strategy parameter set to "stratified". It makes predictions by choosing a label randomly, considering the class distribution of the training data. This makes bias through class imbalance obvious, since the DummyClassifier does not have any other biases.</p><p>All experiments in this section were executed using AIX360 version 0.2 and scikit-learn version 0.21.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Iris Dataset</head><p>The Iris flower dataset consists of 150 examples of 3 types of flowers (iris setosa, iris versicolor, iris viginica). The 3 classes are equally distributed, i. e. there are 50 examples per class. There are 4 real-valued features for petal/sepal length/width.</p><p>Proper manual annotation would require domain knowledge. Since this requirement could not be fulfilled for this paper, a decision tree (standard parameters, no maximum depth) was trained on the full dataset. This overfitted decision tree perfectly predicts every example in the dataset. The leaves (identified by the node number) of this tree were used as explanations (see figure <ref type="figure">1</ref>).</p><p>This results in a total of 9 explanations: With balanced class weights, the SVM and decision tree yield noticably different results (see table <ref type="table" target="#tab_3">4</ref>). Accuracy without explanations is the same. With explanations the prediction performance for both labels and explanations decreases significantly. Label accuracy drops by 12-30 percentage points, and the prediction of explanations is only about two-thirds as good. The Iris dataset is admittedly quite small. The next section repeats the same experiments on a larger, synthetic dataset.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Proactive Retention Dataset</head><p>The TED source code in the IBM AIX360 repository (https://github.com/IBM /AIX360/) includes a synthetic dataset that is used in the Proactive Retention tutorial. The intended use case is finding employees who are at risk of leaving a company.</p><p>The data is generated according to distribution functions. There are 8 features:</p><p>-Position in the company -Organization -Potential -Rating -Rating slope -Salary competitiveness -Tenure (number of months employed) -Position tenure (number of months at current position) Each feature has a non-uniform distribution from which the values are randomly sampled.</p><p>There are two labels, Yes and No, expressing whether proactive retention is necessary for a given employee or not. Explanations are generated by applying 25 rules based on the features. If any of these rules applies to an example, it is labeled Yes with the rule number as explanation. If no rule applies, the example is labeled No. The dataset in the TED source code contains 10,000 examples of which 33.8% are labeled Yes.</p><p>The SVM classifier was trained with default parameters. The decision tree was trained with a maximum depth of 5.</p><p>The results are similar to the ones in section 4.1 with regard to class balance (see table <ref type="table" target="#tab_4">5</ref>), but more in line with the expectation stated in section 3. Without balanced class weights, the TED-augmented versions of the classifiers perform worse than their counterparts without explanations by about 6 percentage points. The DummyClassifier improves very slightly in label prediction. The accuracy for explanations is at roughly the same level as for labels. With balanced class weights, however, both the SVM and the decision tree again perform much worse. Label prediction performance decreases by 18-40 percentage points. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Conclusion</head><p>This paper performs additional experiments based on the work by <ref type="bibr" target="#b2">[3]</ref> and attempts to investigate the unexpected performance gains that arise from making a machine learning problem harder.</p><p>The experiments show that class imbalance can be a problem. If training data, especially small datasets, is not carefully created or extended, the distribution of classes can become an issue. The simple Cartesian product instantiation of TED can be very sensitive to training data. Especially when augmenting existing data with explanations to make it compatible with TED, it can be easy to overlook the introduced or changed bias.</p><p>Using balanced class weights to counteract the bias in the class distribution of the training data results in a large decrease in performance in both experiments (see section 4). This shows that at least some of the accuracy increases in label prediction that TED achieves compared to equivalent classifiers trained without explanations may be based on the aforementioned bias.</p><p>Since manual work on datasets is time-consuming, methods that automatically generate explanations for existing datasets or create synthetic datasets from scratch are very appealing. Such methods have been used in the original work by <ref type="bibr" target="#b2">[3]</ref> as well as in this paper. It would, however, be more interesting to see how TED performs on a more realistic, representative dataset.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 .</head><label>1</label><figDesc>Accuracy for predicting Y and E in Tic-Tac-Toe and Loan Repayment</figDesc><table><row><cell>Training input</cell><cell></cell><cell cols="2">Accuracy (%)</cell><cell></cell></row><row><cell></cell><cell cols="2">Tic-Tac-Toe</cell><cell cols="2">Loan Repayment</cell></row><row><cell></cell><cell>Y</cell><cell>E</cell><cell>Y</cell><cell>E</cell></row><row><cell>X, Y</cell><cell>96.5</cell><cell></cell><cell>99.2</cell><cell></cell></row><row><cell>X, Y, E</cell><cell>97.4</cell><cell>94.3</cell><cell>99.6</cell><cell>99.4</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 .</head><label>2</label><figDesc>Label/explanation distribution for the complete annotated Iris dataset</figDesc><table><row><cell>Label</cell><cell cols="3">Explanation Count Percentage (%)</cell></row><row><cell>Iris setosa</cell><cell>1</cell><cell>50</cell><cell>33.33</cell></row><row><cell cols="2">Iris versicolor 5</cell><cell>47</cell><cell>31.33</cell></row><row><cell></cell><cell>10</cell><cell>2</cell><cell>1.33</cell></row><row><cell></cell><cell>15</cell><cell>1</cell><cell>0.67</cell></row><row><cell cols="2">Iris virginica 6</cell><cell>1</cell><cell>0.67</cell></row><row><cell></cell><cell>8</cell><cell>3</cell><cell>2.00</cell></row><row><cell></cell><cell>11</cell><cell>1</cell><cell>0.67</cell></row><row><cell></cell><cell>14</cell><cell>2</cell><cell>1.33</cell></row><row><cell></cell><cell>16</cell><cell>43</cell><cell>28.67</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3 .</head><label>3</label><figDesc>Accuracy for classifiers on the Iris dataset without class weights</figDesc><table><row><cell>Classifier</cell><cell cols="3">Training input Accuracy (%)</cell></row><row><cell></cell><cell></cell><cell>Y</cell><cell>E</cell></row><row><cell>SVM</cell><cell>X, Y</cell><cell>98</cell></row><row><cell></cell><cell>X, Y, E</cell><cell>98</cell><cell>92</cell></row><row><cell>Decision tree</cell><cell>X, Y</cell><cell>92</cell></row><row><cell></cell><cell>X, Y, E</cell><cell>96</cell><cell>92</cell></row><row><cell cols="2">DummyClassifier X, Y</cell><cell>34</cell></row><row><cell></cell><cell>X, Y, E</cell><cell>46</cell><cell>40</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4 .</head><label>4</label><figDesc>Accuracy for classifiers on the Iris dataset with balanced class weights</figDesc><table><row><cell>Classifier</cell><cell cols="3">Training input Accuracy (%)</cell></row><row><cell></cell><cell></cell><cell>Y</cell><cell>E</cell></row><row><cell>SVM</cell><cell>X, Y</cell><cell>98</cell></row><row><cell></cell><cell>X, Y, E</cell><cell>86</cell><cell>64</cell></row><row><cell cols="2">Decision tree X, Y</cell><cell>92</cell></row><row><cell></cell><cell>X, Y, E</cell><cell>62</cell><cell>60</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 5 .</head><label>5</label><figDesc>Accuracy for classifiers on the Proactive Retention dataset without class weights</figDesc><table><row><cell>Classifier</cell><cell cols="3">Training input Accuracy (%)</cell></row><row><cell></cell><cell></cell><cell>Y</cell><cell>E</cell></row><row><cell>SVM</cell><cell>X, Y</cell><cell>86.15</cell></row><row><cell></cell><cell>X, Y, E</cell><cell>80.5</cell><cell>77.1</cell></row><row><cell>Decision tree</cell><cell>X, Y</cell><cell>91.2</cell></row><row><cell></cell><cell>X, Y, E</cell><cell cols="2">85.55 81.25</cell></row><row><cell cols="2">DummyClassifier X, Y</cell><cell>54.3</cell></row><row><cell></cell><cell>X, Y, E</cell><cell cols="2">54.75 43.85</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 6 .</head><label>6</label><figDesc>Accuracy for classifiers on the Proactive Retention dataset with balanced class weights</figDesc><table><row><cell>Classifier</cell><cell cols="3">Training input Accuracy (%)</cell></row><row><cell></cell><cell></cell><cell>Y</cell><cell>E</cell></row><row><cell>SVM</cell><cell>X, Y</cell><cell>86.5</cell></row><row><cell></cell><cell>X, Y, E</cell><cell>68.75</cell><cell>59.9</cell></row><row><cell cols="2">Decision tree X, Y</cell><cell>92.25</cell></row><row><cell></cell><cell>X, Y, E</cell><cell>48.35</cell><cell>37.4</cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>Prof. Dr. Ute Schmid, head of Cognitive Systems Group at University of Bamberg, held a seminar on Explainable AI during the winter semester 2019/2020 out of which this paper emerged.</p></div>
			</div>

			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>-1 for iris setosa -3 for iris versicolor -5 for iris virginia These are also all possible Y E combinations.  Unlike the labels, the combined labels and explanations are not equally distributed (see table <ref type="table">2</ref>).</p><p>An SVM classifier (default parameters), a decision tree (maximum depth of 2 to prevent overfitting), and a DummyClassifier were trained using a 33% test data split (i. e. 100 training examples). The results of the classifiers without balanced class weights can be seen in table <ref type="table">3</ref>. The accuracy of the SVM and decision tree are consistent with the evaluation in <ref type="bibr" target="#b2">[3]</ref>. Performance on label prediction is either as good (SVM) or even better (a 4 percentage point increase for decision tree) when explanations are included in the training data. The prediction accuracy on explanations is also quite high.</p><p>The DummyClassifier shows a surprising increase in accuracy. Without explanations, labels are predicted with 34% accuracy, which is to be expected from a classifier that chooses randomly from 3 equally distributed labels. With explanations, however, the performance increases by over 10 percentage points.</p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title level="m" type="main">One Explanation Does Not Fit All: A Toolkit and Taxonomy of AI Explainability Techniques</title>
		<author>
			<persName><forename type="first">V</forename><surname>Arya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">K E</forename><surname>Bellamy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">Y</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Dhurandhar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hind</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">C</forename><surname>Hoffman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Houde</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><forename type="middle">V</forename><surname>Liao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Luss</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mojsilović</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Mourad</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Pedemonte</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Raghavendra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Richards</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Sattigeri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Shanmugam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Singh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">R</forename><surname>Varshney</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Wei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhang</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1909.03012</idno>
		<imprint>
			<date type="published" when="2019-09">Sep 2019</date>
		</imprint>
	</monogr>
	<note>cs, stat</note>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">The Use of Multiple Measurements in Taxonomic Problems</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">A</forename><surname>Fisher</surname></persName>
		</author>
		<idno type="DOI">10.1111/j.1469-1809.1936.tb02137.x</idno>
		<ptr target="https://doi.org/10.1111/j.1469-1809.1936.tb02137.x" />
	</analytic>
	<monogr>
		<title level="j">Annals of Eugenics</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="179" to="188" />
			<date type="published" when="1936">1936</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">TED: Teaching AI to Explain its Decisions</title>
		<author>
			<persName><forename type="first">M</forename><surname>Hind</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Wei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Campbell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">C F</forename><surname>Codella</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Dhurandhar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mojsilović</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Natesan Ramamurthy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">R</forename><surname>Varshney</surname></persName>
		</author>
		<idno type="DOI">10.1145/3306618.3314273</idno>
		<ptr target="https://doi.org/10.1145/3306618.3314273" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society</title>
				<meeting>the 2019 AAAI/ACM Conference on AI, Ethics, and Society</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2019-01">Jan 2019</date>
			<biblScope unit="page" from="123" to="129" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Scikit-learn: Machine learning in Python</title>
		<author>
			<persName><forename type="first">F</forename><surname>Pedregosa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Varoquaux</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gramfort</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Michel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Thirion</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Grisel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Blondel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Prettenhofer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Weiss</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Dubourg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Vanderplas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Passos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Cournapeau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Brucher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Perrot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Duchesnay</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Machine Learning Research</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="page" from="2825" to="2830" />
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Why Should I Trust You?</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">T</forename><surname>Ribeiro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Singh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Guestrin</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1602.04938</idno>
	</analytic>
	<monogr>
		<title level="m">Explaining the Predictions of Any Classifier</title>
				<imprint>
			<date type="published" when="2016-08">Aug 2016</date>
		</imprint>
	</monogr>
	<note>cs, stat</note>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Please delete that! Why should I?: Explaining learned irrelevance classifications of digital objects</title>
		<author>
			<persName><forename type="first">M</forename><surname>Siebers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><surname>Schmid</surname></persName>
		</author>
		<idno type="DOI">10.1007/s13218-018-0565-5</idno>
		<ptr target="https://doi.org/10.1007/s13218-018-0565-5" />
	</analytic>
	<monogr>
		<title level="j">KI -Künstliche Intelligenz</title>
		<imprint>
			<biblScope unit="volume">33</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="35" to="44" />
			<date type="published" when="2019-03">Mar 2019</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
