<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Validation of ML Models from the Field of XAI for Computer Vision in Autonomous Driving</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Antonio</forename><surname>Mastroianni</surname></persName>
							<email>antonio.mastroianni@stud.hslu.ch</email>
							<affiliation key="aff0">
								<orgName type="institution">Lucerne University of Applied Sciences and Arts (Hochschule Luzern)</orgName>
								<address>
									<addrLine>Suurstoffi 1</addrLine>
									<postCode>6343</postCode>
									<settlement>Rotkreuz</settlement>
									<country key="CH">Switzerland</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Sibylle</forename><forename type="middle">D</forename><surname>Sager-Müller</surname></persName>
							<email>sibylle.sager@hslu.ch</email>
							<affiliation key="aff0">
								<orgName type="institution">Lucerne University of Applied Sciences and Arts (Hochschule Luzern)</orgName>
								<address>
									<addrLine>Suurstoffi 1</addrLine>
									<postCode>6343</postCode>
									<settlement>Rotkreuz</settlement>
									<country key="CH">Switzerland</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Validation of ML Models from the Field of XAI for Computer Vision in Autonomous Driving</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">3BF20BDB25DC2138CEF5F948A39B3FFD</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T19:37+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Validation</term>
					<term>Metrics</term>
					<term>Captum</term>
					<term>XAI Methods</term>
					<term>Segmentation</term>
					<term>Computer Vision 1</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>We describe Explainable Artificial Intelligence (XAI) methods for image segmentation for autonomous driving applications. The analysis is conducted using metrics such as efficiency, robustness, localization, and complexity. Four XAI methods, namely Gradient-weighted Class Activation Mapping (GradCAM), Local Interpretable Model Agnostic Explanations (LIME), Feature Ablation, and Saliency are applied and assessed on a dataset of street images.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Computer vision through machine learning is a relevant topic in the field of autonomous driving. Within the field of computer vision, there are various tasks such as image classification, object detection, semantic segmentation, and instance segmentation. The focus of this paper lies in the evaluation of Explainable Artificial Intelligence (XAI) methods on segmentation models. Because we want to use the segmentation models for autonomous driving in the future, we focused on images showing street scenes. The current model quality indicates that there is a lot of further work necessary before their implementation in a real-world application.</p><p>The methods refer to algorithms that are included in the Captum package <ref type="bibr" target="#b0">[1]</ref>. For the evaluation of XAI methods, four metrics were used, three of which originate from the categories of complexity, robustness, and localization as listed in Quantus <ref type="bibr" target="#b1">[2]</ref>. Complexity is based on sparseness, robustness on average sensitivity, and localization on Relevance Rank Accuracy (RRA) and accordingly the False Positive Rate of RRA. The fourth metric, efficiency, was measured based on runtime.</p><p>To apply and evaluate XAI methods, it is necessary to have images and a model capable of segmenting images. A publicly available dataset that provides street images with ground truth masks is Cityscapes <ref type="bibr" target="#b2">[3]</ref>. Images from Zurich, Switzerland, were selected from this dataset. For the segmentation, pre-trained models from PyTorch were used <ref type="bibr" target="#b3">[4]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Segmentation Models</head><p>To evaluate the segmentation of street elements, the pre-trained models DeeplabV3 and the Fully Convolutional Network (FCN) both with ResNet101 architecture from PyTorch were analyzed. These models were trained using part of the COCO val2017 dataset and the 20 segmentation classes from Pascal VOC. The decision to use pre-trained models was because these 20 classes include relevant street elements such as bicycles, buses, cars, motorbikes, people, and trains. The models were selected due to their high average Intersection over Union (IoU) scores, which are listed on the PyTorch website <ref type="bibr" target="#b3">[4]</ref>.</p><p>For the evaluation of the segmentation models, two metrics were employed: the IoU value and the Matching Area. The evaluation, shown in Table <ref type="table">1</ref>, involved calculating the average values of 122 images from Zurich for both metrics.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 1 Segmentation Performance Comparison of Deeplabv3 ResNet101 and FCN ResNet101 Models Across Road-Related Classes</head><p>A conclusion drawn from the results of this table is that both the Deeplabv3 and FCN models demonstrate similar performances in segmenting road-related classes. Specifically, the "Car" and "Person" classes tend to exhibit the best results in segmentation for both models. Therefore, only these two classes will be used for further evaluations.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">XAI Methods</head><p>Various XAI methods can be utilized. methods explain the behavior of the model <ref type="bibr" target="#b4">[5]</ref>. In this evaluation, a total of four XAI methods are applied. The selection of methods was partially based on Munn and Pitman <ref type="bibr" target="#b5">[6]</ref>. The authors dedicated a chapter in their book to the topic of explainability for image data. In this chapter, the methods, GradCAM and LIME, were introduced, which was one reason for choosing these two methods. Two further methods were selected: Feature Ablation and Saliency.</p><p>Gradient-weighted Class Activation Mapping (GradCAM) <ref type="bibr" target="#b6">[7]</ref> is a technique that analyzes gradient information for any convolutional layer of a model and generates a heatmap that highlights important regions in the image.</p><p>Local Interpretable Model Agnostic Explanations (LIME) <ref type="bibr" target="#b7">[8]</ref> involves multiple iterations of removing specific regions of an image to determine which specific areas are more or less important. It is a model-agnostic, perturbation-based method.</p><p>Feature Ablation <ref type="bibr" target="#b8">[9]</ref> is also a perturbation-based method. It calculates the difference of the attribution in the model output of each feature when it is active and when it is replaced. Multiple features can be turned off together instead of one at a time.</p><p>Saliency <ref type="bibr" target="#b9">[10]</ref> is a method that follows the gradient of a class through the model using backpropagation. During this process, each pixel is minimally changed, and the resulting changes are observed in the prediction of the class score.</p><p>In the implementation of XAI methods, an image representing a street scene was selected. The following two segmentation classes were examined: cars and persons. The calculated attribution values from the methods were normalized and a heatmap was generated and overlaid on the original image, as shown in Figure <ref type="figure" target="#fig_0">1</ref>. While the original image had dimensions of 2048 x 1024 pixels, it was resized to 512 x 256 pixels to enhance computational speed. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Metrics</head><p>Hedström et al. <ref type="bibr" target="#b1">[2]</ref> compare various XAI libraries of evaluation metrics for XAI methods. Metrics were divided into six categories: Faithfulness, Robustness, Localisation, Complexity, Axiomatic and Randomisation. Three metrics based on these categories were used to evaluate XAI methods and an additional efficiency metric was employed. The selection of the following metrics was guided by the idea to test and understand metrics from different categories but else chosen somewhat arbitrary.</p><p>To measure the efficiency of a method, the runtime was recorded. The time taken to compute the attributions for the "Car" and "Person" classes was measured.</p><p>To calculate robustness, the average sensitivity <ref type="bibr" target="#b10">[11]</ref> was applied. Zhou et al. <ref type="bibr" target="#b11">[12]</ref> describe that models usually do not adapt well to new environments when new factors such as weather or illumination conditions are introduced. For this reason, the original images were modified to various brightness levels while preserving the objects in the image. For calculating the average sensitivity, the following formula (1) was used:</p><formula xml:id="formula_0">𝐴𝑣𝑔𝑆𝑒𝑛𝑠𝑖𝑡𝑖𝑣𝑖𝑡𝑦 = 1 𝑁 𝛴 1=1 𝑁 |𝐴𝑣𝑔𝐴𝑡𝑡𝑟𝑖𝑏𝑢𝑡𝑖𝑜𝑛 𝑜𝑟𝑖𝑔 𝐼𝑚𝑎𝑔𝑒 − 𝐴𝑣𝑔𝐴𝑡𝑡𝑟𝑖𝑏𝑢𝑡𝑖𝑜𝑛 𝑚𝑜𝑑 𝐼𝑚𝑎𝑔𝑒|<label>(1)</label></formula><p>To obtain comparable values, the average sensitivity was calculated using the same method for the "Car" and "Person" classes. A low value of average sensitivity indicates good robustness, where 0 represents the lowest and 1 the highest possible value. Figure <ref type="figure" target="#fig_1">2</ref> illustrates an example of how the average attribution may appear at different brightness levels. Next, we consider the metric Relevance Rank Accuracy (RRA) <ref type="bibr" target="#b12">[13]</ref>, which falls in the category of Localization. This metric measures how many of the highly attributed pixels are located within the ground truth mask. In our evaluation, highly attributed pixels are defined as those falling within the top 20% range. This means when attributions range from 0 to 1, values from 0.8 to 1 are marked as highly attributed. An example of calculating the RRA is illustrated in Figure <ref type="figure" target="#fig_2">3</ref>. Not only is the RRA an important indicator, but also the False Positive (FP) Rate of the RRA. A method might represent all pixels in the image as highly important, resulting in an RRA of 1, which is the highest possible value. Therefore, to achieve more comparable results, the FP RRA was additionally examined. The sparseness <ref type="bibr" target="#b13">[14]</ref> was measured as the last metric, falling in the category of Complexity. Here, the Gini index (G) is used for calculation. In our case, G indicates how scattered or concentrated the attributions are distributed in an image. A value of 0 (the smallest value) indicates a large dispersion, where each pixel is crucial for segmentation. A value of 1 (the largest value) indicates that the important attributions are concentrated. For calculating the Gini index, the following formula (2) was used <ref type="bibr" target="#b14">[15]</ref>:</p><formula xml:id="formula_1">𝐺 = 2 ∑ 𝑖𝑥(𝑖) 𝑛 𝑖=1 𝑛 ∑ 𝑥(𝑖) 𝑛 𝑖=1 − 𝑛 + 1 𝑛<label>(2)</label></formula><p>Let 𝑖 be the indexing for the position of an attribution in an array, 𝑥(𝑖) be the value of the attribution at position 𝑖, and 𝑛 be the number of pixels in the image. The higher the Gini index, the better. A lower value indicates that the importance of all pixels is equal. Our images consist of different objects and only the pixels representing the object should be marked as important, which favors a higher G value.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Evaluation of XAI methods</head><p>In the evaluation, the metrics of average sensitivity, RRA and sparseness were calculated for the targets Car and Person. The target is a parameter that can be selected in every method. The first number in Table <ref type="table">2</ref> represents the Car class, while the second number represents the Person class.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 2</head><p>Evaluation of the methods GradCAM, Feature Ablation, Saliency and LIME on the metrics. The arrow direction indicates whether higher or lower results are considered better.</p><p>In terms of efficiency, GradCAM clearly outperforms all other XAI methods, followed by Saliency, then Feature Ablation and LIME. In terms of robustness, measured by average sensitivity, both GradCAM and Saliency show performance. In particular, Saliency shows almost no change in attribution values despite brightness variations. In the RRA and FP RRA category, Feature Ablation and LIME using Max(0,RRA -FP RRA) show identical top performance. The performance of Saliency in this category is considered significantly off target. The low, almost zero RRA means that Saliency only finds very few highly attributed pixels, present in the ground truth mask. In terms of the sparseness metric, the Saliency method emerges as the top performer. The other three methods show similar values that are significantly different from that of Saliency. Furthermore, it is observed that LIME and Feature Ablation exhibit similar or even identical values. The similarity may arise from the fact that both LIME and Feature Ablation are perturbation-based methods.</p><p>Additionally, it is notable that in most cases, the values from the category "Car" are better than those from the category "Person." One possible reason for the lower performance values in the "Person" class could be the downsizing of the image, as described in Chapter 3. Table <ref type="table">3</ref> shows how the values of the Matching Area, which indicates how well the segmentation aligns with the ground truth mask, get worse when the image is resized. Another reason could be the general size of the objects in the image. A car, in comparison, has a significantly larger area than a person and is therefore easier to detect.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 3 Effects of Image Resizing on Segmentation Accuracy of Class "Person"</head><p>In Figure <ref type="figure" target="#fig_3">4</ref>, the measured metrics are visualized in a radar chart with best scores on the outer circle and worse scores in the inner area.  To evaluate the methods by the used metrics, we need to distinguish between their importance. Two of the most relevant metrics are shown in the lower part of the spider plot, namely the metrics RRA and FP RRA. They are a central aspect in the human interpretability, as they indicate in the heatmap (Figure <ref type="figure" target="#fig_0">1</ref>) what is considered important. Due to the bad interpretability, we decide not to pursue Saliency further. Setting Saliency aside as a method, the three remaining methods can be ordered according to their performance: GradCAM demonstrates superiority, followed by Feature Ablation and LIME.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusion</head><p>In summary, GradCAM emerges in this study as the first choice for evaluating XAI methods for image segmentation, not only because of its favorable metrics but also because of its interpretability in the heatmap. Figure <ref type="figure" target="#fig_0">1</ref> illustrates how GradCAM considers not only the objects themselves but also contextual features such as the road. The heatmap shows a gradient from orange (important) to red (very important), which is missing in other methods. The situation for Saliency was different: While it performed well in the metrics, its heatmap provided almost no information, making interpretation difficult. This highlights the importance of understanding and selecting a comprehensive set of metrics that can provide a clear understanding of the reliability of the methods. We acknowledge that the results are specifically tailored to this case and cannot be directly applied to other fields, such as medicine. Nevertheless, they represent a crucial first step toward future autonomous driving applications.</p><p>The current study is to the best of our knowledge the first one to evaluate a subset of XAI methods for image segmentation, an application area of computer vision that has not received as much attention as, e.g., image classification. Future research should involve a variety of data sets and segmentation models and prioritize the evaluation of a broader range of XAI methods for image segmentation. This research should include the development and application of various evaluation metrics, with a particular focus on interpretability. By optimizing these evaluation criteria, an understanding of the specific strengths and weaknesses of different XAI methods can be achieved, which eventually leads to the identification of the most suitable methods for specific applications.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: The Heatmaps from XAI Methods GradCAM, Feature Ablation, Saliency, and LIME Superimposed on the Original Street Images. The Top Row Shows the "Car" Segments while the Bottom Row Shows the "Person" Segments.</figDesc><graphic coords="3,85.05,425.50,428.33,123.60" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Example with average attribution at different brightness levels. Robustness is ensured when the average attribution remains equal to or only slightly deviates from the original.</figDesc><graphic coords="4,96.48,347.66,402.02,212.85" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Example of calculating the RRA and FP RRA with the ground truth mask.</figDesc><graphic coords="5,96.50,141.79,401.94,226.75" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: Spider plot comparing evaluation metrics of methods (Car left and Person right)</figDesc><graphic coords="7,85.05,85.05,424.90,144.00" type="bitmap" /></figure>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Algorithm Descriptions</title>
		<ptr target="https://captum.ai/docs/attribution_algorithms" />
	</analytic>
	<monogr>
		<title level="j">Captum</title>
		<imprint>
			<date type="published" when="2023-11-25">Nov. 25, 2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Quantus: An Explainable AI Toolkit for Responsible Evaluation of Neural Network Explanations and Beyond</title>
		<author>
			<persName><forename type="first">A</forename><surname>Hedström</surname></persName>
		</author>
		<ptr target="http://arxiv.org/abs/2202.06861" />
	</analytic>
	<monogr>
		<title level="j">arXiv</title>
		<imprint>
			<date type="published" when="2023-04-27">Apr. 27, 2023. Mar. 20, 2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Dataset Overview</title>
		<ptr target="https://www.cityscapes-dataset.com/dataset-overview/" />
	</analytic>
	<monogr>
		<title level="m">Cityscapes Dataset</title>
				<imprint>
			<date type="published" when="2023-12-15">Dec. 15, 2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">&apos;Models and pre-trained weights</title>
		<ptr target="https://pytorch.org/vision/stable/models.html#semantic-segmentation" />
	</analytic>
	<monogr>
		<title level="j">PyTorch</title>
		<imprint>
			<date type="published" when="2023-11-28">Nov. 28, 2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<title level="m" type="main">Interpretable Machine Learning --A Brief History, State-of-the-Art and Challenges</title>
		<author>
			<persName><forename type="first">C</forename><surname>Molnar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Casalicchio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Bischl</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-030-65965-3_28</idno>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="volume">1323</biblScope>
			<biblScope unit="page" from="417" to="431" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><surname>Munn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Pitman</surname></persName>
		</author>
		<title level="m">Explainable AI for practitioners: designing and implementing explainable ML solutions</title>
				<meeting><address><addrLine>Beijing, Sebastopol, CA</addrLine></address></meeting>
		<imprint>
			<publisher>O&apos;Reilly</publisher>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Grad-CAM: Visual Explanations from Deep Networks via Gradient-based Localization</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">R</forename><surname>Selvaraju</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Cogswell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Das</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Vedantam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Parikh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Batra</surname></persName>
		</author>
		<idno type="DOI">10.1007/s11263-019-01228-7</idno>
	</analytic>
	<monogr>
		<title level="j">Int. J. Comput. Vis</title>
		<imprint>
			<biblScope unit="volume">128</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="336" to="359" />
			<date type="published" when="2020-02">Feb. 2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Explaining the Predictions of Any Classifier</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">T</forename><surname>Ribeiro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Singh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Guestrin</surname></persName>
		</author>
		<ptr target="http://arxiv.org/abs/1602.04938" />
	</analytic>
	<monogr>
		<title level="j">arXiv</title>
		<imprint>
			<date type="published" when="2016-02-26">Feb. 26, 2016. May 13, 2024</date>
		</imprint>
	</monogr>
	<note>Why Should I Trust You?</note>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Benchmarking Deep Learning Interpretability in Time Series Predictions</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">A</forename><surname>Ismail</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Gunady</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">C</forename><surname>Bravo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Feizi</surname></persName>
		</author>
		<ptr target="http://arxiv.org/abs/2010.13924" />
	</analytic>
	<monogr>
		<title level="j">arXiv</title>
		<imprint>
			<date type="published" when="2020-10-26">Oct. 26, 2020. Mar. 20, 2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps</title>
		<author>
			<persName><forename type="first">K</forename><surname>Simonyan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Vedaldi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Zisserman</surname></persName>
		</author>
		<ptr target="http://arxiv.org/abs/1312.6034" />
		<imprint>
			<date type="published" when="2014-04-19">Apr. 19, 2014. Mar. 20, 2024</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv</note>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">On the (In)fidelity and Sensitivity for Explanations</title>
		<author>
			<persName><forename type="first">C.-K</forename><surname>Yeh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C.-Y</forename><surname>Hsieh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">S</forename><surname>Suggala</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">I</forename><surname>Inouye</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Ravikumar</surname></persName>
		</author>
		<ptr target="http://arxiv.org/abs/1901.09392" />
	</analytic>
	<monogr>
		<title level="j">arXiv</title>
		<imprint>
			<date type="published" when="2019-11-03">Nov. 03, 2019. Mar. 20, 2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Automated Evaluation of Semantic Segmentation Robustness for Autonomous Driving</title>
		<author>
			<persName><forename type="first">W</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">S</forename><surname>Berrio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Worrall</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Nebot</surname></persName>
		</author>
		<idno type="DOI">10.1109/TITS.2019.2909066</idno>
	</analytic>
	<monogr>
		<title level="j">IEEE Trans. Intell. Transp. Syst</title>
		<imprint>
			<biblScope unit="volume">21</biblScope>
			<biblScope unit="issue">5</biblScope>
			<biblScope unit="page" from="1951" to="1963" />
			<date type="published" when="2020-05">May 2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Ground Truth Evaluation of Neural Network Explanations with CLEVR-XAI</title>
		<author>
			<persName><forename type="first">L</forename><surname>Arras</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Osman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Samek</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.inffus.2021.11.008</idno>
	</analytic>
	<monogr>
		<title level="j">Inf. Fusion</title>
		<imprint>
			<biblScope unit="volume">81</biblScope>
			<biblScope unit="page" from="14" to="40" />
			<date type="published" when="2022-05">May 2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<title level="m" type="main">Concise Explanations of Neural Networks using Adversarial Training</title>
		<author>
			<persName><forename type="first">P</forename><surname>Chalasani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">R</forename><surname>Chowdhury</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Jha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wu</surname></persName>
		</author>
		<ptr target="http://arxiv.org/abs/1810.06583" />
		<imprint>
			<date type="published" when="2020-07-04">Jul. 04, 2020. Mar. 20, 2024</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv</note>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Gini Koeffizient Definition und Berechnung</title>
		<ptr target="https://studyflix.de/wirtschaft/gini-koeffizient-898" />
	</analytic>
	<monogr>
		<title level="j">Studyflix</title>
		<imprint>
			<date type="published" when="2024-03-31">Mar. 31, 2024</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
