<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">XAI-driven Model Improvements in Interpretable Image Segmentation</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Rokas</forename><surname>Gipiškis</surname></persName>
							<email>rokas.gipiskis@mif.vu.lt</email>
							<affiliation key="aff0">
								<orgName type="department">Institute of Data Science and Digital Technologies</orgName>
								<orgName type="institution">Vilnius University</orgName>
								<address>
									<addrLine>4 Akademijos St</addrLine>
									<settlement>Vilnius</settlement>
									<country key="LT">Lithuania</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">XAI-driven Model Improvements in Interpretable Image Segmentation</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">6B8EBFB0ABEBB6AB22F152588E77E7B5</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T19:37+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Explainable AI</term>
					<term>Interpretable AI</term>
					<term>Image segmentation</term>
					<term>XAI</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Semantic image segmentation is the most fine-grained task in computer vision. Its applications range from autonomous vehicles to medical imaging. Despite its deployments in critical areas, interpretable image segmentation remains an underexplored field, especially when compared to explainable AI (XAI) solutions in classification and object detection. Even less attention has been paid to the use of XAI in non-explainability-related scenarios, where XAI methods are applied not for interpretability per se, but rather for other instrumental reasons, such as improving a model's performance. Such use cases can potentially extend to AI safety, specifically in the case of adversarial attacks, self-supervised learning, neural architecture search (NAS), and continual learning (CL). Most of these areas have never been investigated in the context of interpretable segmentation. This work outlines key developments in the field of interpretable image segmentation, with a particular focus on XAI-driven model improvements. We also consider potential uses of interpretable image segmentation for model compression in the case of NAS, and instance-based memory compression in the case of CL.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Image segmentation is a predominant task in computer vision, with applications ranging from medicine to industry. However, evaluation metrics for deep learning (DL) models do not provide a complete view of their performance. Although a model's performance might be good, it could still focus on undesirable spurious correlations. Furthermore, even if the metric is accurate, it does not offer insights into the internal mechanisms of the model. The need for explainable and trustworthy systems is ever-increasing. Yet, there appears to be a clear gap in the explainable AI (XAI) literature between classification and segmentation.</p><p>One could argue that segmentation could be considered a subset or rather an extension of explainable classification. However, this does not resolve the fact that explainable segmentation has its own unique challenges. Firstly, there is the question of how to efficiently generate explanations for segmentation when multiple pixels are involved. Secondly, there is the challenge of how to interpret those explanations when multiple pixels are involved. Another important question, not directly related to enhancements in interpretability, is whether current XAI techniques in image segmentation can be used to improve model performance, specifically in compression-based approaches in neural architecture search (NAS) and continual learning (CL).</p><p>Based on this reasoning, the current doctoral research is framed by the following questions:</p><p>1. How do we improve explainable segmentation techniques for better explainability? This question is directly related to the enhancement of explainable segmentation techniques for a better understanding of segmentation models. 2. How do we use explainable segmentation to enhance performance in tasks not directly related to explainability? This question is related to model improvements by indirectly using XAI-based techniques.</p><p>By combining the results of these two application areas, we can not only identify the most suitable XAI candidates in image segmentation, but also evaluate their potential use in creating more efficient models, both in terms of their weights and memory utilization. Such efficient solutions could be particularly useful in edge computing and in Internet of Things devices.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related Work</head><p>The research into explainable image segmentation began in the late 2010s. Since then, it has seen incremental growth in new methods and applications. For the purposes of this paper, we can divide related works into two groups, based on the end-purpose of XAI application. The first group uses XAI methods to better understand a model's performance or enhance its interpretability. The second group employs XAI techniques not primarily for explainability, but as tools to enhance a model's performance, whether by compressing its weights, improving its memory utilization, or, in the case of self-supervised segmentation, limiting the need for manual annotations.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Explanations for Image Segmentation</head><p>Two influential early applications of XAI in segmentation can be found in <ref type="bibr" target="#b0">[1]</ref> and <ref type="bibr" target="#b1">[2]</ref>. A perturbation-based explainability approach is introduced in <ref type="bibr" target="#b0">[1]</ref> as a method to detect contextual bias. In <ref type="bibr" target="#b1">[2]</ref>, on the other hand, a gradient-based Seg-Grad-CAM solution is proposed and evaluated using U-Net <ref type="bibr" target="#b2">[3]</ref>. Perturbation-based XAI techniques are further investigated in <ref type="bibr" target="#b3">[4]</ref> and <ref type="bibr" target="#b4">[5]</ref>. In <ref type="bibr" target="#b3">[4]</ref>, the focus is on occlusions in the input space, while <ref type="bibr" target="#b4">[5]</ref> explores gradientfree perturbations in the activation space. Simple gradient and Smoothgrad <ref type="bibr" target="#b5">[6]</ref> extensions for segmentation are presented in <ref type="bibr" target="#b6">[7]</ref>, where they are investigated for applications in cyber-physical systems. These methods are further discussed in Section 5. Currently, gradient-based post-hoc approaches are more prevalent in explainable segmentation due to their lower computational costs.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">XAI-driven Model Improvements</head><p>To our knowledge, there are no XAI-driven model improvements specifically for NAS or CL in segmentation. In classification, <ref type="bibr" target="#b7">[8]</ref> proposes a NAS model based on class activation mapping (CAM). The teacher and student models are incorporated into the evolutionary search. The less complex student model has to generate an explanation map that closely approximates the one generated by the teacher model, as measured by the inverse of the Euclidean distance. In <ref type="bibr" target="#b8">[9]</ref>, an input saliency-based NAS is introduced as a way of reweighing different data points. However, the proposed solution only focuses on the features in the input space, leaving investigation of the activation space features for further research. This approach is suitable for differentiable NAS methods, but further investigation is needed for non-differentiable methods, such as evolutionary-algorithm-based NAS. Additional modifications or selecting a non-gradient based optimization algorithm would be required.</p><p>Explainable segmentation techniques have also been investigated for safety and robustness evaluations in segmentation models <ref type="bibr" target="#b6">[7]</ref>. XAI methods can be potentially used in detecting adversarial attacks targeting segmentation. However, the study does not investigate architecturebased changes, so that the safety-critical measures could be incorporated into the model itself. Another non-explainability-related area that uses XAI in model training is self-supervised learning as well as weakly-supervised segmentation. Although, after the initial use of classification saliencies for weakly-supervised object localization <ref type="bibr" target="#b9">[10]</ref>, several other applications have been proposed, the solutions focus on XAI techniques in classification and are not directly related to interpretable segmentation. Typically, explanatory heatmaps for a selected class are used as imprecise segmentation masks that could be employed to increase dataset size, since manual annotations are expensive and time-consuming.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Research</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Research Questions</head><p>Current research questions encompass XAI use for both explainability-related (the first question) and non-explainability-related tasks (the latter two questions):</p><p>1. How can explainability techniques for image segmentation be improved, either in terms of evaluative XAI metrics or computational costs? 2. Can more efficient segmentation models be found by incorporating explainable segmentation techniques in NAS, specifically in the case of teacher-student architectures? 3. Can explainable segmentation techniques be implemented in continual learning for compression in experience replay?</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Hypothesis</head><p>The underlying assumption is that efficient explainable segmentation techniques can identify those regions in the input space or those feature maps in the activation space that are most important for the decision-making of a model and, by extension, its accuracy. Since XAI techniques primarily focus on these areas, their results could be used for model compression in NAS, or memory compression in CL.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.">Objectives</head><p>To better investigate XAI-driven segmentation model enhancements, we define the following objectives:</p><p>1. Identify the most suitable XAI techniques in segmentation based on computational requirements and quantitative XAI metrics. 2. Investigate whether the CAM-NAS application in classification can be successfully extended to segmentation. 3. Evaluate the performance of various explainable segmentation techniques, focusing on their potential uses in NAS. 4. Explore the use of explainable segmentation techniques for memory compression in experience replay by storing only the image areas centered around the most important input features, as identified by selected XAI techniques.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Approach</head><p>Firstly, suitable XAI techniques have to be selected for the experiments. Based on previous research <ref type="bibr" target="#b3">[4]</ref>, <ref type="bibr" target="#b4">[5]</ref>, <ref type="bibr" target="#b6">[7]</ref>, gradient-based methods are preferable for the proposed use-cases due to their lower computational costs. Speed is an important factor when extracting saliency maps, especially when multiple iterations are required. This is further supported by the CAM-NAS <ref type="bibr" target="#b7">[8]</ref> experiments in classification, where gradient-based methods achieve the best results. A simple gradient-based saliency map technique can be used as a baseline. Different implementations of Seg-Grad-CAM <ref type="bibr" target="#b1">[2]</ref> can also be investigated. Since gradient-based techniques can generate a lot of noise, noise-reduction techniques, such as thresholding a certain percentage of pixels based on their importance, might also be considered. Especially when manual human-in-the-loop supervision is involved. NAS focuses on automating the design of neural network architectures. Following <ref type="bibr" target="#b7">[8]</ref>, the initial teacher-student model will be extended to semantic segmentation models. The explanations will be generated based on the summed-up pre-Softmax prediction scores for the selected class of interest (Fig. <ref type="figure" target="#fig_0">1</ref>). A well-trained segmentation model (the teacher) will be paired up with a less complex model (the student). Then, explainable segmentation maps will be generated for the same input images and compared in terms of a similarity score. If the teacher model has truly learned the most important representations in an unbiased way, and if the selected XAI technique can capture the most important features for the model's decision-making, then we would like to have a student model that is also sensitive to the same features. This could be viewed as a knowledge transfer from the teacher model to the student. The original CAM-NAS implementation <ref type="bibr" target="#b7">[8]</ref> uses evolutionary algorithms for the generation of search submodels, and it could serve as an initial starting point for the experiments.</p><p>It is less clear whether XAI-driven model enhancements can be implemented in the case of CL, specifically for memory compression in experience replay. CL focuses on how an already trained model can learn new tasks without forgetting the previous ones. Experience replay is an efficient CL strategy that allows storing the most important examples from old tasks inside the memory so that the model can still be exposed to them in the future. In classification, it is possible to reduce memory utilization by storing just the most salient regions of the data samples <ref type="bibr" target="#b10">[11]</ref>. By cropping the image so that it is centered around the most important regions, memory can be utilized more efficiently. However, it is unclear if the cropping strategy could work in the case of segmentation, as it is a dense prediction task that, unlike image classification, could not be completed if part of the image was missing. In this particular context, compared to compression in classification, segmentation appears to be more sensitive to partial data. Enough critical contextual information would have to be stored for the segmentation to be successful. Perhaps less salient regions could be downsampled, as described in <ref type="bibr" target="#b11">[12]</ref> in the case of classification. Then, enough contextual information could still be preserved to complete segmentation, especially if the right contextual information was identified by the explainable segmentation technique. Following <ref type="bibr" target="#b11">[12]</ref>, once the most important image regions are identified, they can be occluded by a bounding box. The resulting image with unoccluded non-discriminative pixels is then downsampled. Then, the previously occluded salient region is summed up to the downsampled image. The final image occupies significantly less space in memory. To our knowledge, similar experiments have not yet been conducted for CL in image segmentation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Results</head><p>So far, the investigation has focused on explainable segmentation for a better understanding of the model, with an additional focus on XAI safety and robustness <ref type="bibr" target="#b6">[7]</ref>. The first comprehensive survey on XAI in image segmentation has been prepared <ref type="bibr" target="#b12">[13]</ref>, including the initial taxonomy, graphical representations of XAI pipelines for different XAI method categories, and a detailed literature analysis based on evaluation metrics, application domains, and used datasets. Surveyed XAI methods in segmentation have been grouped into gradient-based, perturbation-based, prototype-based, architecture-based, and counterfactual techniques.</p><p>Input perturbation experiments, also known as occlusion sensitivity, have been performed on different segmentation models in <ref type="bibr" target="#b3">[4]</ref>. Both the size and color of the occlusion filter have been investigated using deletion curves, a quantitative XAI evaluation metric. The results indicate that segmentation models are sensitive to varying occlusion colors and sizes, and that this is related to the original colors in the unperturbed input image as well as the ratio between the foreground object of interest and its background. It has also been observed that different occlusion colors can greatly affect the segmentation outputs, even when the same filter size is used and when applied to the same input region. More neutral Gaussian-based filters appear to cause less unnatural distortions to the model's output. In contrast, depending on the background and foreground colors, the black occlusion filter can be mistaken as part of the segmented object. Another interesting finding is that compared to perturbation-based explanations in classification, there is significantly less variance generated in the evaluation scores. Therefore, normalization techniques, such as min-max scaling, can be used to generate clearer explanations with higher color intensities. It is also worth noting that input perturbation methods require significant computational costs, as each occlusion requires an additional inference.</p><p>Perturbations in the activation space have been investigated in <ref type="bibr" target="#b4">[5]</ref>. A new gradient-free Seg-Ablation-CAM technique has been proposed as an extension of <ref type="bibr" target="#b13">[14]</ref>. The method is based on partial or full deactivations of activation maps from the selected neural network layer. The results of foreground and background occlusions indicate that foreground occlusions are more important for the model's output. Based on the qualitative results, the proposed approach provides less noisy and more concentrated saliency visualizations compared to gradient-based XAI methods in segmentation. However, just like other perturbation-based methods, it is computationally expensive, as multiple inferences are required.</p><p>Gradient-based explainable segmentation methods have been described in <ref type="bibr" target="#b6">[7]</ref>, where vanilla gradient and its SmoothGrad <ref type="bibr" target="#b5">[6]</ref> extension have been investigated for industrial applications. This is the first study to focus on adversarial attacks targeting explainable segmentation. It has been demonstrated that segmentation models can be attacked so that the model's output does not change, but its corresponding explanation is arbitrarily manipulated. This is achieved by introducing a three-term loss function which ensures that the perceptible noise is not introduced in the input, that the output is not significantly changed, and that the generated explanation is close to the targeted one, as specified before the attack.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Further Research</head><p>The next research steps will focus on the implementation of CAM-NAS for semantic segmentation. It needs to be evaluated whether the proposed strategy is feasible for higher-resolution images. The final contribution could lead to more efficient image segmentation models. Further research directions could explore how to retain both interpretability and NAS efficiency while investigating the potential trade-offs between the two. This might be related to the fact that some computationally expensive XAI techniques, such as Seg-Ablation-CAM <ref type="bibr" target="#b4">[5]</ref>, are less noisy compared to less computationally demanding techniques, like simple gradients. Other studies in interpretable image segmentation could also investigate whether safety-critical components can be automatically identified by XAI techniques and then incorporated into the proposed architecture.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1:The pipeline for CAM-NAS in segmentation, based on the original implementation<ref type="bibr" target="#b7">[8]</ref> for classification tasks. This is an idealized scenario where the saliency maps generated by both models are identical.</figDesc><graphic coords="5,89.29,84.19,416.70,191.03" type="bitmap" /></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>I extend my gratitude to Prof. Olga Kurasova, my doctoral advisor, and Prof. Chun-Wei Tsai for their support and counsel.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Grid saliency for context explanations of semantic segmentation</title>
		<author>
			<persName><forename type="first">L</forename><surname>Hoyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Munoz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Katiyar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Khoreva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Fischer</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Proceedings of the Advances in Neural Information Processing Systems</title>
		<imprint>
			<biblScope unit="volume">32</biblScope>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Towards interpretable semantic segmentation via gradient-weighted class activation mapping (student abstract)</title>
		<author>
			<persName><forename type="first">K</forename><surname>Vinogradova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Dibrov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Myers</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the AAAI Conference on Artificial Intelligence</title>
				<meeting>the AAAI Conference on Artificial Intelligence</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="volume">34</biblScope>
			<biblScope unit="page" from="13943" to="13944" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">U-net: Convolutional networks for biomedical image segmentation</title>
		<author>
			<persName><forename type="first">O</forename><surname>Ronneberger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Fischer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Brox</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Medical image computing and computer-assisted intervention-MICCAI 2015</title>
				<meeting>the Medical image computing and computer-assisted intervention-MICCAI 2015</meeting>
		<imprint>
			<date type="published" when="2015">2015. 2015</date>
			<biblScope unit="volume">18</biblScope>
			<biblScope unit="page" from="234" to="241" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Occlusion-based approach for interpretable semantic segmentation</title>
		<author>
			<persName><forename type="first">R</forename><surname>Gipiškis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Kurasova</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Iberian Conference on Information Systems and Technologies</title>
				<meeting>the Iberian Conference on Information Systems and Technologies</meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="1" to="6" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Ablation studies in activation maps for explainable semantic segmentation in Industry 4.0</title>
		<author>
			<persName><forename type="first">R</forename><surname>Gipiškis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Chiaro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Annunziata</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Piccialli</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE EUROCON 2023-20th International Conference on Smart Technologies</title>
				<meeting>the IEEE EUROCON 2023-20th International Conference on Smart Technologies</meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="36" to="41" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<author>
			<persName><forename type="first">D</forename><surname>Smilkov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Thorat</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Kim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Viégas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Wattenberg</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1706.03825</idno>
		<title level="m">SmoothGrad: Removing noise by adding noise</title>
				<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">The impact of adversarial attacks on interpretable semantic segmentation in cyber-physical systems</title>
		<author>
			<persName><forename type="first">R</forename><surname>Gipiškis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Chiaro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Preziosi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Prezioso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Piccialli</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Systems Journal</title>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">CAM-NAS: An efficient and interpretable neural architecture search model based on class activation mapping</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Joe</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Applied Sciences</title>
		<imprint>
			<biblScope unit="volume">13</biblScope>
			<biblScope unit="page">9686</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Saliency-aware neural architecture search</title>
		<author>
			<persName><forename type="first">R</forename><surname>Hosseini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Xie</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Proceedings of the Advances in Neural Information Processing Systems</title>
		<imprint>
			<biblScope unit="volume">35</biblScope>
			<biblScope unit="page" from="14743" to="14757" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Learning deep features for discriminative localization</title>
		<author>
			<persName><forename type="first">B</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Khosla</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Lapedriza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Oliva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Torralba</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition</title>
				<meeting>the IEEE Conference on Computer Vision and Pattern Recognition</meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="2921" to="2929" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Online continual learning with saliency-guided experience replay using tiny episodic memory</title>
		<author>
			<persName><forename type="first">G</forename><surname>Saha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Roy</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Machine Vision and Applications</title>
		<imprint>
			<biblScope unit="volume">34</biblScope>
			<biblScope unit="page">65</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Class-incremental exemplar compression for classincremental learning</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Luo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Schiele</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Sun</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</title>
				<meeting>the IEEE/CVF Conference on Computer Vision and Pattern Recognition</meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="11371" to="11380" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<author>
			<persName><forename type="first">R</forename><surname>Gipiškis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C.-W</forename><surname>Tsai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Kurasova</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2405.01636</idno>
		<title level="m">Explainable AI (XAI) in image segmentation in medicine, industry, and beyond: A survey</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Ablation-CAM: Visual explanations for deep convolutional network via gradient-free localization</title>
		<author>
			<persName><forename type="first">S</forename><surname>Desai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">G</forename><surname>Ramaswamy</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision</title>
				<meeting>the IEEE/CVF Winter Conference on Applications of Computer Vision</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="983" to="991" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
