<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Preprocessing Ground-Based Hyperspectral Image Data for Improving CNN-based Classification</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Andreas</forename><surname>Schliebitz</surname></persName>
							<email>a.schliebitz@hs-osnabrueck.de</email>
							<affiliation key="aff0">
								<orgName type="institution">Osnabrück University of Applied Sciences</orgName>
								<address>
									<addrLine>Albrechtstr. 30</addrLine>
									<postCode>49076</postCode>
									<settlement>Osnabrück</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Heiko</forename><surname>Tapken</surname></persName>
							<email>h.tapken@hs-osnabrueck.de</email>
							<affiliation key="aff0">
								<orgName type="institution">Osnabrück University of Applied Sciences</orgName>
								<address>
									<addrLine>Albrechtstr. 30</addrLine>
									<postCode>49076</postCode>
									<settlement>Osnabrück</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Martin</forename><surname>Atzmueller</surname></persName>
							<email>martin.atzmueller@uni-osnabrueck.de</email>
							<affiliation key="aff1">
								<orgName type="department">Semantic Information Systems Group</orgName>
								<orgName type="institution">Osnabrück University</orgName>
								<address>
									<addrLine>Wachsbleiche 27</addrLine>
									<postCode>49090</postCode>
									<settlement>Osnabrück</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
							<affiliation key="aff2">
								<orgName type="department">German Research Center for Artificial Intelligence (DFKI)</orgName>
								<address>
									<addrLine>Hamburger Str. 24</addrLine>
									<postCode>49084</postCode>
									<settlement>Osnabrück</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Preprocessing Ground-Based Hyperspectral Image Data for Improving CNN-based Classification</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">571218444B5C4790DEEEE2A9696B90F9</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T16:19+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>hyperspectral image analysis</term>
					<term>data preprocessing</term>
					<term>artificial neural networks</term>
					<term>image classification CEUR Workshop Proceedings</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Complex data -like hyperspectral image data -requires adequate preprocessing methods for tackling the issue of data quality, as a prerequisite for further machine learning approaches like deep learning. This paper addresses preprocessing in the context of ground-based hyperspectral image data: It presents novel preprocessing methods, and proposes a comprehensive preprocessing pipeline for handling complex hyperspectral image samples. Multiple preprocessing pipelines are applied on a set of hyperspectral images in the context of image classification, analyzing which preprocessing algorithms perform best, in order to draw further conclusions about methods and their combinations in our application context. Our results show trends on the application of specific methods, and indicate that the application of shorter pipelines tends to achieve better results. We also provide empirical evidence suggesting that too intensive dimensionality reduction can have detrimental effects on classifiability, regardless of contamination levels.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>In recent years, deep learning-based techniques have advanced significantly, in particular, in the area of computer vision, e. g., <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b1">2,</ref><ref type="bibr" target="#b2">3,</ref><ref type="bibr" target="#b3">4]</ref>. However, one important prerequisite for their successful application is sufficient data quality <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b5">6]</ref>, requiring appropriate preprocessing: in particular, preprocessing needs to be applied e. g., when working with complex data like hyperspectral images <ref type="bibr" target="#b6">[7,</ref><ref type="bibr" target="#b7">8,</ref><ref type="bibr" target="#b8">9]</ref> or performing advanced sampling approaches for quality estimation, cf. <ref type="bibr" target="#b9">[10,</ref><ref type="bibr" target="#b10">11]</ref>.</p><p>In this paper, we target preprocessing methods in the context of analyzing hyperspectral image data: We investigate several preprocessing algorithms based on a literature review, provide efficient Python implementations (detached from data acquisition), and combine them according to the recommendations of Vidal et al. <ref type="bibr" target="#b10">[11]</ref> into comprehensive preprocessing pipelines. These are applied to artificially contaminated hyperspectral images of potatoes. A hyperspectral image is a data cube consisting of many two-dimensional spatial bands stacked along the spectral dimension. Subsequently, the cleaned datasets are classified using an artificial neural network to assess and identify promising combinations of preprocessing methods <ref type="bibr" target="#b11">[12,</ref><ref type="bibr" target="#b12">13,</ref><ref type="bibr" target="#b13">14]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related Work</head><p>Vidal et al. <ref type="bibr" target="#b10">[11]</ref> discuss preprocessing steps for hyperspectral data, as well as individual algorithms within the preprocessing pipeline shown in Figure <ref type="figure">1:</ref> • Dead pixels can be detected using thresholding techniques relying on median spectra or more robust methods like genetic or evolutionary algorithms <ref type="bibr">[11, p. 143</ref>]. In addition to more sophisticated methods like Minimum Volume Ellipsoid (MVE), defect pixels can also be localized by their abnormal intensity values <ref type="bibr">[11, p. 143</ref>]. • Spectral spike points may be removed using median-modified Wiener filters (MMWF), wavelet transforms (Wt), signal derivatives or statistical methods detecting unlikely deviations from the signal's mean <ref type="bibr">[11, p. 144</ref>]. • Background removal can often be performed using manual or automatic thresholding techniques <ref type="bibr">[11, p. 142</ref>]. Outlier detection should only be attempted if a normal distribution of data points can be assumed <ref type="bibr">[11, p. 145</ref>]. If so, methods like Resampling by Half Means (RHM), Smallest Half Volume (SHV) or more robust estimators like MVE or Minimum Covariance Determinant (MCD) can be used <ref type="bibr">[11, p. 145</ref>]. • Following the paper of Vidal et al. <ref type="bibr" target="#b10">[11]</ref>, spectral preprocessing can be split into noise reduction, scatter correction and the reduction of additive baseline shifts. Noise reduction can be performed using 3D or Daubechies wavelets <ref type="bibr">[11, pp. 140-141]</ref>. Both spectral denoising and the rectification of additive baseline shifts can be attempted using the Savitzky-Golay filter <ref type="bibr">[11, p. 145</ref>  <ref type="bibr">[11, p. 140]</ref>. A classification of hyperspectral data cleaned with a selection of these methods is performed by Amigo et al. <ref type="bibr" target="#b15">[16]</ref>, co-author of <ref type="bibr" target="#b10">[11]</ref>. The authors aim at distinguishing flame retardants contained in six different types of plastics using partial least squares regression. In contrast to this paper, the authors do not attempt a classification via artificial neural networks. In the context of hyperspectral image classification, neural networks, especially convolutional types, can be considered as successors of kernel-based methods like PCA and Support Vector Machines (SVM) <ref type="bibr" target="#b16">[17]</ref>. Yu et al. <ref type="bibr" target="#b17">[18]</ref> demonstrate the superior classification ability of convolutional neural networks (CNN) in a comparative study involving 𝑘-nearest neighbor (KNN) and SVM variations. The results show that the CNN-based classifier performs best on all three hyperspectral datasets, being Indian Pines, PaviaU and Salinas. The literature today includes a number of CNN architectures specifically designed for hyperspectral image classification. Ausdebert et al. implement ten of these CNN-based classifiers in their DeepHyperX toolbox <ref type="bibr" target="#b18">[19,</ref><ref type="bibr" target="#b19">20]</ref>, of which the S-CNN architecture <ref type="bibr" target="#b20">[21]</ref> performs best in our experiments (see Table <ref type="table" target="#tab_6">4</ref>).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Methods</head><p>A preprocessing pipeline can be considered as a composition of functions which takes a hyperspectral data cube (hypercube) as input and applies a set of transformations to it. Each preprocessing step can affect the spatial and spectral dimensions of the resulting hypercube, which will serve as input of the subsequent step. Therefore, algorithms in a preprocessing pipeline Figure <ref type="figure">1</ref>: Structure of a hyperspectral preprocessing pipeline, cf. <ref type="bibr" target="#b10">[11]</ref> must be able to adapt to changes in input dimensionality. Experiments have shown that careless composition of algorithms within preprocessing pipelines can have a negative impact on their effectiveness. Figure <ref type="figure">1</ref> (adapted from Vidal et al. <ref type="bibr" target="#b10">[11]</ref>) depicts a reasonable sequence of individual preprocessing steps also used by the pipelines compiled in this paper. Below, the findings of a literature review are used to instantiate several such preprocessing pipelines with selected algorithms. The task of these pipelines is to remove unwanted noise from hyperspectral images, for three contamination levels (low, medium and high). After this preprocessing phase, the cleansed hyperspectral images are used for training a neural classifier in order to draw conclusions about effective combinations of different preprocessing algorithms. The selected preprocessing algorithms are summarized in Table <ref type="table" target="#tab_1">1</ref>. For hyperspectral data, spectral binning is a dimensionality reduction technique, which merges similar bands along the spectral dimension 𝑧 of the hypercube H ∈ R 𝑦×𝑥×𝑧 where 𝑦 and 𝑥 denote the spatio-temporal and spatial dimensions respectively. More generally, 𝑦 represents the height and 𝑥 the width of each spectral band in pixels. Binning can be done manually for obvious groups of similar bands by specifying static bins. Since this approach is rarely practical for all samples in a dataset, a new automatic spectral binning procedure is presented below. The SRD method is based on the average distance 𝑑 ∈ R &gt;0 of the intensity values inside the averaged spectrum I * ∈ R 𝑧 (self-reference) of the hypercube. For dynamic creation of bins, SRD expects a user defined maximum deviation 𝜀 ∈ (0, 1] from this average distance 𝑑. In summary, the SRD algorithm works as follows:</p><p>1. Calculate the reference spectrum I * of H, where the 𝑖-th entry of I * is the mean 𝑏 𝑖 of the 𝑖-th spectral band of H. 2. Calculate the mean distance 𝑑 of all neighboring intensity values in I * , adding the maximum allowed percentage deviation: 𝑑 := 𝑑 • (1 + 𝜀) 3. Obtain the averages 𝑏 𝑖 from the reference spectrum I * with 1 ≤ 𝑖 ≤ 𝑧 and identify the bins using the following case distinction: a) If 𝑖 = 1 or the distance from the current mean 𝑏 𝑖 to the mean of the bands within the current bin is less than 𝑑: Add the 𝑖-th band from H to the current bin. A small deviation from the current bin's mean indicates a high degree of similarity. b) Otherwise: Close the current bin and start with a new one. 4. Create a reduced hypercube H ′ ∈ R 𝑦×𝑥×𝑧 ′ by averaging the spectral bands of each bin.</p><p>This operation reduces the hypercube's depth by merging multiple similar bands into one. The storage gain 𝐺 SRD achieved by spectral binning is calculated from the product of the depth difference 𝑧 − 𝑧 ′ , the unchanged spatial resolution 𝑦 × 𝑥 and the constant storage size 𝑘 of an intensity value in bits:</p><formula xml:id="formula_0">𝐺 SRD = (𝑧 − 𝑧 ′ ) • 𝑦 • 𝑥 • 𝑘 where 0 &lt; 𝑧 ′ ≤ 𝑧.</formula><p>Here, the storage gain is large if the difference 𝑧 − 𝑧 ′ is large, that is, the number of bins generated is small.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.2.">Spatial Binning using a SID-SAM Sliding Window (SSSW)</head><p>Spatial binning, in contrast to spectral binning, achieves dimensionality reduction by merging similar point spectra located in the 𝑥𝑦 plane of the hypercube H ∈ R 𝑦×𝑥×𝑧 . The spatial binning method presented in this section compares multiple point spectra in a 𝑠 × 𝑡 pixel wide window, which is moved over the 𝑥𝑦-plane of the hypercube with a horizontal step size of 𝑠 pixels. After reaching the right image boundary, the window is moved 𝑡 pixels down and repositioned at the left edge of the image. For each point spectrum within a window region, a similarity score is calculated by comparing it to the hypercube's averaged spectrum I * using the SID-SAM similarity measure <ref type="bibr" target="#b30">[31]</ref>. In doing so, only those spectra are merged that exhibit a high similarity to I * and thus contribute comparatively little to the hypercube's entropy. SSSW can be used with all types of hyperspectral data cubes regardless of the window size, since the median spectrum I * does always exist. The similarity matrix R 𝑦×𝑥 stores the SID-SAM score for each point spectrum. An entry of R is small if the associated point spectrum is similar to I * . Afterwards, the matrix R is subdivided into the original 𝑠 × 𝑡 wide window regions. The SID-SAM values inside these smaller matrices are averaged to obtain a single similarity measure tied to a specific image area. Subsequently, all averaged SID-SAM values are associated with their corresponding point spectra in the 𝑥𝑦-plane of the hypercube. In doing so, a maximum of 𝑢 = (⌊𝑦/𝑠⌋ + 1) • (⌊𝑥/𝑡⌋ + 1) non-overlapping image regions are considered for spatial binning. After sorting the averaged SID-SAM values in ascending order, the point spectra eligible for merging are determined through the use of a user defined percentile 𝜃 ∈ [0, 1]. In this case, it is sufficient to associate the smallest ⌊𝜃 • 𝑢⌋ SID-SAM values with their corresponding window regions and merge the contained spectra by spectral averaging. The storage gain 𝐺 SSSW achieved by the SSSW algorithm increases with the window size 𝑠 × 𝑡 and the percentile 𝜃 used for binning:</p><formula xml:id="formula_1">𝐺 SSSW ≈ ⌊𝜃 • 𝑢⌋ • (𝑠 • 𝑡 − 1) • 𝑧 • 𝑘 bits.</formula><p>The value 𝑘 represents the constant storage size of an intensity value in bits. Strictly speaking, this storage gain calculation is rather an approximation, since the sliding window procedure will generate irregular window sizes if either the image width or height is not a multiple of 𝑠 or 𝑡 respectively. Finally, since merging multiple point spectra creates gaps in the original data cube, a compression is performed by densely rearranging the binned point spectra resulting in different output dimensions. This procedure has the disadvantage of separating spatial features in the 𝑥𝑦-plane of the hypercube.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Preprocessing Pipelines</head><p>The combinatorial complexity associated with instantiating various alternative preprocessing pipelines is significantly reduced by using the fixed order of the individual preprocessing steps depicted in Figure <ref type="figure">1</ref>. The number of eligible pipeline combinations is further limited by purposely leaving out MSC, since the linearly related SNV <ref type="bibr" target="#b31">[32]</ref> is faster to compute. Furthermore, due to the absence of a reference spectrum, there are no advantages justifying the use of MSC in this paper. Additional pipeline combinations are eliminated by always applying SDT and SDF at the beginning of each preprocessing pipeline. This results in the generated pipelines differing only in their preprocessing algorithms for outlier detection, noise reduction, and dimension reduction (see Figure <ref type="figure" target="#fig_0">2</ref>). The input and output degrees of all inner nodes of this graph are initially identical by definition, since each node produces an output and passes it as input to all descendant nodes.  <ref type="table" target="#tab_1">1</ref>.</p><p>For reasons of clarity, the three outgoing edges of each noise reduction node (MNF, Wt, SGF) in Figure <ref type="figure" target="#fig_0">2</ref> are merged into a single edge. The same applies to the outgoing edges of the SNV node. The number of pipelines constructed in this way can be determined by counting all possible paths from the root (SDT) to the leaf nodes (SRD, PCA, SSSW) of the graph. Since preprocessing steps with only one algorithm do not create additional pipelines, only three of the six preprocessing steps have to be considered with three alternatives each. Therefore, the graph shown in Figure <ref type="figure" target="#fig_0">2</ref> produces a total of 3 3 = 27 different pipeline combinations.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.">Reference Dataset</head><p>The raw dataset used in this paper consists of hyperspectral images depicting a total of 728 clean potatoes, which were manually acquired using a Helios Core 0.9-1.7 XF-PA100 240×320/CII/H330 line scan camera.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4.">Dataset Contamination</head><p>The spectra of the reference dataset are artificially contaminated with varying amounts of noise to investigate the effectiveness of different preprocessing pipelines. This procedure became necessary due to an unknown amount of in-camera preprocessing that was applied to the raw sensor data at acquisition time. Contamination is performed in three stages using additive and multiplicative scattering effects, Gaussian noise, spike points and dead pixels. Table <ref type="table" target="#tab_5">3</ref> quantifies the degree of artificial contamination for all three contamination levels using statistical indicators applied to the whole set of intensity values contained within each dataset. The strength of additive and multiplicative scattering effects is dynamically calculated as a function of contamination level and applied to random spectra of a fixed user-defined percentile (24 %). As with the other types of contamination, the spectra are increasingly influenced by additive and multiplicative scattering effects at higher contamination levels. Additive scattering effects can manifest themselves not only in a positive but also negative shift of intensity values. In contrast to additive influences, multiplicative scattering affects the slopes within a spectrum's wavelength profile. In this paper, this phenomenon is simulated by multiplying a spectrum with by factor between 0.5 and 0.9. This ensures that the intensity values of the spectrum are not shifted out of their normalized range of values [0,1]. Since multiplying a spectrum by a number close to one has little effect on its slopes, the highest scaling factor is used at the lowest contamination level. Gaussian noise is applied to every sample with increasing variance at higher contamination levels (0.0004, 0.0012, 0.0020). Simulation of both spike points (max. 1.5 %) and defect pixels (max. 2.4 %) is implemented by randomly setting maximum (1) or minimum (0) intensity values. The influence of defect pixels is simulated within the 𝑖-th band B 𝜆 𝑖 ∈ R 𝑦×𝑥 by randomly replacing columns, such that they consist either of the minimum or maximum value 0, 1 ∈ R 𝑦 , respectively.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.5.">Classification of Preprocessed Datasets</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.5.1.">Dataset Sampling</head><p>Sampling of both the reference dataset and the three contaminated datasets is performed using a special sampling method that guarantees the spatial disjointness of training, test, and validation spectra. The purpose of this sampling procedure is to separate the point spectra of a spatially contiguous dataset in such a way that no two spectra of the same potato appear in more than one of the three dataset splits. The developed sampling strategy achieves this by dividing the ground truth mask of the reference dataset into three overlap-free regions (see Figure <ref type="figure" target="#fig_1">3</ref>). This trisection is performed along the horizontal axis of the contiguous hypercube, since a vertical decomposition could result in a class imbalance within the partial datasets. The reason for this is that the potatoes inside the contiguous hypercube are sorted according to their size in order to save space by reducing the amount of empty background voxels. However, in our dataset, green potatoes were on average significantly smaller than potatoes of other quality classes and therefore located in the lower third of the ground truth mask. Performing vertical partitioning along the horizontal axis showed an improvement in the distribution of potato sizes and classes in each dataset split. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.5.2.">Network Architecture Selection</head><p>Classification of the contaminated datasets as well as those cleaned by the preprocessing pipelines is performed using the DeepHyperX toolbox <ref type="bibr" target="#b18">[19,</ref><ref type="bibr" target="#b19">20]</ref>. For the selection of a suitable classifier, the reference dataset consisting of 8.  <ref type="table" target="#tab_6">4</ref>). Due to the long training duration of the sharma architecture, the second ranked nn architecture is chosen as the classifier to evaluate the performance of the different preprocessing pipelines, serving as a prototypical instance towards more advanced architectures. The nn architecture is not a convolutional neural network, but rather a simple multilayer perceptron (MLP) consisting of only four fully connected layers, each with 2048 neurons.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Results</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Experimental Setup</head><p>The nn architecture is trained on approximately 4.5 million spectra from each of the contaminated datasets in 12 epochs with a batch size of 100. Training is monitored at each epoch with approximately 225 000 validation spectra while the learning rate is dynamically adjusted by the ReduceLROnPlateau scheduler. Cross entropy is used as the loss function in conjunction with the Adam optimizer. As a regularization mechanism, Dropout is applied with a dropout probability of 𝑝 = 0.5. The evaluation of a trained model is performed on approximately 900 000 test spectra. Since the training, test, and validation datasets are created using the spatially disjoint sampling method from Section 3.5.1, no spectra from the same potato can appear in both the training and test datasets. The computed performance metric is the classification accuracy of each model on its respective test dataset.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Reference Dataset</head><p>The nn classifier trained on the reference dataset achieves an accuracy of 53.61 % on the first four defect classes (cf. Table <ref type="table" target="#tab_4">2</ref>), which is used as a baseline for performance comparison below.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3.">Contaminated Datasets</head><p>As expected, the classification accuracy of the nn architecture decreases monotonically but not uniformly as the contamination level of the dataset increases. A test accuracy of 40.77 % was reported for low, 37.53 % for medium and 36.36 % for high contamination. The observed degradation spans 4.41 %, with a gap of 3.24 % from second to first place. This difference is almost three times larger than from third to second rank (1.17 %). Compared to the classification results obtained from the reference dataset, the nn architecture exhibits a drop in test accuracy of 12.84 % at low contamination. Hence, the generalization ability of the nn architecture is found to be significantly weaker on the contaminated datasets.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.4.">Preprocessed Datasets</head><p>The classification results listed in Tables <ref type="table" target="#tab_8">5 and 7</ref> correspond to the pipeline combinations implied by Figure <ref type="figure" target="#fig_0">2</ref>. Also, the 36 rows of the first table together with the 9 rows of the second table constitute the 45 preprocessing runs mentioned in section 3.5.2. Before analyzing the reported classification accuracies, it is important to note that the tables presented in this section do not explicitly list the first two preprocessing algorithms dealing with dead pixel (SDT) and spike detection (SDF). As mentioned in Section 3.2, these procedures are always applied in this order at the beginning of each preprocessing pipeline, such that they are omitted for better readability of the tabular listings. Table <ref type="table">6</ref> quantifies the performance differential between the various dimensionality reduction techniques observed in Table <ref type="table" target="#tab_8">7</ref>. The left half of this listing includes the mean classification accuracies of the first 19 pipeline combinations using spectral binning, which are grouped by their respective contamination level and sorted by test accuracy. This ordering also applies to the right half of Table <ref type="table">6</ref>, which summarizes the same metrics for the last 17 pipelines using PCA for dimensionality reduction. Table <ref type="table" target="#tab_8">7</ref> reports the classification results of those preprocessing pipelines that discard the spatial features of their inputs. This is because, unlike in Table <ref type="table" target="#tab_7">5</ref>, outliers are detected using the COPOD algorithm and dimensionality reduction is performed through spatial (SSSW) rather than spectral binning (SRD). This leads to a new peak classification accuracy of 44.52 %, i. e., 1.81 % higher than reported in Table <ref type="table" target="#tab_7">5</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.5.">Discussion</head><p>The classification results listed in Table <ref type="table" target="#tab_7">5</ref> show a clear dichotomy in their final processing step. The first 19 pipelines largely perform spectral binning as their dimensionality reduction, whereas the last 17 pipelines apply PCA. All pipeline combinations in the upper half of Table <ref type="table" target="#tab_7">5</ref> (rows 1 to 19) were able to improve the classifiability of datasets with medium and high contamination levels. None of the last 17 pipelines succeeded in achieving the same objective. The pipelines which showed the largest improvement in classifiability of their cleaned datasets are located in the first two rows of  this observation, it can be concluded that a drastic dimensionality reduction using PCA can have detrimental effects on the classifiability of hyperspectral datasets. This claim is further substantiated by rows 20 to 36 of Table <ref type="table" target="#tab_7">5</ref>. For reference, the PCA carried out in this paper was set to obtain 99.8 % of a sample's variance.</p><p>In Table <ref type="table">6</ref>, we observe that the contamination levels are reversed in comparison to the classification results of the contaminated datasets (see Section 4.3). On average the best improvements in test accuracy can be obtained on the most contaminated datasets (+3.03 %). Our experiments indicate, that preprocessing datasets with low contamination levels may even lead to a deterioration of their classifiability (∅ -1.48 %). In our tests, this observation holds true for each individual pipeline applied to a dataset with low contamination. The best classification result of a lightly contaminated dataset is 39.71 % (rank 5) behind two more contaminated datasets. The second half of the table shows the negative impact of PCA on the classification results. The greatest losses occur for datasets of the lowest contamination level, whose test accuracy decreases by 7.71 % on average. In second place are the most contaminated datasets, whose mean classification accuracy is least affected by PCA (-3.67 %). Regarding noise reduction, the first five rows of Table <ref type="table" target="#tab_7">5</ref> indicate that the MNF algorithm is superior to SGF and Wt. This coincides with both the visual appearance of the denoised samples and the fact that MNF was specifically designed for hyperspectral noise reduction. Even the popular Savitzky-Golay filter, which seems to work well with the Reed-Xiaoli detector, cannot outperform MNF. In terms of outlier detection, the combinations of SOS and MNF (3) and RXD and MNF (2) occur almost equally often in the Top-6 results of Table <ref type="table" target="#tab_7">5</ref>. However, the highest ranked preprocessing pipeline uses RXD rather than SOS for outlier detection. The importance of RXD increases after the rankings are extending to the first eleven results, where RXD is used by six pipelines and SOS only by five. In the lower half of Table <ref type="table" target="#tab_7">5</ref> we observe that SGF performs poorly in conjunction with PCA, whereas the wavelet transform appears to be more robust. Similar advantages emerge in the midfield, where wavelet filtering is on par with SGF and even outperforms on datasets with higher contamination levels.</p><p>Overall, the best classification accuracy is achieved by a pipeline which discards features in the spatial-temporal domain (see Table <ref type="table" target="#tab_8">7</ref>). This pipeline preprocesses a dataset of low contamination -classified with an accuracy of 44.52 %. This result is 1.81 % higher than the previous peak of 42.71 % (see Table <ref type="table" target="#tab_7">5</ref>). The second best accuracy was previously obtained by a combination of RXD, MNF and spectral binning. For the results shown in Table <ref type="table" target="#tab_8">7</ref>, there is no superior method for noise reduction in the context of COPOD and spatial binning. Both MNF and SGF show promising results in spectral and spatial smoothing, while wavelet transform clearly underperforms.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Conclusions</head><p>This paper considered classification methods on hyperspectral image data, specially presenting novel preprocessing methods, which were integrated into a comprehensive preprocessing pipeline. We performed an analysis of multiple preprocessing pipelines in the context of image classification, in order to analyze which preprocessing algorithms perform best.</p><p>Our results indicate that dimensionality reduction using principal component analysis can potentially have negative effects on the classifiability of hyperspectral data. Furthermore, both the Reed-Xiaoli detector and the MNF transform are preferable as dedicated hyperspectral preprocessing algorithms to general-purpose methods such as SOS or wavelet filtering. The Savitzky-Golay filter, widely used in signal processing, can also be used successfully for smoothing noisy spectra. The results obtained with this method can be considered satisfactory as long as no dimensional reduction is performed with PCA. For this reason, it is recommended that dimensionality reduction should be removed from a preprocessing pipeline and not applied across the board to all samples. If data reduction is still required, then the less invasive but also less effective methods of spatial and spectral binning can be used as an alternative to PCA. Furthermore, a step-by-step calculation of the results listed in the Table <ref type="table" target="#tab_7">5</ref> and 7 showed that shorter pipelines correlate with better classification results. Therefore, preprocessing pipelines should contain as few algorithms as possible, focusing on those that are tailored to the type of contamination at hand. In order to determine which preprocessing steps are really useful, an examination of the contaminated dataset is recommended.</p><p>In our experimentation, advantages in favoring preprocessing pipelines that preserve spatial features were not observed in terms of classification accuracy using a multilayer perceptron. However, it can be conjectured that convolutional neural networks can extract generalizable features from coherent spatio-temporal regions, which is supported by the results shown in Table <ref type="table" target="#tab_6">4</ref>. In summary, for classifying hyperspectral image data we recommend the following strategies:</p><p>1. The preprocessing algorithms should be as domain-specific as possible.</p><p>2. A minimal number of pre-processing algorithms should be used.</p><p>3. The preprocessing algorithms should address the specific impurities of the dataset. 4. The algorithms should be carefully parameterized and arranged according to Figure <ref type="figure">1</ref>. 5. For further classification, a CNN designed for hyperspectral data should be used. For future work, we aim to extend the analysis of the applied neural network architectures. Also, for explainability and interpretability of the applied method <ref type="bibr" target="#b32">[33,</ref><ref type="bibr" target="#b33">34,</ref><ref type="bibr" target="#b34">35,</ref><ref type="bibr" target="#b35">36]</ref> both specialized preprocessing approaches as well as respective feature generation, e. g., using declarative or neuro-symbolic techniques <ref type="bibr" target="#b36">[37,</ref><ref type="bibr" target="#b37">38,</ref><ref type="bibr" target="#b38">39]</ref> provide interesting directions for future research.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Preprocessing pipelines with algorithms from Table 1.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Sampling of training, testing, and validation data from three non-overlapping regions.</figDesc><graphic coords="7,89.29,599.91,419.56,54.88" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head></head><label></label><figDesc>3 million point spectra is limited to a training and test dataset of 200 000 spectra each. The training is monitored with a validation split of 20 000 spectra. The goal of these initial training runs is to determine a neural network architecture that has a low training duration with comparatively high accuracy. Excessively long training durations are unacceptable from a practical point of view, since up to 45 training and testing iterations must be performed with the pipelines constructed in Section 3.2. The seven neural network architectures evaluated in advance to this paper are part of the DeepHyperX toolbox and trained in six epochs with a batch size of 100. Other settings and hyperparameters like learning rate, its scheduler, optimizer and loss function are left at their defaults. All training runs are performed on two Tesla V100 graphics cards of an NVIDIA DGX Station. The comparatively low amount of training spectra is due to the long training duration (4 h 20 min) of the best performing CNN architecture developed by Sharma et al. (see Table</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head></head><label></label><figDesc>]. For scatter correction the use of the Standard Normal Variate (SNV) is advisable if no reference spectrum is available. Otherwise, if a reference spectrum can be recorded a priori, then Multiplicative Scatter Correction (MSC) should be considered instead [11, p. 145] [15]. • Dimensionality reduction of a hyperspectral datacube can be achieved using variable selection techniques based on genetic algorithms or partial least squares regression (e. g. iPLS) [11, p. 140]. Other methods include spatial and spectral binning as well as factor model approaches such as Principal Component Analysis (PCA) or Multivariate Curve Resolution (MCR)</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 1</head><label>1</label><figDesc>Overview of algorithms used for creating preprocessing pipelines.</figDesc><table><row><cell>Preprocessing step</cell><cell>Algorithm</cell><cell>Abbr.</cell><cell>Ref.</cell></row><row><cell>Dead pixel detection</cell><cell>Standard Deviation Threshold</cell><cell>SDT</cell><cell>[11, sec. 6]</cell></row><row><cell>Spike detection</cell><cell>Standard Deviation Factor</cell><cell>SDF</cell><cell>[11, sec. 7]</cell></row><row><cell></cell><cell>Stochastic Outlier Selection</cell><cell>SOS</cell><cell>[22]</cell></row><row><cell>Outlier detection</cell><cell>Copula Based Outlier Selection</cell><cell cols="2">COPOD [23]</cell></row><row><cell></cell><cell>Reed-Xiaoli Detector</cell><cell>RXD</cell><cell>[24]</cell></row><row><cell></cell><cell>Savitzky-Golay Filter</cell><cell>SGF</cell><cell>[25]</cell></row><row><cell>Noise reduction</cell><cell>Daubechies Wavelet Filter</cell><cell>Wt</cell><cell>[26]</cell></row><row><cell></cell><cell>Minimum Noise Fraction</cell><cell>MNF</cell><cell>[27]</cell></row><row><cell>Scatter correction</cell><cell cols="2">Standard Normal Variate Multiplicative Scatter Correction MSC SNV</cell><cell>[28] [29]</cell></row><row><cell></cell><cell>Principal Component Analysis</cell><cell>PCA</cell><cell>[30]</cell></row><row><cell>Dimensionality reduction</cell><cell>Spectral Binning</cell><cell>SRD</cell><cell>-</cell></row><row><cell></cell><cell>Spatial Binning</cell><cell>SSSW</cell><cell>-</cell></row><row><cell>3</cell><cell></cell><cell></cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>.1. Methods for Dimensionality Reduction 3.1.1. Spectral Binning using Self-Reference Deviation (SRD)</head><label></label><figDesc></figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head></head><label></label><figDesc>Table2includes 18 more potatoes for a grand total of 746 since some potatoes are annotated with multiple classes. During dataset creation, samples annotated with multiple classes are duplicated to each assigned quality class. To avoid methodological errors in training and evaluation, potatoes with multiple defect classes are not included in the contiguous reference dataset. Additionally, the raw dataset is restricted to the four largest defect classes from Table2to ensure the availability of sufficient test spectra during evaluation. These measures reduce the size of the raw dataset from 728 to initially 622 and after deduplication to finally 606 potatoes with an average of approximately 13 700 point spectra each.</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 2</head><label>2</label><figDesc>Quality classes of the hyperspectral potato dataset.</figDesc><table><row><cell># Quality class</cell><cell>Quantity</cell><cell>Σ</cell><cell># Quality class</cell><cell>Quantity</cell><cell>Σ</cell></row><row><cell>1 green</cell><cell>264</cell><cell></cell><cell cols="2">7 growth deformity 17</cell><cell></cell></row><row><cell>2 dry rot 3 no defect</cell><cell>171 117</cell><cell>622</cell><cell>8 blue spots 9 withered</cell><cell>17 14</cell><cell>124</cell></row><row><cell>4 growth crack</cell><cell>70</cell><cell></cell><cell>10 wet rot</cell><cell>9</cell><cell></cell></row><row><cell cols="2">5 mechanical damage 43 6 scab 24</cell><cell></cell><cell></cell><cell>Total</cell><cell>746</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 3 :</head><label>3</label><figDesc>Statistical characteristics of the three contamination levels.</figDesc><table><row><cell cols="3">Contam. Avg. SD (𝜎) Var. (𝜎 2 )</cell></row><row><cell>low</cell><cell>0.374 0.241</cell><cell>0.058</cell></row><row><cell>medium</cell><cell>0.353 0.246</cell><cell>0.061</cell></row><row><cell>high</cell><cell>0.333 0.256</cell><cell>0.066</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_6"><head>Table 4</head><label>4</label><figDesc>Classification accuracies and training durations of a hyperspectral CNN (sharma) and multilayer perceptron (nn) on spectra of the reference dataset.</figDesc><table><row><cell>Arch.</cell><cell cols="4">Train loss Val. acc. Test acc. Train 𝑡 Test 𝑡</cell></row><row><cell>sharma [21]</cell><cell>0.228</cell><cell>59.7 %</cell><cell>64.7 %</cell><cell>4:20:08 0:14:57</cell></row><row><cell>nn [19]</cell><cell>0.999</cell><cell>49.4 %</cell><cell>52.8 %</cell><cell>0:02:26 0:06:51</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_7"><head>Table 5 (</head><label>5</label><figDesc>+6.35 %, +5.03 %).</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_8"><head>Table 7</head><label>7</label><figDesc>Classification results of preprocessed datasets with discarded spatial features.</figDesc><table><row><cell cols="5"># Contam. COPOD SGF Wt MNF SNV SSSW Test acc.</cell></row><row><cell>1 low</cell><cell></cell><cell>-</cell><cell>-</cell><cell>44.52 %</cell></row><row><cell>2 high</cell><cell>-</cell><cell>-</cell><cell></cell><cell>42.40 %</cell></row><row><cell>3 medium</cell><cell></cell><cell>-</cell><cell>-</cell><cell>42.35 %</cell></row><row><cell>4 medium</cell><cell>-</cell><cell>-</cell><cell></cell><cell>41.39 %</cell></row><row><cell>5 high</cell><cell></cell><cell>-</cell><cell>-</cell><cell>40.89 %</cell></row><row><cell>6 low</cell><cell>-</cell><cell></cell><cell>-</cell><cell>38.62 %</cell></row><row><cell>7 high</cell><cell>-</cell><cell></cell><cell>-</cell><cell>35.78 %</cell></row><row><cell>8 medium</cell><cell>-</cell><cell></cell><cell>-</cell><cell>35.41 %</cell></row><row><cell>9 low</cell><cell>-</cell><cell>-</cell><cell></cell><cell>34.64 %</cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>This work was partly funded by the German Federal Ministry of Economics and Climate Protection as part of the research project Agri-Gaia under grant number 01MK21004G.</p></div>
			</div>

			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>By examining the actual dimensionality reductions performed in Table <ref type="table">5</ref>, we observe that the highest ranked pipeline (42.71 % acc.) uses spectral binning to perform an average dimensionality reduction of 81 bands per sample. After replacing the spectral binning with PCA, the test accuracy drops by 10.79 % to 31.92 % (rank 31). The dimensionality reduction carried out by the PCA decreases the average depth of the respective dataset from 252 to only 3 bands. Based on</p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Recent advances in convolutional neural networks</title>
		<author>
			<persName><forename type="first">J</forename><surname>Gu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kuen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Shahroudy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Shuai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Cai</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.patcog.2017.10.013</idno>
	</analytic>
	<monogr>
		<title level="j">Pattern recognition</title>
		<imprint>
			<biblScope unit="volume">77</biblScope>
			<biblScope unit="page" from="354" to="377" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">A survey of the recent architectures of deep convolutional neural networks</title>
		<author>
			<persName><forename type="first">A</forename><surname>Khan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Sohail</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><surname>Zahoora</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">S</forename><surname>Qureshi</surname></persName>
		</author>
		<idno type="DOI">10.1007/s10462-020-09825-6</idno>
	</analytic>
	<monogr>
		<title level="j">Artificial Intelligence Review</title>
		<imprint>
			<biblScope unit="volume">53</biblScope>
			<biblScope unit="page" from="5455" to="5516" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Deep Learning for Image Processing Applications</title>
		<author>
			<persName><forename type="first">D</forename><surname>Hemanth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Estrela</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in Parallel Computing</title>
				<imprint>
			<publisher>IOS Press</publisher>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">A Survey on the New Generation of Deep Learning in Image Processing</title>
		<author>
			<persName><forename type="first">L</forename><surname>Jiao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Zhao</surname></persName>
		</author>
		<idno type="DOI">10.1109/ACCESS.2019.2956508</idno>
	</analytic>
	<monogr>
		<title level="j">IEEE Access</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="page" from="172231" to="172263" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Data collection and quality challenges for deep learning</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">E</forename><surname>Whang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J.-G</forename><surname>Lee</surname></persName>
		</author>
		<idno type="DOI">10.14778/3415478.3415562</idno>
	</analytic>
	<monogr>
		<title level="j">Proceedings of the VLDB Endowment</title>
		<imprint>
			<biblScope unit="volume">13</biblScope>
			<biblScope unit="page" from="3429" to="3432" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Data collection and quality challenges in deep learning: A data-centric ai perspective</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">E</forename><surname>Whang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Roh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Song</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J.-G</forename><surname>Lee</surname></persName>
		</author>
		<idno type="DOI">10.1007/s00778-022-00775-9</idno>
	</analytic>
	<monogr>
		<title level="j">The VLDB Journal</title>
		<imprint>
			<biblScope unit="volume">32</biblScope>
			<biblScope unit="page" from="791" to="813" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Recent advances in techniques for hyperspectral image processing</title>
		<author>
			<persName><forename type="first">A</forename><surname>Plaza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>Benediktsson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">W</forename><surname>Boardman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Brazile</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Bruzzone</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Camps-Valls</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Chanussot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Fauvel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Gamba</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gualtieri</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.rse.2007.07.028</idno>
	</analytic>
	<monogr>
		<title level="j">Remote sensing of environment</title>
		<imprint>
			<biblScope unit="volume">113</biblScope>
			<biblScope unit="page" from="S110" to="S122" />
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Near-infrared spectroscopy and hyperspectral imaging: non-destructive analysis of biological materials</title>
		<author>
			<persName><forename type="first">M</forename><surname>Manley</surname></persName>
		</author>
		<idno type="DOI">10.1039/c4cs00062e</idno>
	</analytic>
	<monogr>
		<title level="j">Chemical Society Reviews</title>
		<imprint>
			<biblScope unit="volume">43</biblScope>
			<biblScope unit="page" from="8200" to="8214" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Hyperspectral Imaging and Chemometrics: A Perfect Combination for the Analysis of Food Structure, Composition and Quality</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Amigo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Martí</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gowen</surname></persName>
		</author>
		<idno type="DOI">10.1016/B978-0-444-59528-7.00009-0</idno>
	</analytic>
	<monogr>
		<title level="j">Data handling in science and technology</title>
		<imprint>
			<biblScope unit="volume">28</biblScope>
			<biblScope unit="page" from="343" to="370" />
			<date type="published" when="2013">2013</date>
			<publisher>Elsevier</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Hyperspectral Imaging: A Review of Best Practice, Performance and Pitfalls for in-line and on-line Applications</title>
		<author>
			<persName><forename type="first">B</forename><surname>Boldrini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Kessler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Rebner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Kessler</surname></persName>
		</author>
		<idno type="DOI">10.1255/jnirs.1003</idno>
	</analytic>
	<monogr>
		<title level="j">Journal of Near Infrared Spectroscopy</title>
		<imprint>
			<biblScope unit="volume">20</biblScope>
			<biblScope unit="page" from="438" to="508" />
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Pre-processing of hyperspectral images. essential steps before image analysis</title>
		<author>
			<persName><forename type="first">M</forename><surname>Vidal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Amigo</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.chemolab.2012.05.009</idno>
	</analytic>
	<monogr>
		<title level="j">Chemometrics and Intelligent Laboratory Systems</title>
		<imprint>
			<biblScope unit="volume">117</biblScope>
			<biblScope unit="page" from="138" to="148" />
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Hyperspectral Imaging Applications in Agriculture and Agro-Food Product Quality and Safety Control: A Review</title>
		<author>
			<persName><forename type="first">L</forename><surname>Dale</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Thewis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Boudry</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Rotar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Dardenne</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Baeten</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">Fernández</forename><surname>Pierna</surname></persName>
		</author>
		<idno type="DOI">10.1080/05704928.2012.705800</idno>
	</analytic>
	<monogr>
		<title level="j">Applied Spectroscopy Reviews</title>
		<imprint>
			<biblScope unit="volume">48</biblScope>
			<biblScope unit="page">142</biblScope>
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">A review of deep learning used in the hyperspectral image analysis for agriculture</title>
		<author>
			<persName><forename type="first">C</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Hou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Li</surname></persName>
		</author>
		<idno type="DOI">10.1007/s10462-021-10018-y</idno>
	</analytic>
	<monogr>
		<title level="j">Artificial Intelligence Review</title>
		<imprint>
			<biblScope unit="volume">54</biblScope>
			<biblScope unit="page" from="5205" to="5253" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Schliebitz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Graf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Wamhof</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Tapken</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gertzen</surname></persName>
		</author>
		<title level="m">KI-basiertes Computer-Vision-System zur Qualitäts-und Größenbestimmung von Kartoffeln</title>
				<imprint>
			<publisher>Resiliente Agri-Food-Systeme</publisher>
			<date type="published" when="2023">2023</date>
			<biblScope unit="volume">43</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<author>
			<persName><forename type="first">D</forename><surname>Pelliccia</surname></persName>
		</author>
		<ptr target="https://nirpyresearch.com/two-scatter-correction-techniques-nir-spectroscopy-python/" />
		<title level="m">Two scatter correction techniques for NIR spectroscopy in Python</title>
				<imprint>
			<date type="published" when="2018">2018. 2023-07-15</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Hyperspectral image analysis. A tutorial</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Amigo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Babamoradi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename></persName>
		</author>
		<idno type="DOI">10.1016/j.aca.2015.09.030</idno>
	</analytic>
	<monogr>
		<title level="j">Analytica Chimica Acta</title>
		<imprint>
			<biblScope unit="volume">896</biblScope>
			<biblScope unit="page" from="34" to="51" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Kernel-based methods for hyperspectral image classification</title>
		<author>
			<persName><forename type="first">G</forename><surname>Camps-Valls</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Bruzzone</surname></persName>
		</author>
		<idno type="DOI">10.1109/TGRS.2005.846154</idno>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Geoscience and Remote Sensing</title>
		<imprint>
			<biblScope unit="volume">43</biblScope>
			<biblScope unit="page" from="1351" to="1362" />
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Convolutional neural networks for hyperspectral image classification</title>
		<author>
			<persName><forename type="first">S</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Jia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Xu</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.neucom.2016.09.010</idno>
	</analytic>
	<monogr>
		<title level="j">Neurocomputing</title>
		<imprint>
			<biblScope unit="volume">219</biblScope>
			<biblScope unit="page" from="88" to="98" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<author>
			<persName><forename type="first">N</forename><surname>Audebert</surname></persName>
		</author>
		<ptr target="https://github.com/nshaud/DeepHyperX" />
		<title level="m">DeepHyperX, Online</title>
				<imprint>
			<date type="published" when="2018">2018. 2023-07-15</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Deep Learning for Classification of Hyperspectral Data: A Comparative Review</title>
		<author>
			<persName><forename type="first">N</forename><surname>Audebert</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">Le</forename><surname>Saux</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Lefèvre</surname></persName>
		</author>
		<idno type="DOI">10.1109/MGRS.2019.2912563</idno>
	</analytic>
	<monogr>
		<title level="j">IEEE Geoscience and Remote Sensing Magazine</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="page" from="159" to="173" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<monogr>
		<title level="m" type="main">Hyperspectral CNN for Image Classification &amp; Band Selection, with Application to Face Recognition</title>
		<author>
			<persName><forename type="first">V</forename><surname>Sharma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Diba</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Tuytelaars</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Van Gool</surname></persName>
		</author>
		<idno>KUL/ESAT/PSI/1604</idno>
		<imprint>
			<date type="published" when="2016">2016</date>
			<pubPlace>Leuven, Belgium</pubPlace>
		</imprint>
		<respStmt>
			<orgName>KU Leuven, ESAT</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Technical report</note>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Outlier Selection and One-Class Classification</title>
		<author>
			<persName><forename type="first">J</forename><surname>Janssens</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Series: TiCC Ph.D. Series</title>
		<imprint>
			<biblScope unit="volume">27</biblScope>
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
	<note type="report_type">Ph.D. thesis</note>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">COPOD: Copula-Based Outlier Detection</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Botta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Ionescu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Hu</surname></persName>
		</author>
		<idno type="DOI">10.1109/ICDM50108.2020.00135</idno>
	</analytic>
	<monogr>
		<title level="m">IEEE International Conference on Data Mining (ICDM)</title>
				<imprint>
			<date type="published" when="2020">2020. 2020</date>
			<biblScope unit="page" from="1118" to="1123" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Adaptive multiple-band CFAR detection of an optical pattern with unknown spectral distribution</title>
		<author>
			<persName><forename type="first">I</forename><surname>Reed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Yu</surname></persName>
		</author>
		<idno type="DOI">10.1109/29.60107</idno>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Acoustics, Speech, and Signal Processing</title>
		<imprint>
			<biblScope unit="volume">38</biblScope>
			<biblScope unit="page" from="1760" to="1770" />
			<date type="published" when="1990">1990</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Smoothing and Differentiation of Data by Simplified Least Squares Procedures</title>
		<author>
			<persName><forename type="first">A</forename><surname>Savitzky</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">J E</forename><surname>Golay</surname></persName>
		</author>
		<idno type="DOI">10.1021/ac60214a047</idno>
	</analytic>
	<monogr>
		<title level="j">Analytical Chemistry</title>
		<imprint>
			<biblScope unit="volume">36</biblScope>
			<biblScope unit="page" from="1627" to="1639" />
			<date type="published" when="1964">1964</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Ten Lectures on Wavelets</title>
		<author>
			<persName><forename type="first">I</forename><surname>Daubechies</surname></persName>
		</author>
		<idno type="DOI">10.1137/1.9781611970104</idno>
	</analytic>
	<monogr>
		<title level="j">SIAM</title>
		<imprint>
			<date type="published" when="1992">1992</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">A Transformation for Ordering Multispectral Data in Terms of Image Quality with Implications for Noise Removal</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">A</forename><surname>Green</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Berman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Switzer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">D</forename><surname>Craig</surname></persName>
		</author>
		<idno type="DOI">10.1109/36.3001</idno>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on geoscience and remote sensing</title>
		<imprint>
			<biblScope unit="volume">26</biblScope>
			<biblScope unit="page" from="65" to="74" />
			<date type="published" when="1988">1988</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">Correction to the Description of Standard Normal Variate (SNV) and De-Trend (DT) Transformations</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">J</forename><surname>Barnes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">S</forename><surname>Dhanoa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">J</forename><surname>Lister</surname></persName>
		</author>
		<idno type="DOI">10.1255/jnirs.21</idno>
	</analytic>
	<monogr>
		<title level="j">NIR news</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="page" from="6" to="6" />
			<date type="published" when="1994">1994</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">The Effect of Multiplicative Scatter Correction (MSC) and Linearity Improvement in NIR Spectroscopy</title>
		<author>
			<persName><forename type="first">T</forename><surname>Isaksson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Naes</surname></persName>
		</author>
		<idno type="DOI">10.1366/0003702884429869</idno>
	</analytic>
	<monogr>
		<title level="j">Applied Spectroscopy</title>
		<imprint>
			<biblScope unit="volume">42</biblScope>
			<biblScope unit="page" from="1273" to="1284" />
			<date type="published" when="1988">1988</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<analytic>
		<title level="a" type="main">On lines and planes of closest fit to systems of points in space</title>
		<author>
			<persName><forename type="first">K</forename><surname>Pearson</surname></persName>
		</author>
		<author>
			<persName><surname>Liii</surname></persName>
		</author>
		<idno type="DOI">10.1080/14786440109462720</idno>
	</analytic>
	<monogr>
		<title level="j">The London, Edinburgh, and Dublin philosophical magazine and journal of science</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="559" to="572" />
			<date type="published" when="1901">1901</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<analytic>
		<title level="a" type="main">New hyperspectral discrimination measure for spectral characterization</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Du</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C.-I</forename><surname>Chang</surname></persName>
		</author>
		<idno type="DOI">10.1117/1.1766301</idno>
	</analytic>
	<monogr>
		<title level="j">Optical engineering</title>
		<imprint>
			<biblScope unit="volume">43</biblScope>
			<biblScope unit="page" from="1777" to="1786" />
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b31">
	<analytic>
		<title level="a" type="main">The Link between Multiplicative Scatter Correction (MSC) and Standard Normal Variate (SNV) Transformations of NIR Spectra</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">S</forename><surname>Dhanoa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">J</forename><surname>Lister</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Sanderson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">J</forename><surname>Barnes</surname></persName>
		</author>
		<idno type="DOI">10.1255/jnirs.30</idno>
	</analytic>
	<monogr>
		<title level="j">J. Near Infrared Spectrosc</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="43" to="47" />
			<date type="published" when="1994">1994</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b32">
	<analytic>
		<title level="a" type="main">Explanation and Justification in Machine Learning: A Survey</title>
		<author>
			<persName><forename type="first">O</forename><surname>Biran</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Cotton</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IJCAI-17 Workshop on Explainable AI</title>
				<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b33">
	<analytic>
		<title level="a" type="main">Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI</title>
		<author>
			<persName><forename type="first">A</forename><surname>Barredo Arrieta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Díaz-Rodríguez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Del</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Ser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Bennetot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Tabik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Barbado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Garcia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Gil-Lopez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Molina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Benjamins</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Chatila</surname></persName>
		</author>
		<author>
			<persName><surname>Herrera</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.inffus.2019.12.012</idno>
	</analytic>
	<monogr>
		<title level="j">Information Fusion</title>
		<imprint>
			<biblScope unit="volume">58</biblScope>
			<biblScope unit="page" from="82" to="115" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b34">
	<analytic>
		<title level="a" type="main">Data quality and explainable AI</title>
		<author>
			<persName><forename type="first">L</forename><surname>Bertossi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Geerts</surname></persName>
		</author>
		<idno type="DOI">10.1145/3386687</idno>
	</analytic>
	<monogr>
		<title level="j">Journal of Data and Information Quality (JDIQ)</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="page" from="1" to="9" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b35">
	<analytic>
		<title level="a" type="main">Interpretable Machine Learning: A Brief Survey From the Predictive Maintenance Perspective</title>
		<author>
			<persName><forename type="first">S</forename><surname>Vollert</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Atzmueller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Theissler</surname></persName>
		</author>
		<idno type="DOI">10.1109/ETFA45728.2021.9613467</idno>
	</analytic>
	<monogr>
		<title level="m">Proc. IEEE International Conference on Emerging Technologies and Factory Automation (ETFA 2021)</title>
				<meeting>IEEE International Conference on Emerging Technologies and Factory Automation (ETFA 2021)</meeting>
		<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b36">
	<analytic>
		<title level="a" type="main">Declarative Aspects in Explicative Data Mining for Computational Sensemaking</title>
		<author>
			<persName><forename type="first">M</forename><surname>Atzmueller</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-030-00801-7_7</idno>
	</analytic>
	<monogr>
		<title level="m">Proc. International Conference on Declarative Programming</title>
				<editor>
			<persName><forename type="first">D</forename><surname>Seipel</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Hanus</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Abreu</surname></persName>
		</editor>
		<meeting>International Conference on Declarative Programming<address><addrLine>Heidelberg, Germany</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="97" to="114" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b37">
	<analytic>
		<title level="a" type="main">From machine learning to explainable AI</title>
		<author>
			<persName><forename type="first">A</forename><surname>Holzinger</surname></persName>
		</author>
		<idno type="DOI">10.1109/DISA.2018.8490530</idno>
	</analytic>
	<monogr>
		<title level="m">2018 world symposium on digital intelligence for systems and machines (DISA)</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="55" to="66" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b38">
	<analytic>
		<title level="a" type="main">ZeroC: A Neuro-Symbolic Model for Zero-shot Concept Recognition and Acquisition at Inference Time</title>
		<author>
			<persName><forename type="first">T</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Tjandrasuwita</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Sosic</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Leskovec</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2206.15049</idno>
	</analytic>
	<monogr>
		<title level="j">Advances in Neural Information Processing Systems</title>
		<imprint>
			<biblScope unit="volume">35</biblScope>
			<biblScope unit="page" from="9828" to="9840" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
