<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Online Explainable Forecasting using Regions of Competence</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Amal</forename><surname>Saadallah</surname></persName>
							<email>amal.saadallah@cs.tu-dortmund.de</email>
							<affiliation key="aff0">
								<orgName type="department">Lamarr Institute for Machine Learning and Artificial Intelligence</orgName>
								<orgName type="institution">TU Dortmund University</orgName>
								<address>
									<settlement>Dortmund</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Matthias</forename><surname>Jakobs</surname></persName>
							<email>matthias.jakobs@tu-dortmund.de</email>
							<affiliation key="aff0">
								<orgName type="department">Lamarr Institute for Machine Learning and Artificial Intelligence</orgName>
								<orgName type="institution">TU Dortmund University</orgName>
								<address>
									<settlement>Dortmund</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Online Explainable Forecasting using Regions of Competence</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">2ECF9DCF0C7691DBFA54ACC4DD8572C4</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T18:54+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Time Series Forecasting</term>
					<term>Explainability</term>
					<term>Model Selection</term>
					<term>Ensembling</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Several machine learning models have been applied to time series forecasting. However, it is generally acknowledged that none of these models is universally valid for every application and over time. This is due to the complex and changing nature of time series data that may involve non-stationary processes but can also be explained by the fact that different models have varying expected regions of expertise or so-called Regions of Competence (RoCs) over the time series. Therefore, adequate and adaptive online model selection is often required. In this work, we review online model selection works that exploit the notion of RoCs by summarizing their methods and highlighting their strengths as well as limitations. In particular, we present a taxonomy of these methods and show how they can be exploited for both single model selection and ensemble of models pruning. We additionally discuss how the RoCs can promote explainability. Finally, we suggest future directions to provide useful research guidance.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Time series forecasting is considered a key step in making informed decisions in a wide range of applications <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b1">2,</ref><ref type="bibr" target="#b2">3]</ref>. Several Machine Learning (ML) models have been applied to the time series forecasting task <ref type="bibr" target="#b3">[4,</ref><ref type="bibr" target="#b4">5,</ref><ref type="bibr" target="#b5">6]</ref>. While certain guidelines, such as task complexity or data size, can help select a suitable family of ML models for forecasting <ref type="bibr" target="#b6">[7]</ref>, it is generally acknowledged that no existing ML model is universally applicable to all forecasting problems. This observation aligns with Wolpert's No Free Lunch theorem <ref type="bibr" target="#b7">[8]</ref>, suggesting that no learning algorithm can be optimal for all learning tasks. In addition, ML models may demonstrate a time-dependent performance <ref type="bibr" target="#b8">[9,</ref><ref type="bibr" target="#b9">10]</ref>. This means their accuracy may not remain consistent over time due to the dynamic nature of time series data, which may involve non-stationary processes and, as a result, be subject to the so-called concept drift phenomenon <ref type="bibr" target="#b10">[11]</ref>. Thus, it is clear that different forecasting models have different expected areas of expertise placed over different parts in the input time series. The part where a specific model outperforms certain candidate models is referred to as a Region of Competence (RoC) of that model <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b11">12,</ref><ref type="bibr" target="#b8">9]</ref>.</p><p>Several works in the ML literature have used this notion of RoCs either implicitly <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b12">13,</ref><ref type="bibr" target="#b11">12]</ref> or explicitly <ref type="bibr" target="#b8">[9,</ref><ref type="bibr" target="#b9">10,</ref><ref type="bibr" target="#b13">14,</ref><ref type="bibr" target="#b14">15,</ref><ref type="bibr" target="#b15">16]</ref> to perform online model selection for forecasting that copes with the timeevolving nature of time series and the fact that models have a certain expected level of competence in predicting a particular region in the time series. While some of these works have focused on the online selection of a single model <ref type="bibr" target="#b8">[9,</ref><ref type="bibr" target="#b14">15,</ref><ref type="bibr" target="#b15">16]</ref>, others have extended on the assumption that no single model is an expert all the time, and have proposed to adaptively select and combine multiple forecasting models into a single model using an ensemble technique <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b16">17,</ref><ref type="bibr" target="#b11">12,</ref><ref type="bibr" target="#b9">10,</ref><ref type="bibr" target="#b13">14]</ref>. The selection of individual models that make up the ensemble is referred to as ensemble pruning <ref type="bibr" target="#b17">[18]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Regions of Competence Computation</head><p>Let 𝑋 be a univariate time-series and let 𝑋 𝑡 be the value of that time-series at time 𝑡. Additionally, we denote with 𝑋 𝑎:𝑏 the subsequence entailing values from 𝑡 = 𝑎 to 𝑡 = 𝑏 with 𝑎 &lt; 𝑏. A Region of Competence (RoC) for a forecasting model 𝑓 ∈ P, where P is a set of given models, refers to a set of subsequence 𝑋 𝑎:𝑏 where 𝑓 exhibits superior performance compared to all other models in P.However, pinpointing these subsequences and determining their optimal length through empirical evaluations is computationally infeasible. Recent literature has leveraged machine learning to approximate these RoCs. These approaches fall into two main categories. The first category includes model-agnostic methods that compute RoCs independent of a family of forecasting models. Within this category, two primary machine learning paradigms have been used, namely pattern matching and meta-learning. The second category comprises model-specific methods tailored to a particular class of forecasting models, such as tree-based models or DNNs.</p><p>The main idea of pattern matching methods is to, using the current subsequence, find the closest subsequence in the training or validation data. The model exhibiting the best performance on these patterns is considered as the expert and this recent subsequence is treated as one of its RoCs. In <ref type="bibr" target="#b11">[12]</ref>, 1-Nearest Neighbors is used to determine the closest pattern. In <ref type="bibr" target="#b15">[16]</ref>, clustering is used to cluster the known datapoints first. Then, each model in P is evaluated on all subsequences from each cluster. The Mean Squared Error (MSE) for each model is computed and each cluster is assigned as Regions of Competence (RoC) of the model that minimizes the computed MSE. This ensures that each cluster is associated with an expert model that excels in predicting values based on the dominant pattern in the corresponding cluster.</p><p>In <ref type="bibr" target="#b16">[17,</ref><ref type="bibr" target="#b4">5]</ref>, meta-learning is used to build models capable of modeling the competence of each candidate model across the input time series. Their meta-learning approach is composed of an arbitrating architecture <ref type="bibr" target="#b18">[19]</ref> and a mixture of experts <ref type="bibr" target="#b19">[20]</ref>. A meta-learner is created for each candidate model. While each model 𝑓 𝑖 is trained to model the future values of the time series, its meta-learning associate 𝑔 𝑖 is trained to model the error of 𝑓 𝑖 . The arbiter 𝑔 𝑖 then can make predictions regarding the error that will incur when predicting the future values of the time series. At test time, the candidate models 𝑓 𝑖 are weighted according to their expected degree of competence in 𝑋 𝑡−𝑝:𝑡−1 , estimated by the predictions of the meta-learners. These methods can be viewed as an implicit exploitation of the RoC concept as they do not result in some identified subsequences in the input time series but rather a modeling of the competence of the models using error estimates.</p><p>The works in this category mainly focus on CNN-based models, comprising 1D-convolutional layers with varied filter and kernel sizes <ref type="bibr" target="#b20">[21,</ref><ref type="bibr" target="#b9">10]</ref>. More recently, in <ref type="bibr" target="#b13">[14]</ref>, a hybrid architecture is proposed by concatenating the convolutional layers to a set of heterogeneous models that can be trained jointly using a gradient-based approach. However, these works use the same principle to compute the RoCs. They suggest a modified version of Grad-CAM <ref type="bibr" target="#b21">[22]</ref>, which utilizes spatial information preserved in convolutional layers to identify important regions of the input. To do so, they start by evaluating the Mean Squared Error on a specific validation time window 𝑋 𝑣𝑎𝑙 , denoted as 𝜁 𝑓 𝑖 , for model 𝑓 𝑖 . The objective is to determine the significance of each time point in this window with respect to the computed error. The last feature maps layer of the CNN is utilized for this purpose. Importance weights 𝛼 𝑓 𝑖 associated with 𝜁 𝑓 𝑖 are computed for each activation unit 𝑢 in each generic feature map 𝐴 by calculating the gradient of 𝜁 𝑓 𝑖 relative to 𝐴. Finally, a global average is computed over all units in</p><formula xml:id="formula_0">𝐴: 𝛼 𝑓 𝑖 = 1 𝑈 ∑︀ 𝑢 𝜕𝜁 𝑓 𝑖</formula><p>𝜕𝐴𝑢 , where 𝑈 is the total number of units in 𝐴. We use 𝛼 𝑓 𝑖 to compute a weighted combination between all the feature maps for a given measured value of the error 𝜁 𝑓 𝑖 . Since we are mainly interested in highlighting temporal features contributing most to 𝜁 𝑓 𝑖 , ReLU is used to remove all the negative contributions by: 𝐿 𝑓 𝑖 = 𝑅𝑒𝐿𝑈 ( ∑︀ 𝛼 𝑓 𝑖 𝐴). To identify the RoC in 𝑋 𝑣𝑎𝑙 that primarily contributed to 𝜁 𝑓 𝑖 , 𝐿 𝑓 𝑖 ∈ R 𝑈 is used. Note that 𝑈 is chosen such that 𝑈 is smaller than 𝑋 𝑣𝑎𝑙 size. Note that multiple time windows are created from 𝑋 𝑣𝑎𝑙 using a time-sliding approach to evaluate the performance of the same model on different windows and increase the number of computed RoCs. While the methods presented in the previous section proved successful in providing an explanation for the model selection and ensembling processes, the problem of opaque, non-interpretable base models still persists. Recent work <ref type="bibr" target="#b14">[15]</ref> utilized tree-based models to make the model decision process more transparent in addition to having an interpretable model selection algorithm. The authors create a pool of Decision Trees, Random Forests, and Gradient Boosting Trees that are subsequently trained on 𝑋 𝑡𝑟𝑎𝑖𝑛 . Shapley values, a well-established Explainability method for feature attribution, are used to generate the explanations necessary for the creation of RoCs. Since computing Shapley values is in general NP-hard <ref type="bibr" target="#b22">[23]</ref>, the authors utilize TreeSHAP <ref type="bibr" target="#b23">[24]</ref>, which is an estimation method designed for tree-based models that is able to estimate Shapley values in polynomial time. Similar to <ref type="bibr" target="#b8">[9]</ref>, the loss of the prediction for each model is explained rather than the prediction itself.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Online Model Selection</head><p>At test time 𝑡, after computing the RoCs of the candidate models for selection, an online decision on model selection for forecasting the value of 𝑋 at 𝑡 has to be made. Note that the following can be applied to any future time instant 𝑡 + ℎ, ℎ ≥ 0. For simplicity of notations, assume ℎ = 0. In this part, we focus on the works that explicitly compute the RoCs, i.e., result in identified subsequences within the input time series.</p><p>The models in P are devised such that they use the same 𝑝-lagged values of the time series as input. As a result, at time 𝑡, the input subsequence is 𝑋 𝑡−𝑝:𝑡−1 , 𝑡 ≥ 𝑝. To perform the selection, the distance of the input subsequence 𝑋 𝑡−𝑝:𝑡−1 to the RoCs of each candidate model is measured <ref type="bibr" target="#b8">[9,</ref><ref type="bibr" target="#b11">12,</ref><ref type="bibr" target="#b15">16,</ref><ref type="bibr" target="#b14">15]</ref>. Since the length of each RoC can be different from 𝑝 (i.e. length of 𝑋 𝑡−𝑝:𝑡−1 ), Dynamic Time Wrapping (DTW) <ref type="bibr" target="#b24">[25]</ref> is generally used to measure the similarity between 𝑋 𝑡−𝑝:𝑡−1 and each RoC member. The model with a RoC member of smallest DTW distance to 𝑋 𝑡−𝑝:𝑡−1 is selected for forecasting. In <ref type="bibr" target="#b15">[16]</ref>, the RoCs are smoothed using a moving-average process and filtered to keep only the RoCs with length 𝑝, and DTW is replaced by Euclidean Distance.</p><p>Recent works have extended the RoC concept for single model selection to ensemble pruning <ref type="bibr" target="#b15">[16,</ref><ref type="bibr" target="#b13">14,</ref><ref type="bibr" target="#b9">10]</ref>. They based their reasoning on the expected error of the ensemble that can be decomposed at each time point into a weighted average error term of the individual composing models and an ambiguity term, which is simply the variance of the ensemble around the weighted mean of its composing models. The weighted error term reflects the ensemble accuracy, while the ambiguity term reflects the ensemble diversity <ref type="bibr" target="#b25">[26,</ref><ref type="bibr" target="#b26">27]</ref>. Since one model can have multiple RoCs, the first step is to select one representative RoC for each model based on the topological closeness to the current input subsequence 𝑋 𝑡−𝑝:𝑡−1 . In this manner, we avoid selecting the same model multiple times. This contributes to the ensemble diversity <ref type="bibr" target="#b15">[16,</ref><ref type="bibr" target="#b13">14]</ref>. In <ref type="bibr" target="#b9">[10]</ref>, diversity is promoted further by clustering the RoCs representatives so that selecting models with similar competencies or expertise is avoided. The second step consists of ranking the models using the distance of their representative RoCs to the current input subsequence 𝑋 𝑡−𝑝:𝑡−1 to promote accuracy. The top-𝑀 closest models are selected to make up the ensemble. The top-𝑀 is arbitrarily set to a fixed number in <ref type="bibr" target="#b15">[16]</ref>. The authors in <ref type="bibr" target="#b9">[10]</ref> derived bounds for both error types to set the top-𝑀 number automatically.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Explainability Aspects</head><p>Most studies mentioned earlier emphasized the importance of transparency in models and learning decisions, including the selection of models <ref type="bibr" target="#b8">[9,</ref><ref type="bibr" target="#b9">10,</ref><ref type="bibr" target="#b15">16,</ref><ref type="bibr" target="#b13">14,</ref><ref type="bibr" target="#b14">15]</ref>. They demonstrated that the computed Regions of Competence serve as a valuable tool for explaining why a specific forecast value is generated at a given time and for choosing a particular individual model or ensemble member within a specific time frame or interval. In essence, these RoCs act as an explanatory mechanism, shedding light on the decision-making process behind forecast outputs and model selections. More specific, they are part of the so-called exemplar-based or prototype-based explainability approaches <ref type="bibr" target="#b27">[28,</ref><ref type="bibr" target="#b28">29]</ref>. Thus, local explanations, i.e., why a certain model was chosen at time 𝑡 can be given by visualy inspecting the Visualization of RoCs for AbnormalHeartbeat data using the DNN-base approach for RoCs computation <ref type="bibr" target="#b8">[9]</ref>. Notice that some RoCs can be empty, reflecting that this model did not outperform the anothers at any point during validation.</p><p>closest RoC member for each model. Global insights can be gained by visualizing the (clustered) RoCs for all models, as seen in Figure <ref type="figure">1</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Conclusion</head><p>This work provides a unified view of several methods for online adaptive time series forecasting methods that are based on Regions of Competence. These regions not only enable state-of-the-art selection and ensemble performance but can also be used to gain insights into the selection and ensembling process on a local, as well as a global level. One avenue of further research could be concerned with utilizing model-agnostic explainability methods such as KernelSHAP to enable heterogeneous model selection and ensembling. However, such a method may be limited by runtime, which would need to be improved to allow for concept drift adaption. Another direction of future work might be to improve overall explainability by limiting the model pool to small, transparent models. An open question remains if it is possible to select or ensemble these small models using Regions of Competence in a way that reaches state-of-the-art performance compared to incomprehensible methods such selecting from pools of Deep Neural Networks.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>11 Figure 1 :</head><label>111</label><figDesc>Figure1: Visualization of RoCs for AbnormalHeartbeat data using the DNN-base approach for RoCs computation<ref type="bibr" target="#b8">[9]</ref>. Notice that some RoCs can be empty, reflecting that this model did not outperform the anothers at any point during validation.</figDesc></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>This research has been funded by the Federal Ministry of Education and Research of Germany and the state of North Rhine-Westphalia as part of the Lamarr Institute for Machine Learning and Artificial Intelligence.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">25 years of time series forecasting</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">G</forename><surname>De Gooijer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">J</forename><surname>Hyndman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International journal of forecasting</title>
		<imprint>
			<biblScope unit="volume">22</biblScope>
			<biblScope unit="page" from="443" to="473" />
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Bright-drift-aware demand predictions for taxi networks</title>
		<author>
			<persName><forename type="first">A</forename><surname>Saadallah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Moreira-Matias</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Sousa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Khiari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Jenelius</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gama</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Knowledge and Data Engineering</title>
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Monash time series forecasting archive</title>
		<author>
			<persName><forename type="first">R</forename><surname>Godahewa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Bergmeir</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">I</forename><surname>Webb</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">J</forename><surname>Hyndman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Montero-Manso</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Neural Information Processing Systems Track on Datasets and Benchmarks</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
	<note>Forthcoming</note>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">An empirical comparison of machine learning models for time series forecasting</title>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">K</forename><surname>Ahmed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">F</forename><surname>Atiya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">E</forename><surname>Gayar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>El-Shishiny</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Econometric reviews</title>
		<imprint>
			<biblScope unit="volume">29</biblScope>
			<biblScope unit="page" from="594" to="621" />
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Arbitrated ensemble for time series forecasting</title>
		<author>
			<persName><forename type="first">V</forename><surname>Cerqueira</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Torgo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Pinto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Soares</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Joint European conference on machine learning and knowledge discovery in databases</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page" from="478" to="494" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Time-series forecasting with deep learning: a survey</title>
		<author>
			<persName><forename type="first">B</forename><surname>Lim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Zohren</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Philosophical Transactions of the Royal Society A</title>
		<imprint>
			<biblScope unit="volume">379</biblScope>
			<biblScope unit="page">20200209</biblScope>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<author>
			<persName><forename type="first">V</forename><surname>Cerqueira</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Torgo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Soares</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1909.13316</idno>
		<title level="m">Machine learning vs statistical methods for time series forecasting: Size matters</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">The lack of a priori distinctions between learning algorithms</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">H</forename><surname>Wolpert</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Neural computation</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="page" from="1341" to="1390" />
			<date type="published" when="1996">1996</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Explainable online deep neural network selection using adaptive saliency maps for time series forecasting</title>
		<author>
			<persName><forename type="first">A</forename><surname>Saadallah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Jakobs</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Morik</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Machine Learning and Knowledge Discovery in Databases</title>
				<editor>
			<persName><forename type="first">N</forename><surname>Oliver</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><surname>Pérez-Cruz</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Kramer</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Read</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>Lozano</surname></persName>
		</editor>
		<meeting><address><addrLine>Cham</addrLine></address></meeting>
		<imprint>
			<publisher>Springer International Publishing</publisher>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="404" to="420" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Explainable online ensemble of deep neural network pruning for time series forecasting</title>
		<author>
			<persName><forename type="first">A</forename><surname>Saadallah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Jakobs</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Morik</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Machine Learning</title>
		<imprint>
			<biblScope unit="volume">111</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">A survey on concept drift adaptation</title>
		<author>
			<persName><forename type="first">J</forename><surname>Gama</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Žliobaitė</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bifet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Pechenizkiy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bouchachia</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACM computing surveys (CSUR)</title>
		<imprint>
			<biblScope unit="volume">46</biblScope>
			<biblScope unit="page" from="1" to="37" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title level="m" type="main">Dynamic model selection for automated machine learning in time series</title>
		<author>
			<persName><forename type="first">F</forename><surname>Priebe</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Constructive aggregation and its application to forecasting with dynamic ensembles</title>
		<author>
			<persName><forename type="first">V</forename><surname>Cerqueira</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Pinto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Torgo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Soares</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Moniz</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Joint European Conference on Machine Learning and Knowledge Discovery in Databases</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="620" to="636" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Online deep hybrid ensemble learning for time series forecasting</title>
		<author>
			<persName><forename type="first">A</forename><surname>Saadallah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Jakobs</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Joint European Conference on Machine Learning and Knowledge Discovery in Databases</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="156" to="171" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Explainable Adaptive Tree-based Model Selection for Time-Series Forecasting</title>
		<author>
			<persName><forename type="first">M</forename><surname>Jakobs</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Saadallah</surname></persName>
		</author>
		<idno type="DOI">10.1109/ICDM58522.2023.00027</idno>
	</analytic>
	<monogr>
		<title level="m">2023 IEEE International Conference on Data Mining (ICDM)</title>
				<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="180" to="189" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Online explainable model selection for time series forecasting</title>
		<author>
			<persName><forename type="first">A</forename><surname>Saadallah</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE 10th International Conference on Data Science and Advanced Analytics (DSAA), IEEE</title>
				<imprint>
			<date type="published" when="2023">2023. 2023</date>
			<biblScope unit="page" from="1" to="10" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Arbitrage of forecasting experts</title>
		<author>
			<persName><forename type="first">V</forename><surname>Cerqueira</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Torgo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Pinto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Soares</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Machine Learning</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">An ensemble pruning primer</title>
		<author>
			<persName><forename type="first">G</forename><surname>Tsoumakas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Partalas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Vlahavas</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Applications of supervised and unsupervised ensemble methods</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2009">2009</date>
			<biblScope unit="page" from="1" to="13" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Arbitrating among competing classifiers using learned referees</title>
		<author>
			<persName><forename type="first">J</forename><surname>Ortega</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Koppel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Argamon</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Knowledge and Information Systems</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="page" from="470" to="490" />
			<date type="published" when="2001">2001</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Adaptive mixtures of local experts</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">A</forename><surname>Jacobs</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">I</forename><surname>Jordan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">J</forename><surname>Nowlan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">E</forename><surname>Hinton</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Neural computation</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="page" from="79" to="87" />
			<date type="published" when="1991">1991</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">An actor-critic ensemble aggregation model for time-series forecasting</title>
		<author>
			<persName><forename type="first">A</forename><surname>Saadallah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Tavakol</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Morik</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE ICDE</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Grad-cam: Visual explanations from deep networks via gradient-based localization</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">R</forename><surname>Selvaraju</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Cogswell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Das</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Vedantam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Parikh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Batra</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE international conference on computer vision</title>
				<meeting>the IEEE international conference on computer vision</meeting>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">On the complexity of cooperative solution concepts</title>
		<author>
			<persName><forename type="first">X</forename><surname>Deng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">H</forename><surname>Papadimitriou</surname></persName>
		</author>
		<idno>arXiv:3690220</idno>
	</analytic>
	<monogr>
		<title level="j">Mathematics of Operations Research</title>
		<imprint>
			<biblScope unit="volume">19</biblScope>
			<biblScope unit="page" from="257" to="266" />
			<date type="published" when="1994">1994</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">From local explanations to global understanding with explainable AI for trees</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Lundberg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Erion</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Degrave</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Prutkin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Nair</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Katz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Himmelfarb</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Bansal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S.-I</forename><surname>Lee</surname></persName>
		</author>
		<idno type="DOI">10.1038/s42256-019-0138-9</idno>
	</analytic>
	<monogr>
		<title level="j">Nature Machine Intelligence</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="56" to="67" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Using dynamic time warping to find patterns in time series</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">J</forename><surname>Berndt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Clifford</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">KDD workshop</title>
				<imprint>
			<date type="published" when="1994">1994</date>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="page" from="359" to="370" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Neural network ensembles, cross validation, and active learning</title>
		<author>
			<persName><forename type="first">A</forename><surname>Krogh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Vedelsby</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in Neural Information Processing Systems</title>
				<imprint>
			<publisher>MIT Press</publisher>
			<date type="published" when="1995">1995</date>
			<biblScope unit="volume">7</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Managing Diversity in Regression Ensembles</title>
		<author>
			<persName><forename type="first">G</forename><surname>Brown</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">L</forename><surname>Wyatt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Tiňo</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Machine Learning Research</title>
		<imprint>
			<biblScope unit="volume">6</biblScope>
			<biblScope unit="page" from="1621" to="1650" />
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">A Comprehensive Review on Explainable AI Techniques, Challenges, and Future Scope</title>
		<author>
			<persName><forename type="first">A</forename><surname>Patil</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Patil</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-981-99-3177-4_39</idno>
	</analytic>
	<monogr>
		<title level="m">Intelligent Computing and Networking</title>
				<editor>
			<persName><forename type="first">V</forename><forename type="middle">E</forename><surname>Balas</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">V</forename><forename type="middle">B</forename><surname>Semwal</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Khandare</surname></persName>
		</editor>
		<meeting><address><addrLine>Singapore</addrLine></address></meeting>
		<imprint>
			<publisher>Springer Nature</publisher>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="517" to="529" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<monogr>
		<title level="m" type="main">Interpretable Machine Learning</title>
		<author>
			<persName><forename type="first">C</forename><surname>Molnar</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
