<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Online Explainable Ensemble of Tree Models Pruning for Time Series Forecasting ⋆</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Amal</forename><surname>Saadallah</surname></persName>
							<email>amal.saadallah@tu-dortmund.de</email>
							<affiliation key="aff0">
								<orgName type="institution">Lamarr Institute for Machine Learning and AI</orgName>
								<address>
									<settlement>Dortmund</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Online Explainable Ensemble of Tree Models Pruning for Time Series Forecasting ⋆</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">F23FBC5C73D5EE07D1D0948CEF747EA8</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T19:38+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Tree Models</term>
					<term>Online Ensemble Pruning</term>
					<term>TreeSHAP</term>
					<term>Time Series Forecasting</term>
					<term>Concept-drift</term>
					<term>Explainability</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Tree-based models are commonly used in time series forecasting due to their inherent interpretability, which makes them preferable to more complex black-box models. However, simple tree-based models are prone to overfitting, limiting their applicability in real-world scenarios. Ensembles of tree-based models are employed to mitigate this, but ensemble pruning is challenging, especially in the presence of dynamic time series data and concept drift. In this paper, we use TreeSHAP, a tree-specific explainability tool, to perform online tree-based ensemble pruning that adapts dynamically to changes in the time series, addressing the concept drift issue. Empirical evaluations on real-world time series datasets demonstrate that our method performs on par with or better than state-of-the-art techniques. In future research, we plan to automate the determination of the optimal number of clusters for ensemble pruning by leveraging ensemble properties like diversity, accuracy, and stability. This automation aims to enhance both the flexibility and explainability of the model selection process. Given that this work is in its early stages, we seek feedback and collaboration with experts to create a robust and explainable framework for ensemble-based time series forecasting.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Time series forecasting is crucial for real-time planning and decision-making across various fields like traffic management, weather prediction, and financial markets. However, it is also one of the most challenging tasks due to the complex and dynamic nature of time series data, which often involves non-stationary variations and is susceptible to concept drift <ref type="bibr" target="#b0">[1]</ref>. This makes accurate forecasting inherently difficult, necessitating models that can adapt to changing data patterns <ref type="bibr" target="#b1">[2,</ref><ref type="bibr" target="#b2">3,</ref><ref type="bibr" target="#b3">4,</ref><ref type="bibr" target="#b4">5,</ref><ref type="bibr" target="#b5">6,</ref><ref type="bibr" target="#b6">7]</ref>. Given these challenges, explainability in forecasting models has become increasingly important, especially for safety-critical applications. Tree-based models are often favored for their intrinsic explainability, but identifying appropriate models for specific time series requires adaptability due to time-varying characteristics. Decision Trees and their ensembles, like Random Forests and Gradient-boosted Trees, are commonly used for time series forecasting. However, these models can struggle with dynamic data since they typically operate in a static manner, not inherently considering variations in the underlying time series. In addition, combining multiple models into ensembles can improve forecasting accuracy, but at the cost of explainability. To address these issues, we propose an online ensemble pruning approach for time series forecasting, where the ensemble members are selected based on an adaptive clustering procedure that uses TreeSHAP values to group models with similar modeling paradigms. This methodology not only ensures diversity within the ensemble but also allows for an explainable selection process by indicating which aspects of the time series data contribute most to the predictions.</p><p>In our future research, one key goal is to automatically determine the optimal number of clusters, which corresponds to the ideal number of trees or ensemble members in the ensemble. This would involve using ensemble properties such as diversity, accuracy, and stability to guide the selection of the most suitable cluster count. By automating this process, we aim to improve the ensemble's flexibility and effectiveness in adapting to dynamic time series data. Moreover, we intend to deepen the explainability aspect of our approach by explicitly demonstrating that selecting models based on different TreeSHAP values aligns with distinct modeling paradigms and hypotheses. This could be achieved by visualizing or analyzing how these varying TreeSHAP values translate into different interpretations of the underlying data, providing insights into the rationale behind model selection. Given that this is early-stage work, we plan to engage with experts in the field to exchange ideas and gather feedback. Collaboration with specialists will be instrumental in refining our methodology for selecting the optimal number of trees and enhancing explainability. By incorporating diverse perspectives, we hope to develop a robust and transparent approach that addresses the complexities of time series forecasting while maintaining clarity in model selection and ensemble pruning. This collaborative effort will contribute to building a reliable framework for ensemble-based forecasting, with a particular emphasis on explainability and adaptability.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Methodology</head><p>Our proposed method uses TreeSHAP for online ensemble pruning using model clustering. First, we define the used notation. Second, we describe Shapley values with a focus on TreeSHAP values <ref type="bibr" target="#b7">[8]</ref>. Third, we show how we generate the candidate tree-based models. Finally, we demonstrate how TreeSHAP values are used for model clustering to allow for efficient ensemble pruning and how the whole process is made adaptive to the changes in the time series.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Preliminaries</head><p>A time series 𝑋 is a temporal sequence of values, where are constrained to be positive and sum to one. In addition, it can be seen from the notation that the weights are time-dependent. This is one of the requirements in online ensemble learning, where the weights are required to be set in a timely manner to cope with the dynamic nature of the time series and the time-changing performance of the ensemble members <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b5">6]</ref>. The goal of dynamic online ensemble pruning is to identify the subset of models S ⊂ T that should compose the ensemble at each time step 𝑡 + ℎ such that the expected prediction error of the pruned ensemble is reduced compared the full ensemble 𝑇 ¯T for each forecast.</p><formula xml:id="formula_0">𝑋 1:𝑡 = {𝑥 1 , 𝑥 2 , • • • , 𝑥 𝑡 }</formula><formula xml:id="formula_1">𝑎𝑟𝑔𝑚𝑎𝑥 S⊂T E [︀(︀ 𝑥 𝑡+ℎ − 𝑇 ¯T(𝑥 ^𝑡+ℎ ) )︀ 2 |𝑋 1:𝑡+ℎ−1 ]︀ − E [︀(︀ 𝑥 𝑡+ℎ − 𝑇 ¯S(𝑥 ^𝑡+ℎ ) )︀ 2 |𝑋 1:𝑡+ℎ−1 ]︀<label>(1)</label></formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">TreeSHAP Ensemble Learning</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.1.">Ensemble Pruning</head><p>We divide the time series </p><formula xml:id="formula_2">𝑋 1:𝑡 into 𝑋 𝑡𝑟𝑎𝑖𝑛 𝜔 = {𝑥 1 , 𝑥 2 , • • • , 𝑥 𝑡−𝜔 } and 𝑋 𝑣𝑎𝑙 𝜔 = {𝑥 𝑡−𝜔+1 , 𝑥 𝑡−𝜔+2 , • • • , 𝑥 𝑡 },</formula><formula xml:id="formula_3">𝐼 𝑗 𝑖 = 1 𝜔 𝜔 ∑︁ 𝑘=1 |𝜑 𝑗 𝑖 (𝑥 𝑡−𝜔+𝑘 )|, ∀ 𝑖 ∈ [1, 𝑙 𝑗 ], ∀ 𝑇 𝑗 ∈ 𝒯<label>(2)</label></formula><p>Each model 𝑇 𝑗 ∈ 𝒯 can then be characterized by a vector</p><formula xml:id="formula_4">I 𝑗 = {𝐼 𝑗 1 , 𝐼 𝑗 2 , • • • 𝐼 𝑗 𝑙 𝑗 }.</formula><p>The models can thus be clustered using their SHAP-based lag importance vectors I 𝑗 . However, different models in 𝒯 might be trained using different lag values. As a result, the length of the vectors I 𝑗 can vary between 𝑙 𝑚𝑖𝑛 and 𝑙 𝑚𝑎𝑥 . It exists clustering distance measure that can handle vectors of different lengths <ref type="bibr" target="#b8">[9]</ref>. However, we are mainly interested in grouping models based on the way they represent the relationship between the input lagged values and the output. Therefore, we assume that the models that are trained using a lag value 𝑙 𝑗 lower than 𝑙 𝑚𝑎𝑥 ignore the importance and the contribution of lagged features that are greater than 𝑙 𝑗 . In other words, if the mode 𝑇 𝑗 is trained on 𝑙 𝑗 ≤ 𝑙 𝑚𝑎𝑥 , for each 𝑖, such that 𝑙 𝑗 ≤ 𝑖 ≤ 𝑙 𝑚𝑎𝑥 , the value of its corresponding SHAP-based lag importance 𝐼 𝑗 𝑖 on 𝑖 is set to zero. In this manner, we bring all the vectors I 𝑗 for all the models 𝑇 𝑗 ∈ 𝒯 to the same length 𝑙 𝑚𝑎𝑥 , and we use K-means with Euclidean distance for model clustering. Models belonging to different clusters are expected to have different modeling paradigms of the contributions of different lagged values to the predictions, which contributes to boosting the ensemble diversity. We select only cluster representatives to take part in the ensemble. We simply select the closest model to each cluster center.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.2.">Ensemble Adaptation</head><p>Streaming time series data is prone to significant changes, leading to concept drifts . To account for these shifts, the selection of ensemble members must be updated, allowing for the inclusion of models that can better address newly emerging patterns. Concept drift is detected by monitoring deviations in the mean of the time series over time, using the Hoeffding Bound to evaluate if these deviations are significant. If a drift is detected, an alarm is triggered, the TreeSHAP-based model clustering is updated, and the ensemble is adjusted to reflect the new patterns in the data.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Experiments</head><p>Our method is denoted in the following as OEP-TT: Online explainable Ensemble Pruning of Tree models for Time series forecasting.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Experimental Setup</head><p>We use 100 univariate time series datasets from various application domains, including financial, weather, and synthetic data. These datasets are provided by the Monash Forecasting Repository <ref type="bibr" target="#b9">[10]</ref>. We process each time series 𝑋 by using the first 50% for training (𝑋 𝑡𝑟𝑎𝑖𝑛 𝜔 ), the following 25% for validation (𝑋 𝑣𝑎𝑙 𝜔 ) and the remaining 25% for testing. Due to this way of splitting the time series, we discard series that are shorter than 250 to allow enough training and validation data. All experiments have been performed on consumer hardware, namely on a 2022 MacBook Pro in R.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">OEP-TT Setup</head><p>Tree-based models set-up: We construct a pool 𝒯 of tree-based models using different parameter settings that are summarized in Table <ref type="table" target="#tab_2">1</ref>. The list of parameters and their value ranges Hyper-parameters values of the tree-based models. Different configurations are generated by taking some combinations of these hyper-parameters as described in the last column.</p><p>in Table <ref type="table" target="#tab_2">1</ref> is not exhaustive, and further parameters and values can be considered to generate more base learners. We also vary the lag parameter 𝑙 on which the tree-based models are trained, i.e., 𝑙 ∈ {3, 5, 7, 10, 15, 20}. Considering different combinations of all the parameters, we train a total of 294 tree-based models.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>OEP-TT set-up:</head><p>OEP-TT has also a number hyper-parameters: 𝑀 is the Size of the Pool of the tree-based models T: 294, 𝜔 is the Size of the validation time window :25% of the data length, |𝑆| is the Number of final selected models: 6.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.">State-of-the-Art Methods Setup</head><p>We compare OEP-TT against State-of-the-Art (SoA) methods for online ensemble pruning, treebased ensembles, and time series forecasting in general. These models include: Auto-Regressive Integrated Moving Average (ARIMA) <ref type="bibr" target="#b10">[11]</ref>, Exponential Smoothing (ETS) <ref type="bibr" target="#b10">[11]</ref>, Long Short-Term Memory (LSTM) <ref type="bibr" target="#b11">[12]</ref>, Multi-Layer Perceptron (MLP) <ref type="bibr" target="#b11">[12]</ref>, Convolutional Neural Network with LSTM (CNN-LSTM, Bi-LSTM) <ref type="bibr" target="#b12">[13]</ref>, Random Forest (RF) <ref type="bibr" target="#b13">[14]</ref>, Gradient-Boosted Decisions Trees (GBDT) <ref type="bibr" target="#b14">[15]</ref>, eXtrem Gradient Boosting (XGBoost) <ref type="bibr" target="#b15">[16]</ref>, and Light Gradient-Boosting Machine (LGBM) <ref type="bibr" target="#b16">[17]</ref>.</p><p>To enable a fair comparison with OEP-TT, we feed to these ensemble pruning methods the same pool of tree-based models T that was used for OEP-TT: Ens: Ensemble of all the base modes in T; OCL <ref type="bibr" target="#b4">[5]</ref>: Online drift-aware clustering of the tree-based models in T using covariance-based clustering; OTOP <ref type="bibr" target="#b4">[5]</ref>: Online drift-aware Top best-performing tree-based models ranking using temporal correlation analysis; DEMSC <ref type="bibr" target="#b4">[5]</ref>: Dynamic Ensemble Members Selection using Clustering: Online drift-aware Top best-performing models ranking using temporal correlation analysis combined with covariance-based clustering; ADE <ref type="bibr" target="#b17">[18,</ref><ref type="bibr" target="#b18">19]</ref> was recently developed for an online dynamic ensemble of forecasters construction. A meta-learning strategy that specializes the tree-based models across the input time series. A sequential weighting schema is developed to automatically select ensemble members by setting their weights to zero.</p><p>We also compare OEP-TT to its variants: OEP-TT-ST: Static variant of OEP-TT. Pruning is decided at the initial forecasting instant and kept fixed along testing; OEP-TT-Per: Pruning is updated periodically in a blind manner (i.e. without considering the occurrence of the drift).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4.">Results</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4.1.">Predictive Performance</head><p>Table <ref type="table" target="#tab_4">2</ref> presents the average ranks and their deviation for OEP-TT and its variants and SoA methods for time series forecasting and online ensemble pruning. For the paired comparison, we compare our method OEP-TT against each of the other methods. We counted wins and losses for each dataset using the RMSE scores. We use the non-parametric Wilcoxon Signed Rank test to compute significant wins and losses, which are presented in parenthesis (significance level 0.05). In the results in Table <ref type="table" target="#tab_4">2</ref>, OEP-TT outperforms almost all the baseline methods in terms of ranks and wins/loses in pairwise comparison.</p><p>In this part, we show how initially OEP-TT supports explainability for the reason behind specific tree-based model selection to construct the ensemble at a specific time instant or interval, for model performance, and for the importance of input lagged time series observations.  Comparison (in terms of average rank achieved over 100 datasets) between our method and the baselines The rank column presents the average rank and its standard deviation across different time series. An average rank of 1 means the model was the best performing on all time series.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Method</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4.2.">Explainability Aspects</head><p>Figure <ref type="figure" target="#fig_0">1</ref> shows the TreeSHAP values clusters of the Saugeen River Flow data set. The dots on each lag value stand for the TreeSHAP values taken by the models belonging to the same cluster, while the line connects the mean values to show the TreeSHAP values for each lag of the representative selected model on each cluster (only for visualization purposes). Note that we show the name of the model and the value of the first hyper-parameter plus the lag value on which it is trained to distinguish between selected models belonging to the same family of tree-based models, e.g., RF200(Lag10) and RF50(Lag7). It can be seen that on different clusters different patterns of lagged values contributions to the target time series observations are observed. This confirms that our clustering procedure promotes ensemble diversity by enforcing the selection of models that have different modeling paradigms and distinct views on the importance of specific lag values. For example, while models in cluster 6 favor higher lag values and emphasize the contribution of their corresponding value to the output forecast value, models in cluster 5 are built on the assumption of restricting the memory of the models to lower lag values (𝑙 = 3). We can notice that in 3 clusters out of 6, models rely on restricted lagged values (clusters 2, 3, and 5). Even with this limited width of memory, i.e., historical data, they can excel in terms of predictive performance. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Concluding Remarks and Future Work</head><p>This paper introduces OEP-TT a novel method for online adaptive ensemble of tree-based models pruning. Through the use of TreeSHAP values, we are able to gain insight into its decision-making process, both for model selection, as well as for the input time series points relevance. We showed the advantages of OEP-TT on 100 real-world datasets, both in terms of predictive performance as well as its explainability aspects. In future work, we plan to extend our method to hybrid model pools by using the most efficient Shapley value estimation methods for each model family, such as TreeSHAP for tree-based models, DeepSHAP <ref type="bibr" target="#b7">[8]</ref> for Neural Networks, as well as KernelSHAP <ref type="bibr" target="#b7">[8]</ref> for remaining models, tune the size of the ensemble and dive further into the explainability aspects. Given that this is early-stage work, we plan to engage with experts in the field to exchange ideas and gather feedback.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Comparison of TreeSHAP-based models clusters on the Saugeen River Flow data set.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head></head><label></label><figDesc>is a sequence of 𝑋 until time 𝑡 and 𝑥 𝑖 is the value of 𝑋 at time 𝑖. Denote with T = {𝑇 1 , 𝑇 2 , • • • , 𝑇 𝑀 } the pool of 𝑀 tree-based models trained to approximate a true unknown function 𝑓 that generated 𝑋. 𝑥 𝑡+ℎ ) by each of the models in T. An ensemble model 𝑇 ¯T of T at time instant 𝑡 + ℎ can be formally expressed as a convex combination of the forecasts of the models in</figDesc><table><row><cell>Let 𝑥 ^𝑡+ℎ = (𝑥 ^𝑇 1 𝑡+ℎ , 𝑥 ^𝑇 2 𝑡+ℎ , • • • , 𝑥 ^𝑇 𝑀 𝑡+ℎ ) be the vector of forecast values of 𝑋 at a future time</cell></row><row><cell>instant 𝑡 + ℎ, ℎ ≥ 1 (i.e. T: 𝑇 ¯T(𝑥 ^𝑡+ℎ ) = ∑︀ 𝑀 𝑗=1 𝑤 𝑗 𝑡+ℎ 𝑥 ^𝑇 𝑗</cell></row></table><note>𝑡+ℎ where 𝑤 𝑗 𝑡+ℎ ∈ [1, 𝑀 ] are the ensemble weights. The weights</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head></head><label></label><figDesc>with 𝜔 a provided window size. 𝑋 𝑡𝑟𝑎𝑖𝑛 𝜔 is used for training the models in T and 𝑋 𝑣𝑎𝑙 𝜔 is used to compute the TreeSHAP values. For each tree-based model 𝑇 𝑗 ∈ 𝒯 , for each observation 𝑥 𝑡−𝜔+𝑘 ∈ 𝑋 𝑣𝑎𝑙 𝑙 𝑗 ], where 𝑙 𝑗 is the number of lags on which the model 𝑇 𝑗 is trained. Then, we aggregate absolute SHAP values over all the observations in 𝑋 𝑣𝑎𝑙 𝜔 to acquire SHAP-based lag importance 𝐼 𝑗 𝑖 for each lag 𝑖 ∈ [1, 𝑙 𝑗 ] using the model 𝑇 𝑗 :</figDesc><table /><note>𝜔 with 𝑘 ∈ [1, 𝜔], we compute a TreeSHAP value 𝜑 𝑗 𝑖 (𝑥 𝑡−𝜔+𝑘 ) for each lagged value, i.e., 𝑖 ∈ [1,</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 1</head><label>1</label><figDesc></figDesc><table><row><cell>Tree-based Model</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 2</head><label>2</label><figDesc></figDesc><table /></figure>
		</body>
		<back>

			<div type="funding">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>⋆ This research has partly been funded by the Federal Ministry of Education and Research of Germany and the state of North-Rhine-Westphalia as part of the Lamarr-Institute for Machine Learning and Artificial Intelligence.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">A survey on concept drift adaptation</title>
		<author>
			<persName><forename type="first">J</forename><surname>Gama</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Žliobaitė</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bifet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Pechenizkiy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bouchachia</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACM computing surveys (CSUR)</title>
		<imprint>
			<biblScope unit="volume">46</biblScope>
			<biblScope unit="page" from="1" to="37" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Explainable online deep neural network selection using adaptive saliency maps for time series forecasting</title>
		<author>
			<persName><forename type="first">A</forename><surname>Saadallah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Jakobs</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Morik</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Machine Learning and Knowledge Discovery in Databases</title>
				<editor>
			<persName><forename type="first">N</forename><surname>Oliver</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><surname>Pérez-Cruz</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Kramer</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Read</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>Lozano</surname></persName>
		</editor>
		<meeting><address><addrLine>Cham</addrLine></address></meeting>
		<imprint>
			<publisher>Springer International Publishing</publisher>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="404" to="420" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Explainable online ensemble of deep neural network pruning for time series forecasting</title>
		<author>
			<persName><forename type="first">A</forename><surname>Saadallah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Jakobs</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Morik</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Machine Learning</title>
		<imprint>
			<biblScope unit="volume">111</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Online adaptive multivariate time series forecasting</title>
		<author>
			<persName><forename type="first">A</forename><surname>Saadallah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Mykula</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Morik</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Joint European conference on machine learning and knowledge discovery in databases</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">A drift-based dynamic ensemble members selection using clustering for time series forecasting</title>
		<author>
			<persName><forename type="first">A</forename><surname>Saadallah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Priebe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Morik</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Joint European conference on machine learning and knowledge discovery in databases</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">An actor-critic ensemble aggregation model for time-series forecasting</title>
		<author>
			<persName><forename type="first">A</forename><surname>Saadallah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Tavakol</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Morik</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE ICDE</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><surname>Jakobs</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Saadallah</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2401.01124</idno>
		<title level="m">Explainable adaptive tree-based model selection for time series forecasting</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">A Unified Approach to Interpreting Model Predictions</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Lundberg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S.-I</forename><surname>Lee</surname></persName>
		</author>
		<ptr target="http://papers.nips.cc/paper/7062-a-unified-approach-to-interpreting-model-predictions.pdf" />
	</analytic>
	<monogr>
		<title level="m">Advances in Neural Information Processing Systems 30</title>
				<editor>
			<persName><forename type="first">I</forename><surname>Guyon</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">U</forename><forename type="middle">V</forename><surname>Luxburg</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Bengio</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Wallach</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Fergus</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Vishwanathan</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Garnett</surname></persName>
		</editor>
		<imprint>
			<publisher>Curran Associates, Inc</publisher>
			<biblScope unit="page" from="4765" to="4774" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Using dynamic time warping to find patterns in time series</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">J</forename><surname>Berndt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Clifford</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">KDD workshop</title>
				<imprint>
			<date type="published" when="1994">1994</date>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="page" from="359" to="370" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Monash time series forecasting archive</title>
		<author>
			<persName><forename type="first">R</forename><surname>Godahewa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Bergmeir</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">I</forename><surname>Webb</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">J</forename><surname>Hyndman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Montero-Manso</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Neural Information Processing Systems Track on Datasets and Benchmarks</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
	<note>Forthcoming</note>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">E</forename><surname>Box</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">M</forename><surname>Jenkins</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">C</forename><surname>Reinsel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">M</forename><surname>Ljung</surname></persName>
		</author>
		<title level="m">Time series analysis: forecasting and control</title>
				<imprint>
			<publisher>John Wiley &amp; Sons</publisher>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Applying lstm to time series predictable through time-window approaches</title>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">A</forename><surname>Gers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Eck</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Schmidhuber</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Neural Nets WIRN Vietri-01</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2002">2002</date>
			<biblScope unit="page" from="193" to="200" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Time-series forecasting of indoor temperature using pre-trained deep neural networks</title>
		<author>
			<persName><forename type="first">P</forename><surname>Romeu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Zamora-Martínez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Botella-Rocamora</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Pardo</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International conference on artificial neural networks</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2013">2013</date>
			<biblScope unit="page" from="451" to="458" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Random forests</title>
		<author>
			<persName><forename type="first">L</forename><surname>Breiman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Machine learning</title>
		<imprint>
			<biblScope unit="volume">45</biblScope>
			<biblScope unit="page" from="5" to="32" />
			<date type="published" when="2001">2001</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">A gradient boosting approach to the kaggle load forecasting competition</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">B</forename><surname>Taieb</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">J</forename><surname>Hyndman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International journal of forecasting</title>
		<imprint>
			<biblScope unit="volume">30</biblScope>
			<biblScope unit="page" from="382" to="394" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Xgboost: extreme gradient boosting</title>
		<author>
			<persName><forename type="first">T</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Benesty</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Khotilovich</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Tang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Cho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Mitchell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Cano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Zhou</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">R package version</title>
		<imprint>
			<biblScope unit="volume">0</biblScope>
			<biblScope unit="issue">4-2 1</biblScope>
			<biblScope unit="page" from="1" to="4" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Lightgbm: A highly efficient gradient boosting decision tree</title>
		<author>
			<persName><forename type="first">G</forename><surname>Ke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Meng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Finley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Ye</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T.-Y</forename><surname>Liu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in neural information processing systems</title>
		<imprint>
			<biblScope unit="volume">30</biblScope>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Arbitrated ensemble for time series forecasting</title>
		<author>
			<persName><forename type="first">V</forename><surname>Cerqueira</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Torgo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Pinto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Soares</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Joint European conference on machine learning and knowledge discovery in databases</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page" from="478" to="494" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Arbitrage of forecasting experts</title>
		<author>
			<persName><forename type="first">V</forename><surname>Cerqueira</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Torgo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Pinto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Soares</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Machine Learning</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
