<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Optimal Robust Simplifications for Explaining Time Series Classifications</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Jan</forename><forename type="middle">Arne</forename><surname>Telle</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Department of Informatics</orgName>
								<orgName type="institution">University of Bergen</orgName>
								<address>
									<country key="NO">Norway</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Cèsar</forename><surname>Ferri</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">VRAIN</orgName>
								<orgName type="institution">Universitat Politècnica de València</orgName>
								<address>
									<country key="ES">Spain</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Brigt</forename><surname>Håvardstun</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Department of Informatics</orgName>
								<orgName type="institution">University of Bergen</orgName>
								<address>
									<country key="NO">Norway</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff2">
								<orgName type="department">Explainable AI for Time Series and Data Streams Tutorial-Workshop</orgName>
								<address>
									<addrLine>September 9</addrLine>
									<postCode>2024</postCode>
									<settlement>Vilnus</settlement>
									<country key="LT">Lithuania</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Optimal Robust Simplifications for Explaining Time Series Classifications</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">DE322D51DD9AAF3B2B8DF39DF001EC5E</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T18:54+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Time Series</term>
					<term>XAI</term>
					<term>Simplification</term>
					<term>Explainability</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Given a Time Series Classifier (TSC) and three parameters that balance between error, simplicity and robustness, we define an optimization problem over all possible ways of simplifying a given time series 𝑡𝑠 into straight-line segments. Robustness is fixed as the fraction of perturbations that have the same classification as 𝑡𝑠 under the TSC, and we introduce a novel method for generating a set of perturbations where this information is easy to visualize. We prove that under some mild conditions on the TSC and the three parameters, we can find the optimal solution in time polynomial in the length of 𝑡𝑠, by first doing dynamic programming to solve for error and simplicity, and then adding robustness. We test the resulting Optimal Robust Simplification (ORS)-Algorithm on binary TSCs for three datasets from UCR. We apply the ORS-Algorithm to prototypes of the classes, with varying parameters, to evaluate its power as an explanatory tool for the trained classifiers. We also provide a tool for visualizing the robustness information. We believe the resulting insights show the usefulness of the Optimal Robust Simplifications in explaining TSCs.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Temporal data is encountered in many real-world applications ranging from patient data in healthcare <ref type="bibr" target="#b0">[1]</ref> to the field of cyber security <ref type="bibr" target="#b1">[2]</ref>. Deep learning methods have been successful <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b0">1,</ref><ref type="bibr" target="#b1">2]</ref> for TSC -Time Series Classification -but such methods are not easily interpretable, and often viewed as black boxes, which limits their applications when user trust in the decision process is crucial. To enable the analysis of these black-box models we revert to post-hoc interpretability. Recent research has focused on adapting existing methods to time series, both specific methods like SHAP-LIME <ref type="bibr" target="#b3">[4]</ref>, Saliency Methods <ref type="bibr" target="#b4">[5]</ref> and Counterfactuals <ref type="bibr" target="#b5">[6]</ref>, and also combinations of these <ref type="bibr" target="#b6">[7]</ref>.</p><p>As humans learn and reason by forming mental representations of concepts based on examples, and any machine learning model has been trained on data, then we believe that data e.g. in the form of prototypes and counterfactuals is indeed the natural common language between the user and this model. However, the basic problem is that compared to images and text, time series data are not intuitively understandable to humans <ref type="bibr" target="#b7">[8]</ref>. This makes interpretability of time series extra demanding, both when it comes to understanding how users will react to the provided explanations and to predict what explanatory tools are best. For example, in <ref type="bibr" target="#b8">[9]</ref> a tool was given for explainability of a TSC, that allowed model inspection so the user could form their own mental model of the classifier. However, a user evaluation showed that non-expert end-users were not able to make use of this freedom, supporting the notion that time-series are particularly non-intuitive for humans.</p><p>Theissler et al <ref type="bibr" target="#b9">[10]</ref> give an intriguing taxonomy of XAI for TSC, divided into those focused on (i) Time-points (e.g. SHAP and LIME) or (ii) Subsequences (e.g. Shapelets) or (iii) Instances (Prototypes, CF, Feature-based). Explaining a classifier by instances like prototypes has a definite appeal, but it is a challenge how to highlight the features important for the classification, while at the same time simplifying the instance to mitigate the non-intuitive aspects of this domain, as we attempt in this work. Theissler et al <ref type="bibr" target="#b9">[10]</ref> review the literature on XAI for TSC and of the 9 papers they discuss on Instance-based explanations using Prototypes or Features, it is remarkable that not a single one is specialized for Local explanations, rather they are all Global. In this work we fill this gap, and introduce a way of simplifying a time series by straight-line segments, and doing this in a way that is robust with respect to the classification of a given TSC ℎ. <ref type="foot" target="#foot_0">1</ref> Our ORS-algorithm, for Optimal Robust Simplifications, is model-agnostic, with robustness of a simplification being the fraction of local perturbations that retains the classification under TSC ℎ. There are various ways of perturbing times series, see e.g. <ref type="bibr" target="#b10">[11,</ref><ref type="bibr" target="#b11">12,</ref><ref type="bibr" target="#b12">13]</ref>. In this work we develop a different method of computing a set of perturbations, both since we start with a simplification on straight-line segments and also because we want the aggregate information about their classifications to be visualized in an informative way. When computing robustness we start with a straight-line simplification, and any perturbation must to a certain degree keep that position and shape, while it is also allowed to deviate as the perturbation moves inside a band around the simplification, see Figure <ref type="figure" target="#fig_9">8</ref>.</p><p>The ORS-Algorithm computing a robust simplification 𝑠𝑡𝑠 of a given times series 𝑡𝑠, under a TSC ℎ, has three parameters that allows to balance between error (Squared Euclidean distance between 𝑡𝑠 and 𝑠𝑡𝑠), simplicity (number of segments of 𝑠𝑡𝑠), and robustness (what fraction of perturbations of 𝑠𝑡𝑠 have same classification as 𝑡𝑠 by ℎ), and 𝑠𝑡𝑠 is defined as the minimization over an objective function taking all this into account, see Definition 1. Perhaps surprisingly, in Theorem 4 we show that under some mild conditions on ℎ and the three parameters, we can find the optimal solution in time polynomial in the length of the original time series 𝑡𝑠, by first solving for error and simplicity, and then adding robustness. Note that our algorithm allows not only the balance between error, simplicity and robustness to change, but also their computation, i.e. the algorithm is agnostic to the particular methods used to calculate these values. Hence Squared Euclidean distance can be replaced by some other distance function as long as it is amenable to the dynamic programming, simplicity can depend on other aspects than number of segments, and robustness can be computed by other types of perturbations.</p><p>In the rest of this paper we first discuss related work and cover some standard definitions before presenting the ORS-Algorithm together with proof of correctness and runtime. We then discuss the usefulness of the robust simplifications, with a focus in this paper on time series that are discrete, univariate, evenly sampled, aperiodic, short, and non-stationary, based on UCR datasets Chinatown, Italy Power Demand and ECG200. We also provide an aid to visualize the robustness information. Our findings suggest that the robust simplifications of a prototype 𝑡𝑠 can be used to extract explanations of the classification done by the TSC, as they simplify 𝑡𝑠 to the point where human intuition can more easily be applied. We conclude the paper by discussing some directions for future work.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related Work</head><p>Several techniques have been applied to generate explanations from TSC models <ref type="bibr" target="#b9">[10,</ref><ref type="bibr" target="#b13">14]</ref>. Specifically, techniques previously used on Convolutional Neural Networks or Recurrent Neural Networks have been applied when these techniques have been used for TSC. For instance, the authors in <ref type="bibr" target="#b6">[7]</ref> apply to time series several XAI methods previously used on image and text domain. They also introduce verification techniques specific to times series, in particular a perturbation analysis and a sequence evaluation. Related to the datasets used in our paper, <ref type="bibr" target="#b14">[15]</ref> presents a user study of XAI methods to explain the ECG200 dataset from UCR.</p><p>Another alternative is to produce explanations through examples, and these can be specifically utilized in the time series domain. One type of this explanation method involves giving the nearest example from the training dataset that acts as a prototype to depict the normal behavior of a similar sample <ref type="bibr" target="#b15">[16]</ref>. A method to generate prototypes for time series data using an autoencoder is presented in <ref type="bibr" target="#b16">[17]</ref>. The main novelty of the work is the method to generate diverse prototypes.</p><p>Given that in many cases raw time series can be too complex for humans, several studies have tried to employ simplified versions as explanations. In <ref type="bibr" target="#b17">[18]</ref>, the authors propose a dimensionality reduction technique for time series, called Piecewise Aggregate Approximation, as a tool for similarity search in large time series databases. The work <ref type="bibr" target="#b18">[19]</ref> presents a review of piecewise linear approximations employed for time series data, and the authors also propose a method named SWAB (based on the Sliding Window and Bottom-up approaches). Finally, in <ref type="bibr" target="#b19">[20]</ref> Camponogara and Nazari introduce a range of piecewise-linear models and algorithms for unknown functions. Another option is to segment time series into fixed-width windows and employ these to justify decisions. In <ref type="bibr" target="#b20">[21]</ref> Schlegel et al propose TS-MULE, a model explanation method working to segment and perturb time series data and extending LIME. A study on the effect of segmentation on time series, in particular in the field of finance for the use of pattern matching was presented in <ref type="bibr" target="#b21">[22]</ref>. In <ref type="bibr" target="#b22">[23]</ref>, Si and Yin also apply segmentation to financial time series data, now as a preprocessing step to locate technical patterns.</p><p>Perturbations have also been previously employed for XAI and time series. In <ref type="bibr" target="#b10">[11]</ref>, the authors propose a framework to generate explanations for time series classifications based on a generative model to create with-in-distribution perturbations. Enguehard, in <ref type="bibr" target="#b11">[12]</ref>, introduces a technique to explain multivariate time series predictions using a perturbation-based saliency method. Recently, the approach ContraLSP has been detailed in <ref type="bibr" target="#b12">[13]</ref>. This method is based on a locally sparse model that produces counterfactual samples to build uninformative perturbations but keeps distribution using contrastive learning.</p><p>Some tools have been proposed to allow users to interact with time series and, in that way, to get some insights about their classification, like <ref type="bibr" target="#b23">[24]</ref> where model inspection is the key aspect. In <ref type="bibr" target="#b24">[25]</ref>, the authors develop an interactive XAI tool for loan applications that allows users to experiment with hypothetical input values and inspect their effect on the outcomes of the model and perform a user evaluation on MTurk. <ref type="bibr" target="#b25">[26]</ref> presents a Python package to provide a unified interface for the interpretation of time series classification.</p><p>Some papers have developed a similar approach to our proposal. Im <ref type="bibr" target="#b26">[27]</ref>, Tang et al. presents the method Dual Prototypical Shapelet Networks (DPSN) for few-shot time-series classification. This method trains a neural network from few examples and also interprets the model from two levels of granularity: a global overview with representative time series examples, and local highlights with discriminative shapelets. Another related paper for sequence learning is <ref type="bibr" target="#b27">[28]</ref>. The work proposes ProSeNet a RNN-based method for deep sequence modeling. The approach combines prototype learning and RNNs to construct models with high accuracy and interpretability. The method considers three criteria in constructing prototypes: Simplicity, the prototypes can be subsequences with only the key events determining the output; Diversity, redundant prototypes should be avoided; Sparsity, for each example to explain only a few prototypes are shown to avoid long explanations. The authors show the validity of ProSeNet for time series using the MIT-BIH Arrhythmia ECG dataset.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Standard definitions</head><p>Let us present formal definitions for Time Series Classification (TSC) and recall basic notions.</p><p>Staying consistent with earlier notation <ref type="bibr" target="#b9">[10,</ref><ref type="bibr" target="#b5">6]</ref> a time series 𝑇 = {𝑡 1 , 𝑡 2 , . . . , 𝑡 𝑚 } is an ordered set of 𝑚 real-valued observations (or time steps). Note we may also view each 𝑡 𝑖 as a pair consisting of an 𝑥-value (the time) and a 𝑦-value (the observation). A time series dataset 𝐷 = {𝑇 1 , 𝑇 2 , ..., 𝑇 𝑛 } ∈ 𝑅 𝑛×𝑚 is a collection of such time series where each time series has a class label 𝑐 forming a vector of class labels. In this paper we consider only binary classification tasks. Given such a dataset, Time Series Classification is the task of training a mapping 𝑏 from the space of possible inputs to a probability distribution over the class values. Thus, a black-box classifier 𝑏(𝑇 ) takes a time series 𝑇 as input and predicts a probability output over the class values.</p><p>Prototypes are time series exemplifying the main aspects responsible for a classifier's specific decision outcome. It can be a real instance (which is what we opt for) sampled from the dataset that is important and meaningful because it summarizes the shape of many other similar instances, or a synthetic one, for example a cluster centroid or an instance generated by following some ad-hoc processes.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">An Algorithm for Optimal Robust Simplifications</head><p>In this section we give the ORS-Algorithm that takes a binary Time Series Classifier ℎ and a univariate times series 𝑡𝑠, on 𝑛 points 𝑝 1 , ..., 𝑝 𝑛 given by increasing 𝑥-values, and computes the optimal robust simplification 𝑜𝑝𝑡 𝑠𝑖𝑚𝑝 (𝑡𝑠, ℎ, 𝛼, 𝛽, 𝛾), where 𝛼, 𝛽, 𝛾 are factors freely chosen to balance between error, simplicity and robustness. This simplification will select 𝑘 + 1 of the 𝑛 points 𝑝 𝑖 1 , 𝑝 𝑖 2 , ..., 𝑝 𝑖 𝑘+1 with 𝑖 𝑗 &lt; 𝑖 𝑗+1 , ∀𝑗 to form a simplified time series 𝑠𝑡𝑠 on 𝑘 segments by drawing a line between each consecutive pairs of the 𝑘 + 1 points and by letting the first segment and the last segment extend to the 𝑥-values of 𝑝 1 and 𝑝 𝑛 , respectively. See Figure <ref type="figure">1</ref>. The optimal simplification thus reflects a selected balance of error, simplicity and robustness achieved by freely choosing the 3 factors 𝛼, 𝛽, 𝛾:</p><p>• 𝛼 for error: 𝑒𝑟𝑟(𝑡𝑠, 𝑠𝑡𝑠) is Squared Euclidean distance between 𝑡𝑠 and 𝑠𝑡𝑠, i.e. the sum over all 𝑥-values, of the square of the difference between the corresponding 𝑦-value of 𝑡𝑠 and 𝑠𝑡𝑠. • 𝛽 for simplicity: 𝑘 𝑠𝑡𝑠 is the number of segments • 𝛾 for robustness: 𝑓 𝑟𝑎𝑔(𝑡𝑠, 𝑠𝑡𝑠, ℎ) is a measure of how robust the classifier ℎ is on perturbations of 𝑠𝑡𝑠, measured by the fraction of perturbations that do not fall in the same class as 𝑡𝑠, thus in the range [0, 1] with 0 being most robust (𝑓 𝑟𝑎𝑔 is short for fragility and can be viewed as (1-robustness)). See subsection 4.2.2 for the definition of the perturbations used.</p><p>Perhaps surprisingly, under some mild assumptions this complicated objective function still allows an algorithm that finds the optimal in time polynomial in 𝑛. In this section we describe this algorithm in detail.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">If no robustness: Dynamic Programming</head><p>We first solve the case of 𝛾 = 0, i.e. without giving any weight to robustness. Firstly, Least Squares is a foundational problem in statistics and numerical analysis, given points 𝑝 1 , 𝑝 2 , ..., 𝑝 𝑛 in the plane find the straight line that minimizes the squared error. In the more general Segmented Least Squares (SLS) problem we ask for a sequence of straight-line segments that minimizes cost, which is some balance of error and simplicity (number of segments). The SLS problem is famously solvable in polynomial time by a textbook Dynamic Programming Algorithm based on the following recurrence for 𝑂𝑃 𝑇 (𝑗), the minimum cost for points 𝑝 1 , ..., 𝑝 𝑗 :</p><formula xml:id="formula_0">𝑂𝑃 𝑇 (𝑗) = {︃ 0, if 𝑗 = 0 min 1≤𝑖≤𝑗 {𝑙𝑠𝑒(𝑖, 𝑗) + 𝛽 + 𝑂𝑃 𝑇 (𝑖 − 1)} else</formula><p>where 𝑙𝑠𝑒(𝑖, 𝑗) is the least square error over all single lines that cover 𝑝 𝑖 to 𝑝 𝑗 , and 𝛽 is the the cost of one extra segment, with value of 𝛽 chosen to balance between error and simplicity. This problem formulation differs from our setting in 3 important ways:</p><p>• We want a continuous set of segments, with each segment starting and ending in given points.</p><p>• We want the first (resp. last) segment to be defined by any two points 𝑝 𝑖 , 𝑝 𝑗 and the line drawn between them continued until it hits the 𝑥-value of 𝑝 1 (resp. of 𝑝 𝑛 ). • We want robustness to perturbations as part of the objective function.</p><p>The first two differences are relatively easy to accommodate, as follows:</p><formula xml:id="formula_1">𝑂𝑃 𝑇 (𝑗) = ⎧ ⎪ ⎪ ⎪ ⎪ ⎨ ⎪ ⎪ ⎪ ⎪ ⎩ 0, if 𝑗 = 0 min 1≤𝑖≤𝑗 {[𝛼 • 𝑒𝑟𝑟(𝑖, 𝑗) + 𝛽 + 𝑂𝑃 𝑇 (𝑖)], [𝛼 • 𝑒𝑟𝑟(1, 𝑖, 𝑗) + 𝛽]} if 𝑗 &lt; 𝑛 min 1≤𝑖≤𝑛 min 𝑖+1≤𝑘≤𝑛 {[𝛼 • 𝑒𝑟𝑟(𝑖, 𝑘, 𝑛) + 𝛽 + 𝑂𝑃 𝑇 (𝑖)], [𝛼 • 𝑒𝑟𝑟(1, 𝑖, 𝑘, 𝑛) + 𝛽]} if 𝑗 = 𝑛</formula><p>where • 𝑒𝑟𝑟(𝑖, 𝑗) is the Squared Euclidean distance between the straight line from 𝑝 𝑖 to 𝑝 𝑗 and the part of the original time series 𝑡𝑠 given by points 𝑝 𝑖 , ..., 𝑝 𝑗 • 𝑒𝑟𝑟(𝑖, 𝑗, 𝑛) is the same except the straight line continues past 𝑝 𝑗 to the 𝑥-value of 𝑝 𝑛 and we compare to the final part of 𝑡𝑠 given by 𝑝 𝑖 , ..., 𝑝 𝑛 • 𝑒𝑟𝑟(1, 𝑖, 𝑗) is similar as above except the line continues in the other direction to the 𝑥-value of 𝑝 1 and we compare to the initial part of 𝑡𝑠 𝑝 1 , ..., 𝑝 𝑗 • 𝑒𝑟𝑟(1, 𝑖, 𝑗, 𝑛) compares 𝑡𝑠 to a single line covering all 𝑥-values and going through points 𝑝 𝑖 , 𝑝 𝑗 For the case of 𝑜𝑝𝑡 𝑠𝑖𝑚𝑝 (𝑡𝑠, ℎ, 𝛼, 𝛽, 𝛾) where 𝛾 = 0 i.e. where robustness does not play a part, the argument for correctness of the above recurrence is similar to the textbook argument for correctness of the recurrence of SLS, so we do not repeat it here, except to note the new parts being that • We use 𝑂𝑃 𝑇 (𝑖) rather than 𝑂𝑃 𝑇 (𝑖 − 1) since now for two consecutive segments the endpoint of the first is the startpoint of the next. • When defining 𝑂𝑃 𝑇 (𝑗) we use 𝑒𝑟𝑟(1, 𝑖, 𝑗) so first segment can end in 𝑗 • When defining 𝑂𝑃 𝑇 (𝑛) we use 𝑒𝑟𝑟(𝑖, 𝑘, 𝑛) so last segment can choose any pair 𝑝 𝑖 , 𝑝 𝑘 , and 𝑒𝑟𝑟(1, 𝑖, 𝑘, 𝑛) to allow simplification with a single segment.</p><p>Transforming this recurrence into a dynamic programming algorithm is straightforward, by first computing each of the 𝑂(𝑛 2 ) error terms in time 𝑂(𝑛) each, followed by a nested loop computing the 𝑛 𝑂𝑃 𝑇 (𝑗) values in 𝑂(𝑛) time each. Retrieving the optimal solution is done by a second pass over the values stored, also standard. This gives the following result. Theorem 1. Given a time series 𝑡𝑠 of length 𝑛, a binary TSC ℎ, and factors 𝛼, 𝛽, 𝛾 that balance between error, simplicity and robustness. For the case of 𝛾 = 0 (no robustness) we can compute the optimal simplification achieving the minimum 𝑜𝑝𝑡 𝑠𝑖𝑚𝑝 (𝑡𝑠, ℎ, 𝛼, 𝛽, 𝛾 = 0) in time 𝑂(𝑛 3 ) by dynamic programming.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Accommodating Robustness: DP with Heaps</head><p>Dynamic programming relies on the property that an optimal solution to a larger problem consists of optimal solutions to subproblems. Robustness depends on the TS Classifier ℎ and we cannot expect it to satisfy this property, since its classification on only a part of a time series will not in general correspond to its classification on the full time series. Thus, we cannot incorporate robustness as part of the DP scheme. Instead, our algorithm for Optimal Robust Simplifications (ORS), the ORS-Algorithm, uses a 2-stage approach: Pseudo-code is given in the Appendix. We will here give a high-level explanation, but let us first state the following formal result. Theorem 2. Let the difference in cost between 𝑠𝑡𝑠 1 and 𝑠𝑡𝑠 𝑞 , as computed by Stage 1, be 𝑑. For any value of 𝛾 ≤ 𝑑 the simplification returned by the ORS-Algorithm will then be an optimal one achieving 𝑜𝑝𝑡 𝑠𝑖𝑚𝑝 (𝑡𝑠, ℎ, 𝛼, 𝛽, 𝛾).</p><p>Proof 1. We prove the statement by contradiction. Let the simplification 𝑠𝑡𝑠 returned by the ORS-Algorithm have 𝑘 𝑠𝑡𝑠 segments and assume there is another simplification 𝑠𝑡𝑠 ′ with 𝑘 𝑠𝑡𝑠 ′ segments such that</p><formula xml:id="formula_2">𝛼 • 𝑒𝑟𝑟(𝑡𝑠, 𝑠𝑡𝑠 ′ ) + 𝛽 • 𝑘 𝑠𝑡𝑠 ′ + 𝛾 • 𝑓 𝑟𝑎𝑔(𝑡𝑠, 𝑠𝑡𝑠 ′ , ℎ) &lt; 𝛼 • 𝑒𝑟𝑟(𝑡𝑠, 𝑠𝑡𝑠) + 𝛽 • 𝑘 𝑠𝑡𝑠 + 𝛾 • 𝑓 𝑟𝑎𝑔(𝑡𝑠, 𝑠𝑡𝑠, ℎ)</formula><p>Since Stage 2 optimized this value over all simplifications 𝑠𝑡𝑠 1 , ..., 𝑠𝑡𝑠 𝑞 we cannot have 𝑠𝑡𝑠 ′ among these, and thus by the assumption in the statement of the theorem we have 𝛼</p><formula xml:id="formula_3">• 𝑒𝑟𝑟(𝑡𝑠, 𝑠𝑡𝑠 ′ ) + 𝛽 • 𝑘 𝑠𝑡𝑠 ′ ≥ 𝑑 + 𝛼 • 𝑒𝑟𝑟(𝑡𝑠, 𝑠𝑡𝑠 1 ) + 𝛽 • 𝑘 𝑠𝑡𝑠 1 , with 𝑘 𝑠𝑡𝑠 1 the number of segments of 𝑠𝑡𝑠 1 . Thus we get 𝑑 + 𝛼 • 𝑒𝑟𝑟(𝑡𝑠, 𝑠𝑡𝑠 1 ) + 𝛽 • 𝑘 𝑠𝑡𝑠 1 + 𝛾 • 𝑓 𝑟𝑎𝑔(𝑡𝑠, 𝑠𝑡𝑠 ′ , ℎ) &lt; 𝛼 • 𝑒𝑟𝑟(𝑡𝑠, 𝑠𝑡𝑠) + 𝛽 • 𝑘 𝑠𝑡𝑠 + 𝛾 • 𝑓 𝑟𝑎𝑔(𝑡𝑠, 𝑠𝑡𝑠, ℎ) ≤ 𝛼 • 𝑒𝑟𝑟(𝑡𝑠, 𝑠𝑡𝑠 1 ) + 𝛽 • 𝑘 𝑠𝑡𝑠 1 + 𝛾 • 𝑓 𝑟𝑎𝑔(𝑡𝑠, 𝑠𝑡𝑠 1 , ℎ)</formula><p>with the latter inequality following since 𝑠𝑡𝑠 was the minimum over 𝑠𝑡𝑠 1 , ..., 𝑠𝑡𝑠 𝑞 . Rearranging, this gives</p><formula xml:id="formula_4">𝑑 &lt; 𝛾•𝑓 𝑟𝑎𝑔(𝑡𝑠, 𝑠𝑡𝑠 1 , ℎ)−𝛾•𝑓 𝑟𝑎𝑔(𝑡𝑠, 𝑠𝑡𝑠 ′ , ℎ)+((𝛼•𝑒𝑟𝑟(𝑡𝑠, 𝑠𝑡𝑠 1 )+𝛽•𝑘 𝑠𝑡𝑠 1 )−(𝛼•𝑒𝑟𝑟(𝑡𝑠, 𝑠𝑡𝑠 1 )+𝛽•𝑘 𝑠𝑡𝑠 1 ))</formula><p>with the latter term being zero and thus 𝑑 &lt; 𝛾 • (𝑓 𝑟𝑎𝑔(𝑡𝑠, 𝑠𝑡𝑠, ℎ) − 𝑓 𝑟𝑎𝑔(𝑡𝑠, 𝑠𝑡𝑠 ′ , ℎ))</p><p>Note we have 𝑓 𝑟𝑎𝑔(•) in the range [0, 1] which means the above implies that 𝛾 &gt; 𝑑, contradicting the assumption in the statement of the theorem. □</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.1.">Important details of Stage 1.</head><p>We describe some key details of how to implement Stage where 𝑠 1 (and 𝑠 2 ) is the 𝑦-value of 𝑠𝑡𝑠 that corresponds to the smallest (and largest) 𝑥-value, where smallest is same as 𝑥-value of 𝑝 1 (and largest is same as 𝑥-value of 𝑝 𝑛 ). In Figure <ref type="figure">1</ref>     </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Usefulness of the robust simplifications</head><p>We have implemented<ref type="foot" target="#foot_1">2</ref> the ORS-Algorithm and in this section we illustrate its use for explaining a Time Series Classifier. The algorithm is designed for classifiers working on univariate discrete time series with a binary classification. Moreover, we believe its use is more easily evaluated if the times series are evenly sampled, aperiodic and non-stationary. However, it is important to note that apart from the above constraints we did not want simple datasets where the binary classification is particularly easy or could be described (e.g. by ourselves) in any straightforward way. Based on these criteria we have chosen to focus on three datasets.</p><p>• From UCR <ref type="bibr" target="#b28">[29]</ref>: Chinatown. This dataset shows the number of pedestrians on a street corner of Chinatown in Melbourne over a 24-hour time period, and classifies these into Weekend and Weekdays. • From UCR <ref type="bibr" target="#b28">[29]</ref>: ECG200. This dataset traces the electrical activity recorded in 96 time steps during one heartbeat and classifies the time series into a normal heartbeat and a Myocardial Infarction. • From UCR <ref type="bibr" target="#b28">[29]</ref>: Italy Power Demand. This dataset shows power demand in Italian households over a 24-hour time period, and classifies these into Winter (October-March) and Summer (April-September).  For each of these datasets we trained a fully convolutional neural network (FCN), originally proposed by Wang et al. <ref type="bibr" target="#b29">[30]</ref>. Specifically we implemented a model closely following the work by Delaney et al. <ref type="bibr" target="#b5">[6]</ref>. To select prototypes we used SPOTGreedy by Gurumoorthy et al. <ref type="bibr" target="#b30">[31]</ref>, implemented in the python package InterpretML <ref type="bibr" target="#b31">[32]</ref>. We then ran the robust simplification algorithm on the resulting prototypes, with varying balancing factors, to evaluate their power as explanatory tools for the trained classifiers.</p><p>In this section we focus on three aspects of the robust simplifications to argue that they are helpful explanations of the classifier model.</p><p>• Showing prototypes together with their simplifications is more informative than showing prototypes only • Varying the balance of error vs simplicity vs robustness will highlight different aspects of the prototypes • Extracting a rule for class membership from this information can give simple and powerful explanations</p><p>We cover these aspects in separate subsections.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1.">Simplifications are informative as explanations</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1.1.">Chinatown</head><p>For the Chinatown dataset we have computed 3 prototypes from each of the two classes, that we call Red and Blue. These 6 prototypes are presented in Figure <ref type="figure" target="#fig_4">3</ref>. From these prototypes only it seems difficult to identify with any kind of certainty a rule-of-thumb that can be used to decide the classification of new instances. On the next figure, Figure <ref type="figure" target="#fig_5">4</ref>, we show a single simplification added to each of the prototypes. These simplifications are chosen with a well-balanced mix of error, simplicity and robustness, choosing each of 𝛼, 𝛽, 𝛾 in the mid-range of values <ref type="foot" target="#foot_2">3</ref> These simplifications function as explanations for the classification of each of the prototypes, as they also incorporate the robustness to 10.000 perturbations, and as such they allow a rule-of-thumb to be more readily guessed at. Looking at the first segment of each of the robust simplifications it is striking that the blue slopes are quite sharply downwards, whereas none of the red are. A natural guess is as follows:</p><p>• Rule-of-thumb for Chinatown classification: if the time series starts by a noticable downward slope then it is Blue, otherwise Red </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2.">Varying balancing factors</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2.1.">Chinatown</head><p>In Figure <ref type="figure">5</ref> we see the effect of varying the balancing factors for a Red prototype in the Chinatown dataset. We see that increasing the importance of simplicity from low to middle to high will decrease the number of segments, while naturally also increasing the Euclidean distance to the original time series. We also see that increasing the importance of robustness changes the slope of the first segment from slightly downward (in the other 6-segment simplification) to slightly upward. This means that a higher percentage of time series with first segment having slightly downward slope in this vicinity are classified as Blue, as compared to the percentage classified Blue having slightly upward slope. This change is significant information, and it holds even in the presence of a higher Euclidean distance to the original prototype. Note that this observation constitutes supporting evidence for the rule-of-thumb guessed at previously for Chinatown.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2.2.">Italy Power Demand</head><p>In Figure <ref type="figure">6</ref> we see a single prototype from ItalyPowerDemand in black and two robust simplifications of it in blue. Both simplifications have the same balancing factor for error and simplicity, but (a) has 4 segments while (b) has 5 segments resulting from an increase in the factor for robustness. Simplifying, we could say that a single segment in (a) has been replaced by two segments in (b) to give two peaks. This indicates a rule-of-thumb for the classifier trained on ItalyPowerDemand, namely that two such peaks at or around those 𝑥-values is important for the Blue classification.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2.3.">ECG200</head><p>In Figure <ref type="figure">7</ref> we see a prototype from ECG200 in black and two distinct simplifications of it in blue. Note that even though the factor for robustness is much higher in (b) than (a), the two simplifications are (almost) identical, both on 5 segments. To explain this lack of significance of robustness we turn to Theorem 2 which says that the difference between the best cost and 𝑞th best cost output by Stage 1 of the ORS Algorithm, may limit the values of 𝛾 for which Stage 2 returns the optimal robust simplification.</p><p>For the simplifications of Figure <ref type="figure">7</ref> we ran Stage 1 with the high value of 𝑞 = 1.000.000 and found this difference to be 0.03630-0.02884=0.00746. Thus by Theorem 2 the simplification in (a) may not be optimal for that value of 𝛾 = 0.01 &gt; 0.00746, whereas the one in (b) is likely not optimal, as 𝛾 = 0.1 &gt;&gt; 0.00746. Let us explain why we believe this happens. This time series has a length of 96, and note the DP to get 5 segments optimizes over all possible ways of choosing 6 points from 96. Note that there are many such choices of 6 points that give 5 segments similar to the ones in Figure <ref type="figure">7</ref>.</p><p>Counting the number of ways of choosing these 6 points, starting with the rightmost point and going left, a back-of-the-envelope calculation estimates this to be in the ballpark of 35*20*10*10*3*5 which is over 1 million. Thus, we take this as evidence that almost all the 𝑞 best simplifications returned by the dynamic programming of Stage 1, that do not take robustness into account, are very close to these 5 segments. Thus increasing the importance of robustness cannot help, as the result will always be a simplification similar to the one given. We leave for future work to deal with such cases, for example by a heuristic during the DP that discards simplifications that are only marginally different from others.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3.">Evaluating a classification rule guessed from robust simplifications</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3.1.">Chinatown</head><p>For the Chinatown classification we made a rule-of-thumb based on information gathered from the explanations of the classifications given for 6 prototypes, by means of the robust simplifications. To check if this rule-of-thumb is actually useful we need to make it more specific. Looking at the values on the 𝑦-axis for Figure <ref type="figure" target="#fig_5">4</ref> we see that the first segments of the 3 Blue simplifications are lines starting at point (0, 500) and decreasing to about (5, 0) while the first segments of the Red simplifications are only very slightly decreasing or increasing. Thus, the difference between the 𝑦-values at 𝑥 = 0 and 𝑥 = 5 would distinguish between these Red and Blue prototypes. We guess that this difference also distinguishes between many other time series in the dataset, as this insight is gathered from the robust simplifications. Taking the halfway point of 250 between the 𝑦-values of 500 and 0 as the cutoff the simple classification rule becomes</p><p>• Simple Rule for Chinatown: if the 𝑦-value at 𝑥 = 0 minus the 𝑦-value at 𝑥 = 5 is larger than 250 then 𝑡𝑠 is classified Blue, otherwise Red</p><p>The Chinatown dataset has 363 time series. We checked their classification by the model and compared to the Simple Rule. Indeed, the Simple Rule classifies all but 1 of the model-classified Blue time series as Blue, and it classifies all but 19 of the Red time series as Red, for an accuracy of 94.5 %. Note this is accuracy with respect to the model, not the ground truth, as the Simple Rule is trying to explain the model. Let us mention that the accuracy of the model wrt ground truth is 96.4 %, while accuracy of the simple rule wrt ground truth is 97 %.</p><p>Finally, let us also mention that if we raise the cutoff in the Simple Rule from 250 to 325 it actually achieves an accuracy of 99.7 % (with accuracy of the simple rule vs ground truth down to 96.7%) failing on only one instance. In retrospect, we could ask if there was any information available to us from the robust simplifications that would point to this higher cutoff. In fact, consider Figure <ref type="figure">5</ref> (a) that visualizes the robustness of a simplification whose first segment goes from (0, 400) to <ref type="bibr" target="#b6">(7,</ref><ref type="bibr">50)</ref>. Note the red part of the band around the first segment stretches up above (0, 250). This should have made us wonder if the cutoff for the simple rule should be higher than 250. We could have made a test of this, for example by constructing a time series whose first segment goes from (0, 300) to (5, 0) and visualize its robustness. See Figure <ref type="figure" target="#fig_9">8</ref>, where we have done exactly this, and pay close attention to the colourings of the band around the first segment. This band is predominantly blue down to somewhere above (0, 300), and already then it turns red. This in itself could have made us guess a higher cutoff than 250 for the Simple Rule. We believe these insights show the usefulness of the robust simplifications and their visualization.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3.2.">ItalyPowerDemand</head><p>In the previous subsection we saw that two peaks in the mid-afternoon seemed to signify Blue class for the classifier on the ItalyPowerDemand dataset. Looking more closely at Figure <ref type="figure">6</ref> we note that the difference between the high part of the peaks and the low part is about 0.5. We also note from the original prototype that the first peak is centered at 10, the second peak at 18, and that in (b) the bottom of the simplification is centered at 14. From this we extract the following rule: The dataset ItalyPowerDemand has 1096 time series. We checked their classification by the model and compared to the Simple Rule. Indeed, the Simple Rule classifies all but 50 of the model-classified Blue correctly, and all but 49 of the Red correctly, for an accuracy of 91%. Let us mention that accuracy of the model wrt ground truth is 96.3%, and of the Simple Rule wrt Ground Truth 89.6%. Note that classifiers of ItalyPowerDemand have previously been shown hard to explain to end-users <ref type="bibr" target="#b23">[24]</ref>. Thus, the Simple Rule, developed by examining the robust simplifications in Figure <ref type="figure">6</ref>, has a surprisingly high accuracy.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusions and Future Work</head><p>In this paper, we addressed the challenge of interpretability in Time Series Classification (TSC) by introducing the Optimal Robust Simplifications (ORS) algorithm. Our approach focuses on simplifying time series data into straight-line segments while ensuring that these simplifications remain robust with respect to the classifications made by a given TSC model.</p><p>Our ORS-Algorithm is model-agnostic and allows a balance between error (fidelity with respect to the original instance), simplicity (complexity of the simplification), and robustness (fraction of local perturbations that retain the correct classification). We prove that the ORS-Algorithm, under certain mild conditions, finds the optimal solution in polynomial time, by a first dynamic programming stage solving for error and simplicity, and then incorporating robustness into the solution. This simplification process aids in making time series data more intuitive and interpretable for humans.</p><p>Our experimental results, based on UCR datasets Chinatown, Italy Power Demand, and ECG200, confirm the practical utility of our approach. The robust simplifications provided by the ORS algorithm make time series data more accessible to human intuition, facilitating better user understanding and trust in the TSC models. In conclusion, our work contributes a novel method for enhancing the interpretability of TSC models through robust simplifications.</p><p>Future research could explore other ways to compute the factors employed in the algorithm. For instance, the squared Euclidean distance could be replaced by some other distance function, or simplicity can depend on other aspects than only the number of segments, and robustness can be computed by other types of perturbations. Finally, we plan to conduct experiments including human studies to demonstrate empirically the benefits of using optimal robust simplifications to explain TSC models and compare to related XAI methods.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :Definition 1 .</head><label>11</label><figDesc>Figure 1: Original time series 𝑡𝑠 in black, with selected points given by 6 grey circles, resulting in a simplification 𝑠𝑡𝑠 in red with 5 segments.</figDesc><graphic coords="4,162.25,196.70,270.78,203.08" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head></head><label></label><figDesc>𝑠 1 is the leftmost circled point and 𝑠 2 the rightmost circled point, at 𝑥-values 0 and 23 respectively. For these 𝑘 + 1 pivot points of 𝑠𝑡𝑠 we allow an increase or decrease of its 𝑦-value by an 𝜖 which is 10% of the 𝑦-range of the dataset the classifier ℎ was trained on, i.e. 𝜖 = 0.1 * (𝑦 𝑚𝑎𝑥 − 𝑦 𝑚𝑖𝑛 ) with 𝑦 𝑚𝑎𝑥 and 𝑦 𝑚𝑖𝑛 the maximum and minimum 𝑦-value in this dataset.We compute 10.000 random perturbations of 𝑠𝑡𝑠, each consisting of 𝑘 segments anchored at the same 𝑥-values as the 𝑘 + 1 pivot points of 𝑠𝑡𝑠 but with the 𝑦-value of each pivot point shifted by an amount drawn independently for the 𝑘 + 1 values uniformly at random in the range [−𝜖, +𝜖]. Thus the perturbation will lie inside a band around 𝑠𝑡𝑠 of height 2 • 𝜖 at each pivot point. See Figure2. We then use ℎ to classify each of the perturbations, and define the 𝑓 𝑟𝑎𝑔-value of 𝑠𝑡𝑠 as the fraction of the 10.000 random perturbations that have the opposite classification as 𝑡𝑠.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>( a )</head><label>a</label><figDesc>𝑓 𝑟𝑎𝑔 = 0.45 blue. Not robust. (b) 𝑓 𝑟𝑎𝑔 = 0.04 blue. Very robust. (c) 𝑓 𝑟𝑎𝑔 = 0.02 red. Very robust.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Visualizing robustness of the three simplified time series in dotted black lines. All perturbations consist of straight-line segments that are forced to lie inside the colored band, with pivot points at same 𝑥-values as the simplification, but different 𝑦-values. We compute several perturbations of 𝑠𝑡𝑠 by randomly altering the 𝑦-values of the 𝑘 𝑠𝑡𝑠 + 1 pivot points inside the narrow band, e.g. in (b) 𝑠𝑡𝑠 in black has 5 segments and the 6 pivot points of the perturbations are at 𝑥-values 0,5,9,13,18,24. Each perturbation is drawn with a color according to its classification by the classifier ℎ. Thus if many have same color as 𝑡𝑠 then robustness is good (and 𝑓 𝑟𝑎𝑔-value low).</figDesc><graphic coords="7,72.00,486.17,139.89,104.92" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Six prototypes from Chinatown. Hard to make a rule-of-thumb.</figDesc><graphic coords="8,72.00,538.90,451.27,188.03" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: Same six prototypes from Chinatown, now shown in dotted black lines, with robust simplifications making it easier to extract a rule-of-thumb.</figDesc><graphic coords="9,72.00,65.61,451.27,188.03" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head>(a) 11 Figure 5 :Figure 6 :</head><label>1156</label><figDesc>Figure 5: Prototype from Chinatown in black and four distinct simplifications of it in red. From (a) to (b) to(c) we see that increasing the importance of simplicity from low to middle to high will decrease the number of segments from 11 to 6 to 2, and increase the Euclidean distance. In (d) we see that increasing the importance of robustness, when comparing to (b) which also has 6 segments, changes the slope of the first segment even though this increases the Euclidean distance, which is significant information about the red classification.</figDesc><graphic coords="10,103.96,421.50,170.27,127.70" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_7"><head>(a) 5 Figure 7 :</head><label>57</label><figDesc>Figure 7: Prototype from ECG200 in black and two distinct simplifications of it in blue. Note that even though the factor for robustness is much higher in (b) than (a), the two simplifications are almost identical (difference being that in (a) the third segment ends in 𝑥 = 40 whereas in (b) it ends in 𝑥 = 39).</figDesc><graphic coords="11,138.44,65.61,157.95,118.46" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_8"><head>•</head><label></label><figDesc>Simple Rule for ItalyPowerDemand: let 𝑚𝑎𝑥1 = maximum of the 𝑦-values over 𝑥 = 9, 10, 11, let 𝑚𝑖𝑛2= minimum of the 𝑦-values over 𝑥 = 13, 14, 15 and 𝑚𝑎𝑥3 = maximum of the 𝑦-values over 𝑥 = 17, 18, 19. If 𝑚𝑎𝑥1 − 𝑚𝑖𝑛2 ≥ 0.5 and 𝑚𝑎𝑥3 − 𝑚𝑖𝑛2 ≥ 0.5 then classify this instance Blue, else Red.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_9"><head>Figure 8 :</head><label>8</label><figDesc>Figure 8: Visualization of robustness of a time series starting at (0, 300) suggesting that a cutoff somewhat higher than (0, 250) for Blue vs Red would have been a better guess for the Simple Rule for Chinatown.</figDesc><graphic coords="13,162.25,65.61,270.78,203.08" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head></head><label></label><figDesc>Stage 1 Use DP to compute a 2-dimensional table of size 𝑛 × 𝑞 that in cell 𝐷[𝑗, 𝑘] stores the 𝑘th least costly simplification for the points 𝑝 1 , ..., 𝑝 𝑗 of the input time series 𝑡𝑠, considering only error and simplicity but not robustness, as in above DP. Stage 2 Consider the 𝑞 least costly simplifications 𝑠𝑡𝑠 1 , ..., 𝑠𝑡𝑠 𝑞 , ordered by non-decreasing cost under error and simplicity as stored in 𝐷[𝑛, 1]..𝐷[𝑛, 𝑞] respectively. Compute robustness for any simplification that ℎ classifies same as 𝑡𝑠. The ORS-Algorithm then returns the simplification having lowest overall total cost, i.e. the 𝑠𝑡𝑠 minimizing the value of value of 𝛼 • 𝑒𝑟𝑟(𝑡𝑠, 𝑠𝑡𝑠) + 𝛽 • 𝑘 𝑠𝑡𝑠 + 𝛾 • 𝑓 𝑟𝑎𝑔(𝑡𝑠, 𝑠𝑡𝑠, ℎ)</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Definition of Perturbations and Robustness in Stage 2.</head><label></label><figDesc>1, computing a table of size 𝑛 × 𝑞 with 𝐷[𝑗, 𝑘] storing the 𝑘th least costly simplification of points 𝑝 1 , ..., 𝑝 𝑗 under error and simplicity only. The difference to the earlier DP is that we compute not only the least costly but the 𝑞 least costly with 𝑞 &gt; 1. After computing error terms we run an outer loop from 𝑗 = 1 to 𝑛 with two inner loops one after the other. The first loop from 𝑖 = 1 to 𝑗 − 1 is similar to the previous DP, computing the cost 𝐷[𝑖, 1] + 𝑒𝑟𝑟(𝑖, 𝑗) of having the last segment go from 𝑝 𝑖 to 𝑝 𝑗 , but now adding all these values to a heap 𝐻 𝑗 . The second loop from 𝑘 = 1 to 𝑞 fills the 𝐷[𝑗, 𝑘] entries by extracting the minimum element from 𝐻 𝑗 , say this element is 𝐷[𝑖, 𝑟] + 𝑒𝑟𝑟(𝑖, 𝑗), then first set 𝐷[𝑗, 𝑘] to this element, but most importantly also add to the heap 𝐻 𝑗 the new value 𝐷[𝑖, 𝑟 + 1] + 𝑒𝑟𝑟(𝑖, 𝑗) since the 𝑟 + 1st least costly solution up to 𝑝 𝑖 may be lower than many other solutions in the heap. For details, in particular regarding the edge-cases which are cumbersome to implement properly, see the pseudo-code in the appendix.</figDesc><table><row><cell>4.2.2.</cell></row></table><note>A simplified time series 𝑠𝑡𝑠 on 𝑘 segments is computed by choosing 𝑘 + 1 points 𝑝 𝑖 1 , ..., 𝑝 𝑖 𝑘+1 from an original time series 𝑡𝑠 on points 𝑝 1 , ..., 𝑝 𝑛 . When computing the robustness of 𝑠𝑡𝑠 focus on the following 𝑘 + 1 pivot points 𝑠 1 , 𝑝 𝑖 2 , ..., 𝑝 𝑖 𝑘 , 𝑠 2 of 𝑠𝑡𝑠 (note the first and last may not be points of 𝑡𝑠)</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head></head><label></label><figDesc>The initialization of error terms is 𝑂(𝑛3 ) as in the first DP version. We use binary heaps having 𝑂(log 𝑛) insertion and extraction operations. Then, for each of the 𝑛 values of 1 ≤ 𝑗 ≤ 𝑛 we have two for-loops, one inserting at most 𝑛 values into heap 𝐻 𝑗 (for time 𝑂(𝑛 log 𝑛)), and the other extracting and inserting at most 𝑞 values of 𝐻 𝑗 (for time 𝑂(𝑞 log 𝑛)). Note there are never more than 𝑛 elements in any heap 𝐻 𝑗 as the first loop never inserts more than 𝑛 elements and the latter loop extracts one element and inserts at most one new element. This completes the proof. The optimal simplification 𝑜𝑝𝑡 𝑠𝑖𝑚𝑝 (𝑡𝑠, ℎ, 𝛼, 𝛽, 𝛾) can be computed in polynomial time for a time series 𝑡𝑠 on 𝑛 points, a binary TSC ℎ, and balancing factors 𝛼, 𝛽, 𝛾, if there exists a constant 𝑐, such that the difference is ≥ 𝛾 between the least costly simplification that does not take robustness into account and the 𝑛 𝑐 th least costly one. This follows from Theorems 2 and 3 by running the ORS Algorithm with 𝑞 = 𝑛 𝑐 .</figDesc><table><row><cell>Theorem 4. Proof 3.</cell></row></table><note>Theorem 3. The runtime of the ORS-ALgorithm is 𝑂(𝑛 • max{𝑛 2 , 𝑞 log 𝑛}).Proof 2.</note></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">Note that if one pieces together several such local explanations, say for prototypes of each class, then this can still form a global explanation of the given TSC ℎ.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">Available on Github: https://github.com/BrigtHaavardstun/kSimplification</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">Exact values used for parameters can be found on the github page. Here we only refer to them as Low-Mid-High to keep the presentation simple.</note>
		</body>
		<back>
			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0" />			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Scalable and accurate deep learning with electronic health records</title>
		<author>
			<persName><forename type="first">A</forename><surname>Rajkomar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Oren</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">M</forename><surname>Dai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Hajaj</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hardt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">J</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Marcus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Sun</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">NPJ digital medicine</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="1" to="10" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<title level="m" type="main">Time-series classification methods: Review and applications to power systems data, Big data application in power systems</title>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">A</forename><surname>Susto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Cenedese</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Terzi</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="179" to="220" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Deep learning for time series classification: a review</title>
		<author>
			<persName><forename type="first">H</forename><surname>Ismail Fawaz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Forestier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Weber</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Idoumghar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P.-A</forename><surname>Muller</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Data mining and knowledge discovery</title>
		<imprint>
			<biblScope unit="volume">33</biblScope>
			<biblScope unit="page" from="917" to="963" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Agnostic local explanation for time series classification</title>
		<author>
			<persName><forename type="first">M</forename><surname>Guillemé</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Masson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Rozé</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Termier</surname></persName>
		</author>
		<idno type="DOI">10.1109/ICTAI.2019.00067</idno>
		<ptr target="https://doi.org/10.1109/ICTAI.2019.00067.doi:10.1109/ICTAI.2019.00067" />
	</analytic>
	<monogr>
		<title level="m">IEEE International Conference on Tools with Artificial Intelligence, ICTAI 2019</title>
				<meeting><address><addrLine>Portland, OR, USA</addrLine></address></meeting>
		<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2019">November 4-6, 2019. 2019</date>
			<biblScope unit="page" from="432" to="439" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Benchmarking deep learning interpretability in time series predictions</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">A</forename><surname>Ismail</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">K</forename><surname>Gunady</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">C</forename><surname>Bravo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Feizi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Annual Conference on Neural Information Processing Systems 2020</title>
				<meeting><address><addrLine>NeurIPS; virtual</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2020-12-06">2020. December 6-12, 2020. 2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Instance-based counterfactual explanations for time series classification</title>
		<author>
			<persName><forename type="first">E</forename><surname>Delaney</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Greene</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">T</forename><surname>Keane</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Case-Based Reasoning Research and Development -Int. Conference, ICCBR</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2021">2021. 2021</date>
			<biblScope unit="page" from="32" to="47" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Towards a rigorous evaluation of xai methods on time series</title>
		<author>
			<persName><forename type="first">U</forename><surname>Schlegel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Arnout</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>El-Assady</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Oelke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">A</forename><surname>Keim</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE/CVF International Conference on Computer Vision Workshop (ICCVW)</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2019">2019. 2019</date>
			<biblScope unit="page" from="4197" to="4201" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Tsviz: Demystification of deep learning models for time-series analysis</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">A</forename><surname>Siddiqui</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Mercier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Munir</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Dengel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ahmed</surname></persName>
		</author>
		<idno type="DOI">10.1109/ACCESS.2019.2912823</idno>
		<ptr target="https://doi.org/10.1109/ACCESS.2019.2912823.doi:10.1109/ACCESS.2019.2912823" />
	</analytic>
	<monogr>
		<title level="j">IEEE Access</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="page" from="67027" to="67040" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">An interactive tool for interpretability of time series classification</title>
		<author>
			<persName><forename type="first">B</forename><surname>Håvardstun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Ferri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>Telle</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note>To be published</note>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Explainable AI for time series classification: A review, taxonomy and research directions</title>
		<author>
			<persName><forename type="first">A</forename><surname>Theissler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Spinnato</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><surname>Schlegel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Guidotti</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Access</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="page" from="100700" to="100724" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Explaining time series classifiers through meaningful perturbation and optimisation</title>
		<author>
			<persName><forename type="first">H</forename><surname>Meng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Wagner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Triguero</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.ins.2023.119334</idno>
		<ptr target="https://doi.org/10.1016/j.ins.2023.119334" />
	</analytic>
	<monogr>
		<title level="j">Information Sciences</title>
		<imprint>
			<biblScope unit="volume">645</biblScope>
			<biblScope unit="page">119334</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Learning perturbations to explain time series predictions</title>
		<author>
			<persName><forename type="first">J</forename><surname>Enguehard</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference on Machine Learning</title>
				<meeting><address><addrLine>PMLR</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="9329" to="9342" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Explaining time series via contrastive and locally sparse perturbations</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Luo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Du</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Fan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Wen</surname></persName>
		</author>
		<ptr target="https://openreview.net/forum?id=qDdSRaOiyb" />
	</analytic>
	<monogr>
		<title level="m">The Twelfth International Conference on Learning Representations</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<author>
			<persName><forename type="first">T</forename><surname>Rojat</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Puget</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Filliat</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Del</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Ser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Gelin</surname></persName>
		</author>
		<author>
			<persName><surname>Díaz-Rodríguez</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2104.00950</idno>
		<title level="m">Explainable artificial intelligence (xai) on timeseries data: A survey</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Example or prototype? learning concept-based explanations in time-series</title>
		<author>
			<persName><forename type="first">C</forename><surname>Obermair</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Fuchs</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Pernkopf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Felsberger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Apollonio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Wollmann</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Asian Conference on Machine Learning</title>
				<meeting><address><addrLine>PMLR</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="816" to="831" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Weighted kNN and constrained elastic distances for time-series classification</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Geler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Kurbalija</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ivanović</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Radovanović</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Expert Systems with Applications</title>
		<imprint>
			<biblScope unit="volume">162</biblScope>
			<biblScope unit="page">113829</biblScope>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<title level="m" type="main">Explaining deep classification of time-series data with learned prototypes</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">H</forename><surname>Gee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>García-Olano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Ghosh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Paydarfar</surname></persName>
		</author>
		<ptr target="https://ceur-ws.org/Vol-2429/paper3.pdf" />
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Dimensionality reduction for fast similarity search in large time series databases</title>
		<author>
			<persName><forename type="first">E</forename><surname>Keogh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Chakrabarti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Pazzani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Mehrotra</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Knowledge and information Systems</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="page" from="263" to="286" />
			<date type="published" when="2001">2001</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Segmenting time series: A survey and novel approach</title>
		<author>
			<persName><forename type="first">E</forename><surname>Keogh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Chu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Hart</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Pazzani</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Data mining in time series databases</title>
				<imprint>
			<publisher>World Scientific</publisher>
			<date type="published" when="2004">2004</date>
			<biblScope unit="page" from="1" to="21" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Models and algorithms for optimal piecewise-linear function approximation</title>
		<author>
			<persName><forename type="first">E</forename><surname>Camponogara</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">F</forename><surname>Nazari</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Mathematical Problems in Engineering</title>
		<imprint>
			<biblScope unit="volume">2015</biblScope>
			<biblScope unit="page">876862</biblScope>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<monogr>
		<author>
			<persName><forename type="first">U</forename><surname>Schlegel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">V</forename><surname>Lam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">A</forename><surname>Keim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Seebacher</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2109.08438</idno>
		<title level="m">Ts-mule: Local interpretable model-agnostic explanations for time series forecast models</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Effect of segmentation on financial time series pattern matching</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Wan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Gong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y.-W</forename><surname>Si</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.asoc.2015.10.012</idno>
		<ptr target="https://doi.org/10.1016/j.asoc.2015.10.012" />
	</analytic>
	<monogr>
		<title level="j">Applied Soft Computing</title>
		<imprint>
			<biblScope unit="volume">38</biblScope>
			<biblScope unit="page" from="346" to="359" />
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Obst-based segmentation approach to financial time series</title>
		<author>
			<persName><forename type="first">Y.-W</forename><surname>Si</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Yin</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.engappai.2013.08.015</idno>
		<ptr target="https://doi.org/10.1016/j.engappai.2013.08.015" />
	</analytic>
	<monogr>
		<title level="j">Engineering Applications of Artificial Intelligence</title>
		<imprint>
			<biblScope unit="volume">26</biblScope>
			<biblScope unit="page" from="2581" to="2596" />
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">XAI for time series classification: Evaluating the benefits of model inspection for end-users</title>
		<author>
			<persName><forename type="first">B</forename><surname>Håvardstun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Ferri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Flikka</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>Telle</surname></persName>
		</author>
		<ptr target="https://xaiworldconference.com/2024/" />
	</analytic>
	<monogr>
		<title level="m">Explainable Artificial Intelligence</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note>to be published</note>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">GAM coach: Towards interactive and usercentered algorithmic recourse</title>
		<author>
			<persName><forename type="first">Z</forename><forename type="middle">J</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">W</forename><surname>Vaughan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Caruana</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">H</forename><surname>Chau</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. Conf. on Human Factors in Computing Systems</title>
				<meeting>Conf. on Human Factors in Computing Systems</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2023">2023</date>
			<biblScope unit="volume">835</biblScope>
			<biblScope unit="page">20</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Tsinterpret: A python package for the interpretability of time series classification</title>
		<author>
			<persName><forename type="first">J</forename><surname>Höllig</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Kulbach</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Thoma</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">J. Open Source Softw</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="page">5220</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Interpretable time-series classification on few-shot samples</title>
		<author>
			<persName><forename type="first">W</forename><surname>Tang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Long</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">2020 International Joint Conference on Neural Networks (IJCNN), IEEE</title>
				<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="1" to="8" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">Interpretable and steerable sequence learning via prototypes</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Ming</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Qu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Ren</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery &amp; Data Mining</title>
				<meeting>the 25th ACM SIGKDD International Conference on Knowledge Discovery &amp; Data Mining</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="903" to="913" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<monogr>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">A</forename><surname>Dau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Keogh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Kamgar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C.-C</forename><forename type="middle">M</forename><surname>Yeh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Gharghabi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">A</forename><surname>Ratanamahatana</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Yanping</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Begum</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bagnall</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Mueen</surname></persName>
		</author>
		<author>
			<persName><surname>Batista</surname></persName>
		</author>
		<title level="m">The UCR time series classification archive</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
	<note>Hexagon-ML</note>
</biblStruct>

<biblStruct xml:id="b29">
	<monogr>
		<author>
			<persName><forename type="first">Z</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Yan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Oates</surname></persName>
		</author>
		<idno>CoRR abs/1611.06455</idno>
		<ptr target="http://arxiv.org/abs/1611.06455.arXiv:1611.06455" />
		<title level="m">Time series classification from scratch with deep neural networks: A strong baseline</title>
				<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<monogr>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">S</forename><surname>Gurumoorthy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Jawanpuria</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Mishra</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2103.10159</idno>
		<title level="m">Spot: A framework for selection of prototypes using optimal transport</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b31">
	<monogr>
		<author>
			<persName><forename type="first">I</forename><surname>Contributors</surname></persName>
		</author>
		<ptr target="https://github.com/interpretml/interpret" />
		<title level="m">Interpretml: Fit interpretable models. explain blackbox machine learning</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note>GitHub repository</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
