<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">GLEAMS: Bridging the Gap Between Local and Global Explanations</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Giorgio</forename><surname>Visani</surname></persName>
							<email>giorgio.visani@leitha.eu</email>
							<affiliation key="aff0">
								<orgName type="department">Leithà S.r.l. -Unipol Group</orgName>
								<address>
									<addrLine>via Stalingrado 53</addrLine>
									<settlement>Bologna</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Vincenzo</forename><surname>Stanzione</surname></persName>
							<email>v.m.stanzione@gmail.com</email>
							<affiliation key="aff1">
								<orgName type="department">Computer Science Department</orgName>
								<orgName type="institution">Bologna University</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Damien</forename><surname>Garreau</surname></persName>
							<email>damien.garreau@uni-wuerzburg.de</email>
							<affiliation key="aff2">
								<orgName type="institution">Julius-Maximilians-Universität Würzburg</orgName>
								<address>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">GLEAMS: Bridging the Gap Between Local and Global Explanations</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">68EE83C656819CA6453345DA36788C4D</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T18:48+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Explainable AI</term>
					<term>Global / local</term>
					<term>Partition-based machine learning</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The explainability of machine learning algorithms is crucial, and numerous methods have emerged recently. Local, post-hoc methods assign an attribution score to each feature, indicating its importance for the prediction. However, these methods require recalculating explanations for each example. On the other side, while there exist global approaches they often produce explanations that are either overly simplistic and unreliable or excessively complex. To bridge this gap, we propose GLEAMS, a novel method that partitions the input space and learns an interpretable model within each sub-region, thereby providing both faithful local and global surrogates. We demonstrate GLEAMS' effectiveness on both synthetic and real-world data, highlighting its desirable properties and human-understandable insights.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>In numerous applications, the data present themselves as large arrays-tabular data, where each line corresponds to a data point and each column to a feature. In this setting, deep learning is becoming as popular as it already is for more structured data types, and deep neural networks achieve state-of-the-art in many scenarios <ref type="bibr" target="#b0">[1]</ref>. A caveat of this approach is the difficulty to obtain insights on a specific prediction: the intricate architectures and the sheer number of parameters prevent the user of getting a clear picture of what is actually important to the model.</p><p>Since getting such explanations is desirable in many applications, and could even become mandated by law in certain regions <ref type="bibr" target="#b1">[2]</ref>, many interpretability methods have been proposed. Without getting into details (we refer to the recent surveys <ref type="bibr" target="#b2">[3]</ref> and <ref type="bibr" target="#b3">[4]</ref> for this purpose), we want to make two distinctions clear. First, it is challenging for interpretable models to have the same accuracy as non-interpretable ones. As a consequence, many explainability methods rely on a post hoc approach, i.e., producing explanations in a second step, exploiting already trained black-box models. This is the road we take, since post-hoc methods allow us to deal with any given model and not being dependent on a specific architecture. Second, the explanations can either be local (related to a specific example), or global (for all the input space). A problem with local explanations is that they have to be computed individually for each new example, which can be time-consuming. For instance, LIME <ref type="bibr" target="#b4">[5]</ref> generates 5, 000 new datapoints for each example to explain. The black-box model then has to be queried on each of these points.</p><p>In this paper, we propose GLEAMS (Global &amp; Local ExplainAbility of black-box Models through Space partitioning), a post hoc interpretability method providing both global and local explanations. Our goal is to build a global surrogate 𝑓 ^of 𝑓 that shares the global shape of the black-box model 𝑓 but, at the same time, has a simple local structure. Our main motivation in doing so is to extract simultaneously from 𝑓 ^: i) local explanations for any instance; ii) overall importance of the features. Secondary objectives are the production of "what-if"-type explanations, and visual summaries. In more details, the core idea of GLEAMS is to recursively split the input space in rectangular cells on which the black-box model can be well-approximated by linear models. We describe the method in Section 2. The local explanations are then simply given by the feature coefficients of these local surrogate models, while global indicators can be readily computed from those, without querying 𝑓 model further. We describe how to obtain the other indicators in Section 2.4, and show experimentally that GLEAMS compares favorably with existing methods in Section 3. GLEAMS code as well as the experiments for this paper are publicly available. <ref type="foot" target="#foot_0">1</ref> Related work can be found in Appendix A. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Description of the method</head><p>We consider the following general setting: 𝑓 is a mapping from the input space 𝒳 to the output space 𝒴. This corresponds to the usual situation in machine learning, where 𝑦 = 𝑓 (𝑥) is a good prediction of the true label of 𝑥. In this paper, we assume that 𝒴 ⊆ R. We want to stress that this can correspond both to the classification and regression setting. In the regression setting, for any input 𝑥 ∈ 𝒳 , 𝑓 (𝑥) is simply a real-valued quantity of interest, whereas in the classification case 𝑓 (𝑥) corresponds to the pseudo-probability to fall into a given class.</p><p>The construction of the global surrogate on which GLEAMS relies to provide explanations is a two-step process. The first step is to sample 𝑁 points 𝑥 1 , . . . , 𝑥 𝑁 , evenly spread in 𝒳 . For each of these points, we query 𝑓 to obtain 𝑦 𝑖 = 𝑓 (𝑥 𝑖 ). Thus we effectively create a dataset (𝑥 𝑖 , 𝑦 𝑖 ), which we call measurement points.</p><p>Subsequently, the idea is to recursively split the input space 𝒳 and to approximate the model 𝑓 by a linear model on each cell of the partition until the linear model in each rectangular leaf fits the measurement points with a sufficiently good accuracy. This process is summarized in Figure <ref type="figure" target="#fig_0">1</ref>. Once this global surrogate constructed, GLEAMS can provide both local and global explanations for any new example to explain. Intuitively, given a partition with not too many cells, it is convenient for the user to look at individual features and to visualize globally the model as a piecewise-linear function. Concurrently, any new example to explain will fall into a cell, and the user can get a local explanation solely by looking at the coefficients of the local linear model. We now give more details on each of these steps.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Measurement points</head><p>Our main assumption, going into the description of the method, is the following: Assumption 1. The input space 𝒳 is a hyper-rectangle of R 𝑑 . That is, there exist reals numbers</p><formula xml:id="formula_0">𝑎 𝑗 , 𝑏 𝑗 ∈ R for any 𝑗 ∈ [𝑑] such that 𝒳 = ∏︀ 𝑑 𝑗=1 [𝑎 𝑗 , 𝑏 𝑗 ].</formula><p>Intuitively, this corresponds to a situation where each variable has a prescribed range (for instance, age takes values in [0, 120]). In particular, we assume that further examples to explain all fall within this range. If they do not, we attribute to these points the explanations corresponding to the projection of these points on 𝒳 . Note that we do not assume normalized data: 𝑎 𝑗 and 𝑏 𝑗 can be arbitrary. In practice, the boundaries are either inferred from a training set, or directly provided by the user.</p><p>A consequence of Assumption 1 is that we can easily find a set of points covering 𝒳 efficiently. To achieve this, we use Sobol sequences <ref type="bibr" target="#b5">[6]</ref>. In a nutshell, a Sobol sequence is a quasi-random sequence of points 𝑥 1 , . . . , 𝑥 𝑁 filling the unit cube [0, 1] 𝑑 as 𝑁 grows. We use Sobol sequences because (a) we want to be able to provide explanations for the whole input space 𝒳 , thus the measurement points should cover 𝒳 as 𝑁 grows; and (b) the decision surface of the black-box model can be quite irregular, thus we want low discrepancy.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Splitting criterion</head><p>A very convenient way to partition the input space is to recursively split 𝒳 , cutting a cell if some numerical criterion is satisfied. Additionally, we consider splits that are parallel to the axes. The main reason for doing so is the tree structure that emerges, allowing later on for fast query time of new inputs (linear in the depth of the final tree). In this section, we describe the criterion used by GLEAMS.</p><p>A popular approach is to fit a constant model on each cell <ref type="bibr" target="#b6">[7,</ref><ref type="bibr">CART]</ref>. We see two problems with constant fits. First, this tends to produce large and thus hard to interpret trees (see, e.g., Chan and Loh <ref type="bibr" target="#b7">[8]</ref>). Second, the local interpretability would be lost, since the local models would not depend on the features in that case. Thus we prefer to fit linear models on each leaf, and our numerical criterion should indicate, on any given leaf, if such simple model is a good fit for 𝑓 . GLEAMS relies on a variant of model-based recursive partitioning <ref type="bibr" target="#b8">[9,</ref><ref type="bibr">MOB]</ref> based on score computations, described in detail in Appendix B.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3.">Global surrogate model</head><p>To build the global surrogate model 𝑓 ^, starting from the hyper-rectangle 𝒳 , we iteratively split along the best dimension. This process is stopped once one of two criteria is met: i) leaves have 𝑅 2 higher than 0.95 (chosen empirically) or ii) number of points in the leaf is less than 𝑛 min threshold, set to 𝑛 min = min(20, 𝑑 + 1). This recursive procedure is described in Algorithm 2, in Appendix B. 𝑓 ^is then achieved by sampling Sobol points on 𝒳 , computing the value of the black-box model at these points, and then using Algorithm 2, Appendix B.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.4.">GLEAMS</head><p>Let us now assume that we computed the global surrogate model 𝑓 ^, that is, a piecewise-linear model based on a partition of 𝒳 . At query time, each new example to explain 𝜉 ∈ 𝒳 is rooted to the corresponding hyper-rectangle 𝑅 in the tree structure, and thus associated to a linear model 𝛽 ^𝜉. From this local linear model, we can directly get different sorts of explanations.</p><p>Local explanations. Simply looking at the coefficient 𝛽 ^𝜉 𝑗 tells us how important feature 𝑗 is to predict 𝑓 (𝜉), locally. The intuition here is that 𝑓 is well-approximated by 𝑥 ↦ → (𝛽 ^𝜉) ⊤ 𝑥 on the cell 𝑅, thus local variations of the model along specific axes are well-described by the coefficients of 𝛽 ^𝜉. These local explanations can be obtained in constant time, without re-sampling and querying the model. Moreover, we have precise insights regarding the scale of these explanations, which is given by the size of 𝑅. We see this as an advantage of using GLEAMS: the portion of space on which the black-box model is locally approximated is well-identified and accessible to the user. This is not the case with other methods such as LIME.</p><p>Global explanations. We define the global importance of feature 𝑗 as the aggregation of the local importance of feature 𝑗 on all cells of the partition 𝒫 of 𝒳 . There are several ways to aggregate those, GLEAMS outputs the weighted average of the absolute value of the local importance. More precisely, let us define 𝛽 ^𝑅 the local model on hyper-rectangle 𝑅 ∈ 𝒫, then the global importance of feature 𝑗 is given by</p><formula xml:id="formula_1">𝐼 𝑗 := 1 𝒱(𝒳 ) ∑︁ 𝑅∈𝒫 𝒱(𝑅) ⃒ ⃒ ⃒𝛽 ^𝑅 𝑗 ⃒ ⃒ ⃒ ,<label>(1)</label></formula><p>where 𝒱(𝐸) denotes the 𝑑-dimensional volume of 𝐸. The motivation for weighting by the cells' volume in Eq. ( <ref type="formula" target="#formula_1">1</ref>) is to take into account the fact that even though feature 𝑗 is very important in a tiny region of the space, if it is not on all other cells then it is not globally important.</p><p>We take the absolute value to take into account the possible sign variations: if feature 𝑗 is positively important on many cells but negatively important on a similar number of cells, then simply taking the average would tell us that the feature is not globally important. Such global attributions provide a faithful variables ranking, based on the impact they show on 𝑓 predictions. </p><p>This can be computed quickly, since, again, there is no sampling and further calls to 𝑓 . This allows us to present the user with a cursor which can be moved to set the feature value and present explanations in real time (see Figure <ref type="figure" target="#fig_1">2</ref>).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Experiments on real data</head><p>Our goal here is to show that GLEAMS provides explanations whose quality is comparable to that of other attribution approaches, while providing these attributions in constant time and additional insights. We benchmark the various methods on the Wine dataset <ref type="bibr" target="#b9">[10]</ref>, House sell dataset<ref type="foot" target="#foot_1">2</ref> and Parkinson telemonitoring dataset <ref type="bibr" target="#b10">[11]</ref>. We refer to Appendix D for more experiments.</p><p>Models. We considered two different models, trained similarly across all datasets. The first is an XGBoost model <ref type="bibr" target="#b11">[12]</ref>. XGBoost iteratively aggregates base regressors greedily minimizing a proxy loss. Following Friedman <ref type="bibr" target="#b12">[13]</ref> prescriptions, we employ simple CART trees with maximum depth 2 as base regressors, learning rate of 0.05, and early stopping rounds set to 100. The second is a multi-layer perceptron (MLP), composed by two hidden dense layers of 264 neurons each, trained by adaptive stochastic gradient descent [14, ADAM] and early stopping on a hold-out validation set. Both these models achieve state-of-the-art accuracy in many settings and are not inherently interpretable. Methods. We compared GLEAMS to three methods, namely LIME for tabular data, SHAP, and PDP. We used default parameters, except for the number of feature combinations tested in SHAP for the recall metric, which were shrinked to 500 to limit computing time. GLEAMS results have been obtained using 𝑁 = 15 (number of Sobol points), which guaranteed a running time in the order of 5 minutes for each of the three datasets at hand.</p><p>Evaluation. We compare GLEAMS to existing methodology using monotonicity and recall.</p><p>For each interpretability method to evaluate, for each point in the test set, we compute the vectors of attributions 𝛼 ∈ R 𝑑 and expected loss 𝑒 ∈ R 𝑑 restricted to the corresponding feature. Monotonicity is then defined as the Spearman rank correlation between |𝛼| and 𝑒 <ref type="bibr" target="#b14">[15]</ref>. Depending on the region where the 𝑒 values are computed we obtain local monotonicity, i.e., inside each specific GLEAMS leaf, or global monotonicity when considering the whole input space. Let us now consider a model trained to use only a subset 𝑇 of the features. We can actually be certain that some feature attributions should be 0. For any given example 𝜉, we can define 𝑇 𝜉 the set of the top |𝑇 | features, ranked by |𝛼 𝑗 |. Recall of Important Features <ref type="bibr" target="#b4">[5]</ref> is the fraction of 𝑇 features retrieved in 𝑇 𝜉 , averaged among the test units. More on this in the Appendix C.</p><p>Results. We present our results in Table <ref type="table" target="#tab_0">1</ref>, reporting the average metrics taken on the test set. All evaluation metrics are described in Appendix C. PDP generally achieves better monotonicity metrics, since it is designed precisely to obtain feature attributions related to the feature impact on the predictions, but it performs worse on the recall task. We notice that GLEAMS shows better results on local and global monotonicity compared to LIME and SHAP on the wine dataset, while GLEAMS performances start to degrade with an increased number of features. In particular we notice the poor performance on local monotonicity of the MLP model for the Parkinson dataset. The degradation is due to the higher dimensionality of the two datasets, which makes models surface more complex (MLP in particular). Increasing 𝑁 would likely improve the results, enabling smaller partitions and more accurate local models. Unfortunately, we did not have time to run experiments with higher 𝑁 . Regarding the recall metric, SHAP stands as the best in class, while GLEAMS is usually slightly worse but still comparable with LIME, and systematically better than PDP.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Conclusion</head><p>In this paper, we presented GLEAMS, a new interpretability method that approximates a blackbox model by recursively partitioning the input space and fitting linear models on the elements of the partition. Both local and global explanations can then be obtained in constant time, for any new example, in sharp contrast with existing post hoc methods such as LIME and SHAP.</p><p>The main limitation of the method is the assumption that all features are continuous. As future work, we aim to extend GLEAMS to mixed features (both categorical and continuous). Another limitation of the method is the initial number of model queries needed. One possible way to reduce it would be to evaluate the local regularity of 𝑓 and to sample less when the model is smooth.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A. Related work</head><p>The idea of approximating globally a complex model such as a deep neural network by a simpler surrogate model relying on a tree-based partition of the space can be dated back at Craven and Shavlik <ref type="bibr" target="#b15">[16]</ref>. The proposed method, TREPAN, approximates the outputs of the network on the training set by a decision tree, choosing the splits using the gain ratio criterion <ref type="bibr" target="#b16">[17]</ref>. More recently, inspired by Gibbons et al. <ref type="bibr" target="#b17">[18]</ref>, Zhou and Hooker <ref type="bibr" target="#b18">[19]</ref> proposed to use another split criterion, based upon an asymptotic expansion of the Gini criterion. Both these approaches are limited to the classification setting, and also require access to the training data (which is implicitly assumed to have a good covering of the input space).</p><p>From the interpretability side, a standard way to explain 𝑓 , which is closely related to our approach, is to compute attributions scores for each feature. Lei et al. <ref type="bibr" target="#b19">[20]</ref> exclude a specific feature from 𝑓 and define its impact as the deterioration of model predictions. On the contrary, Partial Dependence Plots (PDP) developed by Friedman <ref type="bibr" target="#b12">[13]</ref> measure the sensitivity of 𝑓 to changes in the specific feature, isolating its effect from the ones of the other features. Ribeiro et al. <ref type="bibr" target="#b4">[5]</ref> propose LIME, which proposes as attribution for a specific example the coefficients of a linear model approximating 𝑓 locally. Similarly, Lundberg and Lee <ref type="bibr" target="#b20">[21]</ref> introduce the idea of decomposing 𝑓 predictions into single variables attributions, using SHAP. The technique samples combinations of features 𝑆 and average changes in single instance prediction between the restricted 𝑓 (𝑆 ∪ 𝑖) against 𝑓 (𝑆), obtaining local attributions for the feature 𝑖. Global SHAP attributions arise by averaging local attributions over the entire dataset.</p><p>Attribution methods usually have to be recomputed for each new example, a caveat that we propose to overcome with GLEAMS, computing a global surrogate and relying on it to provide explanations. We are not the first to propose to provide explanations on a global scale: Ribeiro et al. <ref type="bibr" target="#b21">[22]</ref> propose Anchors, a method extracting local subsets of the features (anchors) that are sufficient to recover the same prediction, while having a good global coverage. A caveat of the method is that local explanations are not quantitative since all members of a given anchor are equivalent. Additionally, the anchors can have many elements if the example at hand is near the decision boundary. Let us also mention Setzu et al. <ref type="bibr" target="#b22">[23]</ref>, which proposes to aggregate local explanations: starting from local decision rules, GlocalX combines them in a hierarchical manner to form a global explanation of the model. In the ad hoc setting, Harder et al. <ref type="bibr" target="#b23">[24]</ref> proposes to train an interpretable and differentially private model by using several local linear maps per class: the pseudo-probability of belonging to a given class is the softmax of a mixture of linear models. This approach is limited to the classification setting, as with GlocalX and Anchors.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>B. Splitting criterion</head><p>Intuitively, any well-behaved function can be locally approximated by a linear component, yielding a statistical model which we now describe. On any given leaf, for any given feature 𝑗, we can reorder the measurement points such that the points contained in the leaf are 𝑥 1 , . . . , 𝑥 𝑛 , with 𝑥 1,𝑗 ≤ • • • ≤ 𝑥 𝑛,𝑗 . To these points correspond model values 𝑦 1 , . . . , 𝑦 𝑛 . In accordance with the intuition given above, we write 𝑦 𝑖 = 𝛽 ⊤ 𝑥 ˜𝑖 + 𝜖 𝑖 for all 𝑖 ∈ [𝑛], where 𝛽 ∈ R 𝑑+1 are the coefficients of our local linear model, 𝑥 ˜𝑖 ∈ R 𝑑+1 is the concatenation of 𝑥 𝑖 with a leading 1 to take the intercept into account, and 𝜖 𝑖 is an error term. Let us now follow Merkle and Zeileis <ref type="bibr" target="#b24">[25]</ref> and make an i.i.d. Gaussian assumption on the 𝜖 𝑖 s, giving rise to the likelihood</p><formula xml:id="formula_3">ℒ(𝛽; 𝑥 𝑖 ) = 1 (2𝜋𝜎 2 ) 𝑑/2 exp (︂ −1 2𝜎 2 (𝑦 𝑖 − 𝛽 ⊤ 𝑥 ˜𝑖) 2 )︂ ,<label>(3)</label></formula><p>where 𝜎 2 &gt; 0 is the variance of 𝜖 𝑖 for all 𝑖 ∈ [𝑛]. Taking the log in Eq. ( <ref type="formula" target="#formula_3">3</ref>), the log-likelihood of a single observation is given by</p><formula xml:id="formula_4">ℓ(𝛽; 𝑥 𝑖 ) = −1 2 {︂ 1 𝜎 2 (𝑦 𝑖 − 𝛽 ⊤ 𝑥 ˜𝑖) 2 + 𝑑 log 2𝜋𝜎 2 }︂ .<label>(4)</label></formula><p>The individual scores are obtained by taking the partial derivatives of the log-likelihood with respect to the parameters, that is, for any 𝑖 ∈ [𝑛], A straightforward computation yields</p><formula xml:id="formula_5">𝑠(𝛽; 𝑥 𝑖 ) = (︂ 𝜕ℓ(𝛽; 𝑥 𝑖 ) 𝜕𝛽 0 , 𝜕ℓ(𝛽; 𝑥 𝑖 ) 𝜕𝛽 1 , . . . , 𝜕ℓ(𝛽; 𝑥 𝑖 ) 𝜕𝛽 𝑑 )︂ ⊤ .<label>(5)</label></formula><formula xml:id="formula_6">𝑠(𝛽; 𝑥 𝑖 ) = (︁ 𝑦 𝑖 • 𝑥 ˜𝑖 − 𝑥 ˜𝑖𝑥 ˜⊤ 𝑖 𝛽 )︁ 𝜎 −2 .<label>(6)</label></formula><p>The maximum likelihood estimate 𝛽 ^is obtained by solving ∑︀ 𝑖 𝑠(𝛽; 𝑥 𝑖 ) = 0: in this case this coincides with the ordinary least-squares estimate, in a sense the best linear fit on the entire leaf.</p><p>However, what we want is to split this leaf if the deviations from the linear model computed on the whole leaf are too strong. Therefore, we compute the cumulative score process defined by</p><formula xml:id="formula_7">∀𝑡 ∈ [0, 1], 𝐵 (𝑗) (𝑡; 𝛽 ^) = 1 √ 𝑛 ⌊𝑛𝑡⌋ ∑︁ 𝑖=1 𝑠(𝛽 ^; 𝑥 𝑖 ) ,<label>(7)</label></formula><p>and take as a criterion the maximum of the 𝐿 1 norm of 𝐵 (𝑗) (𝑡), across all features. The computation of the best split is described in Algorithm 1 and we provide an example in Figure <ref type="figure" target="#fig_2">3</ref>. Note that Eq. ( <ref type="formula" target="#formula_7">7</ref>) corresponds to Eq. ( <ref type="formula">13</ref>) in Merkle and Zeileis <ref type="bibr" target="#b24">[25]</ref>, with the notable difference that Eq. ( <ref type="formula">13</ref>) is normalized (by the square root of the estimated covariance matrix of the scores). We observe empirically that normalizing yields worse results when detecting the splits. compute scores 𝑠(𝛽 ^; 𝑥 𝑖 ) according to Eq. (</p><p>compute 𝐵 (𝑗) according to Eq. (</p><p>store</p><formula xml:id="formula_10">𝑚 𝑗 = max 𝑡∈[0,1] ⃦ ⃦ 𝐵 (𝑗) (𝑡) ⃦ ⃦<label>1 10:</label></formula><p>store (𝑗, 𝑡 𝑗 ) = GetBestSplit(𝑥, 𝑦) </p><formula xml:id="formula_11">𝑡 𝑗 = arg max 𝑡∈[0,1] ⃦ ⃦ 𝐵 (𝑗) (𝑡) ⃦ ⃦</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>C. Evaluation</head><p>Evaluating interpretability methods is quite challenging: for a given machine learning model, there is rarely a ground-truth available, telling us which features were used for a given prediction. Zhou et al. <ref type="bibr" target="#b25">[26]</ref> outline two main evaluation methodologies: human-centered evaluations, and functionality-grounded (quantitative, note that the terminology was introduced by Doshi-Velez and Kim <ref type="bibr" target="#b26">[27]</ref>). We focus on the latter, since human-centered evaluations are hard to obtain and can be unreliable. Among those, we choose to compare GLEAMS to existing methodology using monotonicity and recall. Below we briefly describe these two metrics.</p><p>Monotonicity is defined as the correlation between the absolute value of the attributions at a given point and the expected loss of the model restricted to the corresponding feature. Intuitively, this takes high values if the model is very imprecise when it does not know the feature at hand. This is Metric 2.3 in Nguyen and Martínez <ref type="bibr" target="#b14">[15]</ref>. Following their notation, for each interpretability method to evaluate, for each point in the test set, we compute on the one hand a vector of attributions 𝛼 ∈ R 𝑑 and on the other hand a vector of expected loss 𝑒 ∈ R 𝑑 . Monotonicity is then defined as the Spearman rank correlation <ref type="bibr" target="#b27">[28]</ref> between |𝛼| and 𝑒. Formally, for all 𝑗 ∈ [𝑑], 𝑒 𝑗 is defined as</p><formula xml:id="formula_12">𝑒 𝑗 = ∫︁ 𝑏 𝑗 𝑎 𝑗 ℓ(𝑓 (𝜉), 𝑓 𝑗 (𝑥))𝑝(𝑥)d𝑥 ,<label>(8)</label></formula><p>where 𝑓 𝑗 is the restriction of 𝑓 to coordinate 𝑗 (keeping all other coordinates equal to those of 𝜉), 𝑝 is a probability distribution on [𝑎 𝑗 , 𝑏 𝑗 ] (recall that 𝑎 𝑗 and 𝑏 𝑗 denote the boundaries of 𝒳 along dimension 𝑗, see Assumption 1), and ℓ is the squared loss. Let us define 𝒰(𝐶) the uniform distribution on a compact set 𝐶. In our experiments, we consider two natural choices of 𝑝 in Eq. <ref type="bibr" target="#b7">(8)</ref>. First, 𝑝 ∼ 𝒰([ℓ 𝑗 , 𝑟 𝑗 ]), where ℓ 𝑗 (resp. 𝑟 𝑗 ) are the left (resp. right) boundaries of 𝑅 along feature 𝑗, where 𝑅 is the cell containing 𝜉. This corresponds to local monotonicity: how well the attributions reflect the variations in prediction locally, that is, from the point of view of GLEAMS, on 𝑅. Second, 𝑝 ∼ 𝒰([𝑎 𝑗 , 𝑏 𝑗 ]), which corresponds to global monotonicity: how well the attributions reflect the variations in prediction on the whole input space along feature 𝑗.</p><p>Let us conclude this section by defining formally the Recall of Important Features. For certain models, we can actually be certain that some feature attributions should be 0. This is the case, for instance, if we force our model 𝑓 to use only a subset of the features. As a measure of the quality of explanations, one can then compare the top features selected by an attribution method to the set of true features, which we know. Let us call 𝑇 the set of true features. As a first sanity check, we ran the partition scheme of GLEAMS on simple models and checked whether the surrogate model 𝑓 ^was indeed close to the true model 𝑓 . In order to have a ground-truth for the partition, we chose partition-based models, i.e., there exist a partition 𝒫 of 𝒳 such that the parameters of 𝑓 are constant on each rectangular cell 𝑅 ∈ 𝒫. We considered a simple setting where the model is piecewise-linear with rectangular cells, a situation in which GLEAMS could, in theory, recover perfectly both the underlying partition and the coefficients of the local models. We present here two examples: (i) a discontinuous piecewise-linear model with arbitrary coefficients, and (ii) a piecewise-linear model with continuity constraints. Example (i) is a generalization of the surface produced by a Random Forest regressor, while example (ii) is reminiscent of a fully-connected, feedforward, ReLU activated neural network. Note however that for the latter, the cells of the partition are more complicated than simple rectangle, see Montufar et al. <ref type="bibr" target="#b28">[29]</ref> (in particular Figure <ref type="figure" target="#fig_1">2</ref>). In each case, GLEAMS recovers perfectly the underlying model, as demonstrated in Figure <ref type="figure" target="#fig_1">2</ref>.</p><p>It is interesting to notice that, despite having a stopping criterion on the coefficient of determination, not all leaves satisfy this criterion since the other stopping criterion is a minimal number of points per leaf. Thus it is possible to end up in situations where the 𝑅 2 on a given leaf is sub par. Nevertheless, these leaves are by construction quite small. For instance, in example (i), the average 𝑅 2 is equal to 1, and in example (ii) it is equal to .98.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Overview of the global surrogate model construction. Left panel: the black-box model maps the input space (here 𝒳 = [0, 1] 2 ) to R, which we can visualize as a surface. Middle panel: we generate 𝑁 Sobol points on 𝒳 , giving rise to 𝑁 measurement points on the surface (in blue). Right panel: we fit a piecewise-linear global surrogate model 𝑓 ^on the measurement points by recursively splitting 𝒳 .</figDesc><graphic coords="2,305.29,325.76,164.51,100.85" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Gleams explanations consist of both: i) Local importance (feature attribution) and ii) counterfactuals, i.e., answering what-if questions. On the left feature importance for a regressor trained on the house sell dataset (Section 3), while figure on the right answers to "What if my house was bigger?" Moving along feature sqft_living, for a given example (yellow dot), we can visualize specific values of 𝑓 (Sobol points in blue) as well as the GLEAMS linear approximations (in green).</figDesc><graphic coords="5,264.20,90.74,214.09,76.82" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Evolution of the norm of the cumulative score process. Top panel: measurements of a piecewise-linear model (dark dots) visualized along one axis. Bottom panel: evolution of ‖𝐵(𝑡)‖ 1 as a function of 𝑡 (solid blue line). The process presents a clear maximum, which gives us a candidate split for this axis (vertical red line).</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: Approximating piecewise-linear models with GLEAMS surrogate model. Left panels: example (i), model on the left and GLEAMS surrogate on the right. Right panels: example (ii), model on the left and GLEAMS surrogate on the right.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>Experimental results (higher is better), 5th and 95th confidence intervals are to be found in brackets. GLEAMS is on par with state-of-the-art explanation methods in terms of reliability of explanations.</figDesc><table><row><cell>dataset</cell><cell>evaluation</cell><cell cols="2">LIME</cell><cell cols="2">SHAP</cell><cell cols="2">PDP</cell><cell cols="2">GLEAMS</cell></row><row><cell></cell><cell></cell><cell>XGB</cell><cell>MLP</cell><cell>XGB</cell><cell>MLP</cell><cell>XGB</cell><cell>MLP</cell><cell>XGB</cell><cell>MLP</cell></row><row><cell>wine</cell><cell>local</cell><cell>0.38</cell><cell>0.51</cell><cell>0.40</cell><cell>0.53</cell><cell>0.56</cell><cell>0.84</cell><cell>0.51</cell><cell>0.77</cell></row><row><cell></cell><cell></cell><cell>[-0.09, 0.77]</cell><cell>[0.19, 0.84]</cell><cell>[-0.08, 0.76]</cell><cell>[0.22, 0.78]</cell><cell>[0.23, 0.83]</cell><cell>[0.69, 0.95]</cell><cell>[0.15, 0.78]</cell><cell>[0.7, 0.86]</cell></row><row><cell>wine</cell><cell>global</cell><cell>0.24</cell><cell>0.74</cell><cell>0.27</cell><cell>0.78</cell><cell>0.70</cell><cell>0.89</cell><cell>0.36</cell><cell>0.85</cell></row><row><cell></cell><cell></cell><cell>[-0.27, 0.66]</cell><cell>[0.53, 0.9]</cell><cell>[-0.19, 0.65]</cell><cell>[0.53, 0.93]</cell><cell>[0.44, 0.9]</cell><cell>[0.78, 0.97]</cell><cell>[-0.19, 0.79]</cell><cell>[0.69, 0.95]</cell></row><row><cell>wine</cell><cell>recall (6 true features)</cell><cell>0.89</cell><cell>0.80</cell><cell>1.0</cell><cell>0.87</cell><cell>0.50</cell><cell>0.50</cell><cell>0.77</cell><cell>0.75</cell></row><row><cell></cell><cell></cell><cell>[0.67, 1.0]</cell><cell>[0.67, 1.0]</cell><cell>[1.0, 1.0]</cell><cell>[0.83, 1.0]</cell><cell>[0.5, 0.5]</cell><cell>[0.5, 0.5]</cell><cell>[0.5, 0.83]</cell><cell>[0.5, 1.0]</cell></row><row><cell>house</cell><cell>local</cell><cell>0.05</cell><cell>0.63</cell><cell>0.52</cell><cell>0.82</cell><cell>0.51</cell><cell>0.83</cell><cell>0.37</cell><cell>-0.19</cell></row><row><cell></cell><cell></cell><cell>[-0.25, 0.32]</cell><cell>[0.37, 0.81]</cell><cell>[0.22, 0.79]</cell><cell>[0.73, 0.89]</cell><cell>[0.73, 0.9]</cell><cell>[0.74, 0.9]</cell><cell>[0.26, 0.47]</cell><cell>[-0.3, -0.05]</cell></row><row><cell>house</cell><cell>global</cell><cell>0.23</cell><cell>0.63</cell><cell>0.37</cell><cell>0.81</cell><cell>0.38</cell><cell>0.82</cell><cell>0.47</cell><cell>0.24</cell></row><row><cell></cell><cell></cell><cell>[-0.09, 0.55]</cell><cell>[0.36, 0.1]</cell><cell>[0.07, 0.64]</cell><cell>[0.69, 0.87]</cell><cell>[0.26, 0.54]</cell><cell>[0.71, 0.9]</cell><cell>[0.07, 0.72]</cell><cell>[-0.02, 0.57]</cell></row><row><cell>house</cell><cell>recall (10 true features)</cell><cell>0.85</cell><cell>0.74</cell><cell>0.99</cell><cell>0.69</cell><cell>0.40</cell><cell>0.40</cell><cell>0.61</cell><cell>0.58</cell></row><row><cell></cell><cell></cell><cell>[0.7, 1.0]</cell><cell>[0.6, 0.8]</cell><cell>[0.9, 1.0]</cell><cell>[0.5, 0.8]</cell><cell>[0.4, 0.4]</cell><cell>[0.4,0.4]</cell><cell>[0.4, 0.7]</cell><cell>[0.4,0.7]</cell></row><row><cell>Parkinson</cell><cell>local</cell><cell>0.10</cell><cell>0.26</cell><cell>0.47</cell><cell>0.55</cell><cell>0.28</cell><cell>0.49</cell><cell>0.50</cell><cell>0.01</cell></row><row><cell></cell><cell></cell><cell>[-0.11, 0.41]</cell><cell>[-0.04, 0.59]</cell><cell>[0.3, 0.66]</cell><cell>[0.28, 0.76]</cell><cell>[0.16, 0.34]</cell><cell>[0.26, 0.76]</cell><cell>[0.36, 0.6]</cell><cell>[-0.2, 0.23]</cell></row><row><cell>Parkinson</cell><cell>global</cell><cell>0.31</cell><cell>0.46</cell><cell>0.41</cell><cell>0.71</cell><cell>0.52</cell><cell>0.66</cell><cell>0.30</cell><cell>0.36</cell></row><row><cell></cell><cell></cell><cell>[0.02, 0.59]</cell><cell>[0.16, 0.71]</cell><cell>[0.08, 0.71]</cell><cell>[0.47, 0.88]</cell><cell>[0.29, 0.71]</cell><cell>[0.43, 0.83]</cell><cell>[-0.06, 0.63]</cell><cell>[0.01, 0.68]</cell></row><row><cell>Parkinson</cell><cell>recall (10 true features)</cell><cell>0.86</cell><cell>0.77</cell><cell>0.97</cell><cell>0.80</cell><cell>0.40</cell><cell>0.40</cell><cell>0.50</cell><cell>0.60</cell></row><row><cell></cell><cell></cell><cell>[0.7, 1.0]</cell><cell>[0.5, 0.8]</cell><cell>[0.9, 1.0]</cell><cell>[0.4, 0.7]</cell><cell>[0.4, 0.4]</cell><cell>[0.4, 0.4]</cell><cell>[0.3, 0.7]</cell><cell>[0.43, 0.6]</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Algorithm 1</head><label>1</label><figDesc>GetBestSplit: Get the best split for a given leaf. Require: 𝑥 1 , . . . , 𝑥 𝑛 : Sobol points in current leaf; 𝑦 1 , . . . , 𝑦 𝑛 : corresponding 𝑓 values 1: for 𝑗 = 1 to 𝑑 do 𝑥 1 , . . . , . . . , 𝑥 𝑛 according to feature 𝑗</figDesc><table><row><cell cols="3">2: re-order 3: re-order 𝑦 1 , . . . , 𝑦 𝑛 accordingly</cell></row><row><cell>4:</cell><cell cols="2">compute 𝛽 ^= OLS(𝑦; 𝑥)</cell></row><row><cell>5:</cell><cell cols="2">compute predictions 𝑦 ^𝑖</cell></row><row><cell>6:</cell><cell>set 𝜎 2 = 1 𝑛</cell><cell>∑︀ 𝑛 𝑖=1 (𝑦 𝑖 − 𝑦 ^𝑖) 2</cell></row><row><cell>7:</cell><cell></cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>1 11 :</head><label>11</label><figDesc>return 𝑗 ⋆ = arg max 𝑗∈[𝑑] 𝑚 𝑗 , feature to split 12: return 𝑡 ⋆ 𝑗 , value of the split 13: return 𝛽 ^, local linear model Sobol points inside 𝑅; 𝑉 : values of 𝑓 on 𝑆 1: initialize 𝑇 ← 𝒳 , tree storing partition and models 2: for 𝐿 ∈ 𝑇.leaves do {𝑥 1 , . . . , 𝑥 𝑛 } = 𝑆 ∩ 𝐿</figDesc><table><row><cell cols="2">Algorithm 2 RecTree: Recursive construction of 𝑓 ^.</cell></row><row><cell>3:</cell><cell>if 𝑅 2 (𝐿) ≤ 𝜌 and 𝑛(𝐿) ≥ 2𝑛 min then</cell></row><row><cell cols="2">4: set 𝑥 = 5: set 𝑦 = {𝑦 1 , . . . , 𝑦 𝑛 } accordingly</cell></row><row><cell>6:</cell><cell></cell></row></table><note>Require: 𝑅: hyper-rectangle; 𝑆:</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>7 :</head><label>7</label><figDesc>𝐿 ℓ ← 𝐿 ∩ {𝑥 : 𝑥 𝑗 ≤ 𝑡 𝑗 } 𝐿 𝑟 ← 𝐿 ∩ {𝑥 : 𝑥 𝑗 &gt;𝑡 𝑗 } GlobalSurrogate: Compute the global surrogate model used by GLEAMS. Require: 𝑓 : black-box model; 𝒳 or its boundaries; 𝑁 : number of Sobol points; 𝜌: 𝑅 2 threshold; 𝑛 min : min. number of points per leaf 1: generate 𝑆 = {𝑥 1 , . . . , 𝑥 𝑁 } Sobol points in 𝒳 2: compute 𝑦 𝑖 = 𝑓 (𝑥 𝑖 ) 3: 𝑉 ← {𝑦 1 , . . . , 𝑦 𝑁 } 4: return 𝑇 ← RecTree(𝒳 , 𝑆, 𝑉 )</figDesc><table><row><cell cols="2">Algorithm 3</cell></row><row><cell>8:</cell><cell></cell></row><row><cell>9:</cell><cell>𝐿.left_child ← RecTree(𝐿 ℓ , 𝑆, 𝑉 )</cell></row><row><cell>10:</cell><cell>𝐿.right_child ← RecTree(𝐿 𝑟 , 𝑆, 𝑉 )</cell></row><row><cell></cell><cell>return 𝑇</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">https://github.com/giorgiovisani/GLEAMS</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">available at https://www.openml.org/search?type=data&amp;status=active&amp;id=42092</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Deep neural networks and tabular data: A survey</title>
		<author>
			<persName><forename type="first">V</forename><surname>Borisov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Leemann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Seßler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Haug</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Pawelczyk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Kasneci</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Neural Networks and Learning Systems</title>
		<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Why a Right to Explanation of Automated Decision-Making Does Not Exist in the General Data Protection Regulation</title>
		<author>
			<persName><forename type="first">S</forename><surname>Wachter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Mittelstadt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Floridi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Data Privacy Law</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="page" from="76" to="99" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)</title>
		<author>
			<persName><forename type="first">A</forename><surname>Adadi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Berrada</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Access</title>
		<imprint>
			<biblScope unit="volume">6</biblScope>
			<biblScope unit="page" from="52138" to="52160" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Explainable AI: A review of machine learning interpretability methods</title>
		<author>
			<persName><forename type="first">P</forename><surname>Linardatos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Papastefanopoulos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kotsiantis</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Entropy</title>
		<imprint>
			<biblScope unit="volume">23</biblScope>
			<biblScope unit="page">18</biblScope>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Why Should I Trust You?: Explaining the Predictions of Any Classifier</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">T</forename><surname>Ribeiro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Singh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Guestrin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</title>
				<meeting>the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="1135" to="1144" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Points which uniformly fill a multidimensional cube</title>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">M</forename><surname>Sobol</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Mathematics, cybernetics series</title>
		<imprint>
			<biblScope unit="page">32</biblScope>
			<date type="published" when="1985">1985</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Breiman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">H</forename><surname>Friedman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">A</forename><surname>Olshen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">J</forename><surname>Stone</surname></persName>
		</author>
		<title level="m">Classification and regression trees</title>
				<imprint>
			<publisher>Wadsworth &amp; Brooks / Cole Advanced Books &amp; Software</publisher>
			<date type="published" when="1984">1984</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">LOTUS: An algorithm for building accurate and comprehensible logistic regression trees</title>
		<author>
			<persName><forename type="first">K.-Y</forename><surname>Chan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W.-Y</forename><surname>Loh</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Computational and Graphical Statistics</title>
		<imprint>
			<biblScope unit="volume">13</biblScope>
			<biblScope unit="page" from="826" to="852" />
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Model-based recursive partitioning</title>
		<author>
			<persName><forename type="first">A</forename><surname>Zeileis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Hothorn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Hornik</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Computational and Graphical Statistics</title>
		<imprint>
			<biblScope unit="volume">17</biblScope>
			<biblScope unit="page" from="492" to="514" />
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Modeling wine preferences by data mining from physicochemical properties</title>
		<author>
			<persName><forename type="first">P</forename><surname>Cortez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Cerdeira</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Almeida</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Matos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Reis</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Decision support systems</title>
		<imprint>
			<biblScope unit="volume">47</biblScope>
			<biblScope unit="page" from="547" to="553" />
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Accurate telemonitoring of parkinson&apos;s disease progression by non-invasive speech tests</title>
		<author>
			<persName><forename type="first">A</forename><surname>Tsanas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Little</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Mcsharry</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Ramig</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Nature Precedings</title>
		<imprint>
			<biblScope unit="page" from="1" to="1" />
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">XGBoost: A Scalable Tree Boosting System</title>
		<author>
			<persName><forename type="first">T</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Guestrin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</title>
				<meeting>the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="785" to="794" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Greedy function approximation: a gradient boosting machine</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">H</forename><surname>Friedman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">The Annals of Statistics</title>
		<imprint>
			<biblScope unit="page" from="1189" to="1232" />
			<date type="published" when="2001">2001</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">P</forename><surname>Kingma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Ba</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1412.6980</idno>
		<title level="m">Adam: A method for stochastic optimization</title>
				<imprint>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<author>
			<persName><forename type="first">A.-P</forename><surname>Nguyen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">R</forename><surname>Martínez</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2007.07584</idno>
		<title level="m">On quantitative aspects of model interpretability</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Extracting tree-structured representations of trained networks</title>
		<author>
			<persName><forename type="first">M</forename><surname>Craven</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">W</forename><surname>Shavlik</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in Neural Information Processing Systems</title>
		<imprint>
			<biblScope unit="page" from="24" to="30" />
			<date type="published" when="1996">1996</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<title level="m" type="main">Programs for Machine Learning</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">R</forename><surname>Quinlan</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1993">1993</date>
			<publisher>Elsevier</publisher>
			<biblScope unit="volume">4</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">The CAD-MDD: A computerized adaptive diagnostic screening tool for depression</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">D</forename><surname>Gibbons</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Hooker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">D</forename><surname>Finkelman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">J</forename><surname>Weiss</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">A</forename><surname>Pilkonis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Frank</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Moore</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">J</forename><surname>Kupfer</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">The Journal of clinical psychiatry</title>
		<imprint>
			<biblScope unit="volume">74</biblScope>
			<biblScope unit="page">669</biblScope>
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Hooker</surname></persName>
		</author>
		<title level="m">Interpreting Models via Single Tree Approximation</title>
				<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Distribution-free predictive inference for regression</title>
		<author>
			<persName><forename type="first">J</forename><surname>Lei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Sell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Rinaldo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">J</forename><surname>Tibshirani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Wasserman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of the American Statistical Association</title>
		<imprint>
			<biblScope unit="volume">113</biblScope>
			<biblScope unit="page" from="1094" to="1111" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">A unified approach to interpreting model predictions</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Lundberg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S.-I</forename><surname>Lee</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in neural information processing systems</title>
		<imprint>
			<biblScope unit="volume">30</biblScope>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Anchors: High-precision model-agnostic explanations</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">T</forename><surname>Ribeiro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Singh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Guestrin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Thirty-Second AAAI Conference on Artificial Intelligence</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Glocalx-from local to global explanations of black box AI models</title>
		<author>
			<persName><forename type="first">M</forename><surname>Setzu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Guidotti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Monreale</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Turini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Pedreschi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Giannotti</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Artificial Intelligence</title>
		<imprint>
			<biblScope unit="volume">294</biblScope>
			<biblScope unit="page">103457</biblScope>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Interpretable and differentially private predictions</title>
		<author>
			<persName><forename type="first">F</forename><surname>Harder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Bauer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Park</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the AAAI Conference on Artificial Intelligence</title>
				<meeting>the AAAI Conference on Artificial Intelligence</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="volume">34</biblScope>
			<biblScope unit="page" from="4083" to="4090" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Tests of Measurement Invariance without Subgroups: A Generalization of Classical Methods</title>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">C</forename><surname>Merkle</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Zeileis</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Psychometrika</title>
		<imprint>
			<biblScope unit="volume">78</biblScope>
			<biblScope unit="page" from="59" to="82" />
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Evaluating the quality of machine learning explanations: A survey on methods and metrics</title>
		<author>
			<persName><forename type="first">J</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">H</forename><surname>Gandomi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Holzinger</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Electronics</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="page">593</biblScope>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<monogr>
		<title level="m" type="main">Towards a rigorous science of interpretable machine learning</title>
		<author>
			<persName><forename type="first">F</forename><surname>Doshi-Velez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Kim</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1702.08608</idno>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">The proof and measurement of association between two things</title>
		<author>
			<persName><forename type="first">C</forename><surname>Spearman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">The American journal of psychology</title>
		<imprint>
			<biblScope unit="volume">100</biblScope>
			<biblScope unit="page" from="441" to="471" />
			<date type="published" when="1904">1904</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">On the number of linear regions of deep neural networks</title>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">F</forename><surname>Montufar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Pascanu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Cho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Bengio</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in neural information processing systems</title>
		<imprint>
			<biblScope unit="volume">27</biblScope>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
