<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>eXplainable Random Forest</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Guy Amit</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Shlomit Gur</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>IBM Research</institution>
          ,
          <addr-line>Haifa</addr-line>
          ,
          <country country="IL">Israel</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Advanced machine learning models have become widely adopted in various domains due to their exceptional performance. However, their complexity often renders them dificult to interpret, which can be a significant limitation in high-stakes decision-making scenarios where explainability is crucial. In this study we propose eXplainable Random Forest (XRF), an extension of the Random Forest model that takes into consideration, crucially, during training, explainability constraints stemming from the users' view of the problem and its feature space. While numerous methods have been suggested for explaining machine learning models, these methods often are suitable for use only after the model has been trained. Furthermore, the explanations provided by these methods may include features that are not human-understandable, which in turn may hinder the user's comprehension of the model's reasoning. Our proposed method addresses these two limitations. We apply our proposed method to six public benchmark datasets in a systematic way and demonstrate that XRF models manage to balance the trade-of between the models' performance and the users' explainability constraints.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Explainability</kwd>
        <kwd>Machine Learning</kwd>
        <kwd>Random Forest</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        In recent years Artificial Intelligence (AI) has been widely adopted in many domains and aspects
of lives [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], including healthcare [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], finance [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], games [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], and other complex cognitive tasks.
This wide adoption can be attributed to improved performance of AI models, which has been
made possible by increasingly complex, and therefore also increasingly dificult to explain,
models [
        <xref ref-type="bibr" rid="ref5 ref6">5, 6</xref>
        ]. However, the goal of these AI models is often to aid or make decisions for humans,
thus giving cause for concern, especially in domains with potentially severe individual or
societal consequences. For example, the use of Machine Learning (ML) algorithms in healthcare
to predict a medical outcome or a patient’s diagnosis [
        <xref ref-type="bibr" rid="ref7 ref8 ref9">7, 8, 9</xref>
        ] may raise concerns regarding their
reliability, trustworthiness, and ethics [
        <xref ref-type="bibr" rid="ref10 ref11">10, 11</xref>
        ], especially in terms of the patients’ characteristics
these models rely on.
      </p>
      <p>
        These concerns have spurred public debates, which in turn gave rise to legislation (e.g., [
        <xref ref-type="bibr" rid="ref12 ref13 ref14">12,
13, 14</xref>
        ]) and high-volume research in the domain of eXplainable Artificial Intelligence (XAI) [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ].
While many XAI approaches have been proposed, they are predominantly post-hoc. That is,
they only deem the ML model explainable or not, after it is set (i.e., done training). If the ML
model is deemed not explainable, there is often no known course of action to rectify it. One
could try, for example, other ML models, modify hyper-parameters, or modify the training data,
but none of these options is guided by any form of explainability. As such, these solutions could
result in no more, if not less, explainable ML models. A few exceptions to this rule include
data-type-specific (often unstructured data, such as images or text) methods for Deep Neural
Networks (DNNs) [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] and Teaching Explanations for Decisions (TED) [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ].
      </p>
      <p>In this work we propose eXplainable Random Forest (XRF), a method for incorporating a
form of user-defined explainability into the training stage of Random Forest (RF) models. The
user-defined explainability in the current context is a global Feature Preference (FP) vector,
which is provided as input to the model by the user. The preference is dependent on the user and
can be based on, for example, human-understandability, political correctness, or actionability
of the features. The proposed method can be viewed as a user-driven "soft" feature selection
where, unlike traditional feature selection methods, all the features remain available for the ML
model to use.</p>
      <p>The objective of the proposed method is to foster a sound trade-of between the performance
of the ML model and its explainability, as defined by the adherence of the model’s Feature
Importance (FI) values to the user-defined FP vector. Thus, we employed a systematic method
that challenges RF’s FI values. Our testing method encouraged the ML model to use  features
with the lowest FI values and discouraged the use of  features with the greatest FI values. The
rest of the features were treated with neutral preference. The generality of this method allowed
us to test its applicability using multiple benchmark datasets in a way that is neither
datasetnor domain-specific. It is our belief that other studies could benefit from this type of systematic
testing in the future.</p>
      <p>Our results demonstrate that our method manages to balance the trade-of between the
performance and explainability of the ML model. For example, if a feature does not hold great
predictive power, a strong user-defined preference will not inflate its importance in the model at
any cost to the model’s performance. Alternatively, a high-importance feature with a negative
user-defined preference can be substituted by another feature or a set of features, even if the
model’s performance is negatively impacted.</p>
      <p>The main contributions of the current work are as follows:
1. We introduce XRF, an extension of RF that balances performance and user-defined
explainability during training.
2. We propose an "eXplainability Score", a metric to quantify explainability and demonstrate
on six benchmark datasets XRF’s ability to balance the trade-of between the model’s
performance and its explainability, as measured by this metric.
3. We present a novel testing scheme to empirically evaluate the trade-of between
performance and explainability in the absence of a human agent.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Background</title>
      <sec id="sec-2-1">
        <title>2.1. eXplainable Artificial Intelligence</title>
        <p>
          XAI refers to the ability to understand and interpret the decision-making process involving an AI
model. It serves multiple purposes [
          <xref ref-type="bibr" rid="ref15 ref18">18, 15</xref>
          ], one of which is increasing trust and accountability in
AI-based (thus also ML-based) systems. This is especially imperative in high-stakes applications
with potentially severe individual or societal consequences (e.g., judicial, health-related, or
ifnancial decisions).
        </p>
        <p>
          Some ML models are interpretable [
          <xref ref-type="bibr" rid="ref15 ref18">18, 15</xref>
          ], that is, their inner workings are inherently
human-understandable (e.g., Decision Trees (DTs)). These models are often very simple and
do not perform well on real-world complex tasks. Conversely, ML models that perform well
on real-world tasks are usually complex and not interpretable. XAI methods can be used to
make non-interpretable models explainable. Explainable ML models are models whose inner
workings are not human-understandable, but can be explained using proxies (e.g., FI of surrogate
interpretable models [19, 20]).
        </p>
        <p>XAI methods for ML models can be categorized into local, global, and hybrid methods. Local
explanation methods, such as Local Interpretable Model-Agnostic Explanations (LIME) [19]
or SHapley Additive exPlanations (SHAP) [20], provide an explanation for a specific model
prediction. These methods usually provide an explanation in the form of a list. This list can
indicate, for example, the respective contribution of each of the features to the particular
prediction or alternative values that would result in a diferent prediction (i.e., counterfactual
explanation). In contrast, global XAI methods explain a ML model over the entire input space,
as captured by the training data. Hybrid methods explain a ML model over sub-samples of the
input space. In this paper we focus on adjusting the global explanation provided by XRF models
to accommodate the user’s FP.</p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Genetic Algorithms</title>
        <p>Genetic Algorithms (GAs) [21, 22] are a family of gradient-free optimization algorithms inspired
by the mechanics of natural selection and genetics. They are known for solving a variety of
optimization and search problems [23, 24, 25, 26]. GAs encode the solution space of a problem
as a set of individuals (i.e., a population) and then repeatedly apply genetic operators such
as selection, crossover (recombination), and mutation in order to make the individuals (i.e.,
solutions) evolve with the end goal of finding a suficient solution to the problem.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Related Work</title>
      <p>Our work is another step in the ongoing quest to make AI more transparent and
humanunderstandable. Some of the earliest XAI techniques include LIME [19] and SHAP [20]. These
techniques are model-agnostic and propose feature-attribution-based explanations to a ML
model using a surrogate ML model. While LIME uses a linear surrogate model, SHAP makes use
of Shapley values [27], borrowed from the field of cooperative game theory. More recent papers
introduced Bayesian aspects to these techniques [28, 29]. For example, the Local
InterpretationDriven Abstract Bayesian Network (LINDA-BN) method proposed using a Probabilistic Graphical
Model to provide a local explanation. That said, these methods are applied post-hoc and,
therefore, do not incorporate consideration of explainability in the training phase.</p>
      <p>
        Several approaches have been proposed to improve a ML model’s explainability during its
training. One approach is to use the TED [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] framework, which predicts both a label and an
explanation from a predefined set of explanations. This is done using a multi-label classifier,
trained to predict the original label coupled with the explanation. TED is model-agnostic and
can be applied to every model capable of performing multi-label classification, but it sufers
from the following limitations: (1) the training set must include both labels and explanations,
(2) no new explanations outside the predefined set can be "learned" from the data, and (3) there
is nothing to prevent a predicted explanation from contradicting features in the explained data
point.
      </p>
      <p>An example of a model-specific approach [ 30] proposed a framework that trains an AdaBoost
ensemble while enforcing the fairness of the model. In this approach the samples’ weights, used
for training each weak learner, are chosen while considering SHAP values. In each iteration, a
surrogate model is fitted in addition to a weak learner and the SHAP values from the surrogate
model are used together with the weak learner’s error rate to update the training samples’
weights. Another model-specific approach [ 31] comes from the domain of data privacy. In this
ifeld of study some of the dataset features are considered more sensitive and are encouraged to be
used less, in order to protect against Membership Inference attacks [32]. This approach extends
the DT’s training process: in each feature split, a predefined weight is considered in addition to
the entropy score. This training process yields a DT whose FI values are aligned with the privacy
requirements that were defined by the user. Although this approach is meant for enforcing
privacy requirements, it can also be used to improve a DT’s explainability. However, the user
has no way of balancing model performance against privacy requirements or explainability.</p>
      <p>
        Finally, it is worth noting that there is a large body of work that focuses on DNNs [
        <xref ref-type="bibr" rid="ref16">16, 33, 34</xref>
        ].
In this work, however, we focus on DT ensemble models.
      </p>
    </sec>
    <sec id="sec-4">
      <title>4. Explainable Random Forest</title>
      <p>In this part of the paper we present our proposed model, the XRF. XRF is an adaptation of the
RF model, which allows the user to influence the model’s dependency on specific features in the
training data. More accurately, given a FP vector,   ∈ R , the model is adjusted to use all the
 features in the training data, while attempting to assign weights to features in accordance
with   .</p>
      <p>Training an XRF model consists of two steps: (1) constructing the DTs (the FP vector does
not afect this step) and (2) optimizing the Objective Function (OF) (equation 2) with respect to
the FP vector. In the following subsections we will describe each of these steps.</p>
      <sec id="sec-4-1">
        <title>4.1. Constructing the Decision Trees</title>
        <p>As previously mentioned, XRF is an extension of RF. The RF model employs a bagging method,
which means that each DT in the RF is constructed independently. During inference, all the
DTs are used in a voting scheme to get a final prediction from the ensemble. In RF, each DT is
given an equal weight and the prediction for a data point  is computed as follows:
 1 
^ = {∑︁  · (|)} = {  · ∑︁ (|)}
∈ =1 ∈ =1
(1)
Where  is the set of possible classes, ^ is the predicted class,  is the number of DTs in the RF,
 is the ℎ DT, and  is its weight. Notice that RF assumes that ∑︀
=1  = 1 and gives each
DT an equal vote. Therefore, in plain RF,  = 1 for all 1 ≤  ≤  ∈ N.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Optimization Process</title>
        <p>In this step the user-defined FP vector is used to influence the model’s use of the features.
During the optimization process the weights of the DTs,  for 1 ≤  ≤  ∈ N, are modified
such that the FI values of the XRF adhere to the specifications in the FP vector as much as
possible.</p>
        <p>We optimize an OF that balances between the model’s performance and a similarity measure
between the model’s FI values and the input FP vector. We refer to this similarity measure as
"eXplainability Score" (XS) and denote it by . Formally, we define the OF of the XRF as
follows:
 (,   ) =  ( ) +  · ( ,   )
(2)
Where  is a hyper-parameter controlling the similarity metric’s weight during the optimization
process and   is a vector of the FI values of the XRF model.</p>
        <p>It is worth noting that the representation of the FP vector should match the choice of XS. For
example: if XS cannot handle negative values or 0, then the FP vector should hold only positive
values; and if XS expects probability distributions, then the values in the FP vector should sum
up to 1.</p>
        <p>
          Taking into consideration the FI values makes the OF a parametric non-diferentiable function
and, therefore, our method must use a gradient-free optimization algorithm, such as the GA used
in our implementation [
          <xref ref-type="bibr" rid="ref19">35</xref>
          ]. The GA searches for the weights of the DTs,  = [1, . . . , ], to
maximize the value of the OF over a validation set. Formally, this can be written as:
max  (,   )

(3)
To ensure that ∑︀
        </p>
        <p>=1  = 1, which is a requirement in plain RF, a normalization strategy is
used in our proposed method. This normalization is applied each time the GA calculates the OF,
as well as when the final weights for the DTs are returned by the optimization process.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Experimental Setting</title>
      <p>
        Our evaluation consisted of compering the performance and the explainability of XRFs for
diferent  values, using plain RFs as baselines. We performed the evaluation on six publicly
available benchmark datasets of varying sizes (see Table 1) and domains: Yeast 1 [
        <xref ref-type="bibr" rid="ref20">36</xref>
        ], Adult
Income 2 [
        <xref ref-type="bibr" rid="ref21">37</xref>
        ], German Credit 3 [
        <xref ref-type="bibr" rid="ref22">38</xref>
        ], Nursery 4 [
        <xref ref-type="bibr" rid="ref23">39</xref>
        ], Iris 5 [
        <xref ref-type="bibr" rid="ref24">40</xref>
        ], and Breast Cancer 6 [
        <xref ref-type="bibr" rid="ref25">41</xref>
        ]. Due
1https://archive.ics.uci.edu/ml/datasets/yeast
2http://archive.ics.uci.edu/ml/datasets/Adult
3http://archive.ics.uci.edu/ml/datasets/statlog+(german+credit
+data)
4https://archive.ics.uci.edu/ml/datasets/nursery
5https://archive.ics.uci.edu/ml/datasets/iris
6https://archive.ics.uci.edu/dataset/17/breast+cancer+wisconsin
+diagnostic
#Samples
1,299
32,561
1,000
12,960
150
569
to the scarcity of some of the labels in the German Credit and Nursery datasets, we used only
their top four 7 and top three 8 labels, respectively. These datasets are popular, also in related
work [
        <xref ref-type="bibr" rid="ref26 ref27 ref28">42, 31, 43, 44</xref>
        ]. Additionally, the sizes of their feature spaces allow us to study, as well as
comprehensively visualize, the method’s performance (e.g., the change in feature attribution as
a function of performance-vs-explainability trade-of).
      </p>
      <p>The XRF relies on the user to specify a FP vector. However, datasets with user-defined FP
vectors, as well as expected explanations, are not readily available. Moreover, human preferences
and feedback are subjective and could vary greatly, possibly adding complexity that would
detract from the methodological focus of the current work. Therefore, we chose to not involve
humans in the evaluation. In the absence of human involvement in the evaluation process, we
developed an automated testing procedure to challenge the XRF’s ability to afect the FI values.
Inverting the FI values of a trained model can be a complicated task, since often the signal
supporting the model prediction is embedded within the features with the greatest FI values.
Thus, to test the efect of XRF’s OF, we employed the following testing scheme:
1. We trained a RF model on a dataset 
2. We extracted the FI values of ’s features for that RF
3. Finally, we set the FP vector such that the  top-importance and  bottom-importance
7CYT, NUC, MIT, and ME3
8not_recom, priority, and spec_prior
features are penalized and rewarded, respectively. The rest of the features were assigned
neutral preference (default). We used  = 1 if || ≤ 5 and  = 2 if || &gt; 5, where ||
is the number of features in .</p>
      <p>For XS we used the cosine similarity measure:
 ( ,   ) =</p>
      <p>·  
‖ ‖·‖   ‖
(4)
(5)
As the performance of XRF is bounded between 0 and 1 (for both accuracy and 1 score), it
is more convenient if XS is bounded too (as is the case with cosine similarity: − 1 &lt;  &lt; 1)
for ease of choice of the hyper-parameter  (see equation 2). As previously mentioned, the
representation of the FP vector should match the choice of XS. For simplicity, we considered FP
values,    ∈ R, such that (1)    &gt; 0 indicates that the ℎ feature should be preferred by
the model, (2)    &lt; 0 indicates that the model should avoid it as much as possible, and (3)
   = 0 indicates that the model may use it as needed (default; neutral preference).</p>
      <p>
        For example: let  be a dataset with four features (|| = 4). Then for a RF model on 
with FI values of [0.15, 0.3, 0.2, 0.35], the first and last features are the least ( 0.15) and most
(0.35) important features in the RF, respectively. Thus, for the computation of XS in the XRF’s
optimization process, the FP vector would be set to: [
        <xref ref-type="bibr" rid="ref1">1, 0, 0, − 1</xref>
        ].
      </p>
      <p>For the optimization step, we employed in our implementation a simple version of the GA
from the DEAP framework 9. We used CxOnePoint crossover operator, a tournament selection,
and an additive Gaussian noise as a mutation operator. We chose softmax to normalize the
weights:
We performed in addition an ablation study, examining the efect of a diferent normalization
strategy and a diferent XS.</p>
      <p>In all of our experiments we performed a grid search over the hyper-parameters of the plain
RF and the GA (searched per dataset 10). To reduce complexity, we split the grid search into two
steps: (1) RF-related features (i.e., number of trees ∈ [10, 20, ..., 120], max depth ∈ [1, 2, ..., 7],
and max samples ∈ [0.2, 0.4, ..., 1]) were first searched using a plain RF and then (2) fixing
these values, the GA-related features (i.e., number of generations ∈ [10, 20, ..., 120], mating
probability ∈ [0.1, 0.3, 0.5], and mutation probability ∈ [0.1, 0.4, 0.7]) were searched using an
XRF with  = 0. This split also ensures that XRF results could be fairly compared to RF results.
Table 2 summarizes the results of these grid searches. We then used these hyper-parameters to
evaluate three XRF models for three diferent values of  , the hyper-parameter controlling the
XS’s weight in the XRF’s OF. To measure the prediction performance of the model, we used
the 1 score and accuracy metrics, while to measure the adherence to the FP vector, we used
the XS. Notice that performance-wise, the model was trained in respect to its accuracy. We ran
these experiments ten times per dataset, using a diferent random seed in each run.
9https://deap.readthedocs.io/en/master/
10random seed = 42</p>
      <p>XS per Alpha
0.9
0.9
)
%
(y0.8
rcccauA0.7
0.6</p>
      <p>Dataset</p>
      <p>AdultIncome
BreastCancer
GermanCredit
Iris
Nursery
Yeast</p>
    </sec>
    <sec id="sec-6">
      <title>6. Evaluation</title>
      <p>In this section we present the results of our evaluations. The XRF optimization process balances
between the task performance (e.g., accuracy or 1 score) and the XS based on the choice
of  . Therefore, we started out by testing how  afects the XRF model’s performance and
explainability. To this end we computed three metrics for the resulting XRF model: (1) accuracy,
(2) 1 score, and (3) XS. We performed this evaluation on six public benchmark datasets (see
Table 1) for three diferent  values (0.5, 1.0, and 2.0; see Figure 1).</p>
      <p>The classification performance metrics (i.e., accuracy and 1) were compared to a baseline
model: a plain RF classifier that was trained using comparable configuration (number of trees,
max depth, and max samples) as the respective XRF models (see RF in Figure 1). Our results
demonstrate that for most of the datasets the classification performance metrics decreased
gradually, compered to the respective baselines models, as  increased. Intriguingly, for
highdimensional datasets, such as the Breast Cancer and German Credit datasets, this decrease
was less pronounced. This can be attributed to the optimization process employed in the XRF
algorithm, which is responsible for increasing the similarity between the FI values and FP
vector. As the number of features in the dataset increases, the space of possible solutions widens,
allowing for the discovery of solutions that strike a better balance between explainability and
performance (i.e., reduce the performance less). Conversely, as desired, for most of the datasets
XS gradually increased as  increased. For some datasets (e.g., Nursery) we observed an increase
in the performance metrics of the XRF model with  = 0.5, as compared to the baseline RF
model. This might be attributed to the optimization step in XRF, which searches for weights to
be assigned to the DTs in the ensemble such that the OF (equation 2) is maximal, as for  &lt; 1
the OF assigns greater importance to the performance term than to the XS term.</p>
      <p>To further evaluate the efects of the OF on the FI values, we plotted, per dataset, the FI value
of each feature as a function of  (Figure 2). Solid blue and magenta lines in the plots indicate
features that were set (in the FP vectors) to be rewarded and penalized, respectively. In some
datasets (e.g., Iris and Nursery) we saw a clear exchange in the FI values between penalized
and rewarded features, while in others (e.g., Adult Income) we saw a clear exchange between
penalized and neutral features. Finally, we observed in some datasets (e.g., Adult Income and
Yeast) that the FI values of rewarded features remained fairly unchanged, suggesting that they
were not found suficiently useful by the model. Overall, these results lend support to our
claim that the XRF’s optimization process indeed shifts the FI values of rewarded and penalized
Config. Metric</p>
      <p>Acc.</p>
      <p>Abs. + CS 1</p>
      <p>Acc.</p>
      <p>Abs. + CE 1</p>
      <p>Acc.</p>
      <p>SM + CS 1</p>
      <p>Acc.</p>
      <p>SM + CE 1

to this configuration and the user can choose to use other similarity metrics for XS and other
normalization strategies for the DTs’ weights. In this part of the paper we consider XRF models
that use alternative configurations.</p>
      <p>
        During these experiments we considered defining the input FP vector as a probability
distribution (  ∈ [
        <xref ref-type="bibr" rid="ref1">0, 1</xref>
        ] and ∑︀
      </p>
      <p>=1   = 1). Under this definition, the model’s usage of features
with high or low probabilities in the FP vector will be encouraged or discouraged, respectively,
by the optimization process. Cross-Entropy (CE) is a well-defined distance metric between two
probability distributions:

(, ) = − ∑︁  · log() (6)
=1
Where  and  are two probability distributions over the same  variables in corresponding
order. Thus, we used for XS a CE -based similarity measure. Given the FP vector and the
FI values, which satisfy the conditions for probability distributions, we define the following
similarity measure:
Where  for 1 ≤  ≤  ∈ N is the weight of the ℎ DT in the XRF model.</p>
      <p>Finally, in these experiments we considered only the modified Yeast dataset (with four labels
after removing relatively scarce labels) and performed a grid search over the hyper-parameters
for each of the combinations of normalization strategy and XS. As these choices (of normalization
strategy and XS) do not afect the performance of RF, there was no need to repeat the grid search
over RF-related hyper-parameters (i.e., number of trees, max depth, and max samples). Thus,
we performed grid searches over the hyper-parameters that are related to the optimization
process (the GA’s hyper-parameters: number of generations, mating probability, and mutation
probability) using XRFs with  = 0. The results of these grid searches were identical to the
results using the cosine similarity measure and the softmax normalization strategy (see top row
in Table 2). Then, similarly to our main experimental setting, we evaluated the performance of
three XRF models with  ∈ [0.5, 1.0, 2.0] based on ten runs, using a diferent random seed in
each run. Overall, our results (see Table 3) suggest that the configuration we chose to use in the
main experimental setting (cosine similarity for XS and the softmax normalization strategy) is
either comparable to or better than the alternatives considered here, according to our metrics.
10
20
30
40</p>
      <p>Another important hyper-parameter of the XRF model, with the potential to considerably
afect the model’s performance, is the number of DTs in the ensemble. In the main experimental
setting we selected the number of DTs based on a grid search on a plain RF. That is, the number
of DTs was determined based on performance only and without optimization. As we aim to
create a robust method, we examined how the number of DTs in the XRF ensemble afects the
XS. To this end we trained multiple XRF models on the modified Yeast dataset (same variation
as above), each with a diferent number of DTs. We set the rest of the hyper-parameters, as well
as the XS and normalization strategy, in accordance with the main experimental setting (see
top row in Table 2). We repeated this experiment with 11  values for every number of DTs
(see Figure 3). Our results did not reveal a clear trend in XS when plotted against the number of
DTs, thereby suggesting that the XS is relatively unafected by the number of DTs in the XRF.
Thus, we conclude that the number of DTs in the XRF is a hyper-parameter that can be left to
the user to set, either manually or by using an automatic grid search.</p>
      <p>It is worth noting that these results could be due to a limitation in our experimental setting.
As we chose to use low-dimensional datasets (30 features at most), the number of significantly
unique DTs for them is capped. In turn, it is possible that in higher-dimensional datasets the
number of DTs will afect the XS.</p>
    </sec>
    <sec id="sec-7">
      <title>7. Conclusion</title>
      <p>
        In this paper we introduce XRF, a novel family of DT ensemble models that optimize a balance
between a performance metric (e.g., accuracy or 1 score) and a user-defined form of
explainability during training. In this context, the explainability is defined as adherence of the model’s
feature attribution (i.e., FI values) to a user-defined FP vector. Through this vector the user may
define which features they prefer the model to use, which features they prefer the model to
avoid, and which features they are neutral about. Additionally, our method allows the user to
determine the desired balance between the model’s explainability and its predictive performance.
That is, the user can determine to what degree the model should prioritize explainability over
performance (or performance over explainability). In contrast to our proposed method, existing
XAI methodologies for structured data are primarily either post-hoc [19, 20] or limited to a
predefined set [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], thereby limiting the insight one could gain from the ML model.
      </p>
      <p>The results of our experiments demonstrate that the FI values in the XRF model are afected
as expected by the FP vector, and in a controlled manner. For example, the FI of a rewarded
feature is increased only if either (1) the feature provides information that could be used for
prediction by the model or (2) explainability is highly prioritized over performance (1 &lt;&lt;  ).
Similarly, the FI of a penalized feature is decreased only if either (1) the predictive signal it
provides to the model can be replaced by another feature or features, or (2) explainability is
highly prioritized over performance (1 &lt;&lt;  ). Finally, our results also suggest that the XS of
the XRF model is relatively unafected by the number of DTs in the model. Although this should
be reexamined in higher-dimensional settings.</p>
      <p>We hope that this work serves as a building block towards the development of more ML
models that balance the trade-of between performance and explainability during training in
an informed manner. As such, we recognize some calculated limitations of the current study,
including the low-dimensionality of the datasets and the time complexity introduced by the
GA algorithm. These will need to be addressed in future work. Additional future research
directions may include (1) extending models that are more complex than RF, (2) developing other
forms of human-defined or metric-derived explainability, and (3) developing other explainability
quantification metrics.</p>
    </sec>
    <sec id="sec-8">
      <title>Acknowledgments</title>
      <p>This project has received funding from the European Union’s Horizon research and innovation
programme under grant agreements no 101094905 (AI4GOV) and 101092639 (FAME).
stakeholders, and future research opportunities, Information Systems Management 39
(2022) 53–63.
[19] M. T. Ribeiro, S. Singh, C. Guestrin, " why should i trust you?" explaining the predictions
of any classifier, in: Proceedings of the 22nd ACM SIGKDD international conference on
knowledge discovery and data mining, 2016, pp. 1135–1144.
[20] S. M. Lundberg, S.-I. Lee, A unified approach to interpreting model predictions, Advances
in neural information processing systems 30 (2017).
[21] M. Mitchell, An introduction to genetic algorithms, MIT press, 1998.
[22] K. De Jong, Learning with genetic algorithms: An overview, Machine learning 3 (1988)
121–138.
[23] R. Leardi, Genetic algorithms in chemistry, Journal of Chromatography A 1158 (2007)
226–233.
[24] A. E. Drake, R. E. Marks, Genetic algorithms in economics and finance: Forecasting
stock market prices and foreign exchange—a review, Genetic algorithms and genetic
programming in computational finance (2002) 29–54.
[25] S. H. Mousavi-Avval, S. Rafiee, M. Sharifi, S. Hosseinpour, B. Notarnicola, G. Tassielli, P. A.</p>
      <p>Renzulli, Application of multi-objective genetic algorithms for optimization of energy,
economics and environmental life cycle assessment in oilseed production, Journal of
Cleaner Production 140 (2017) 804–815.
[26] H. A. Saleh, R. Chelouah, The design of the global navigation satellite system surveying
networks using genetic algorithms, Engineering Applications of Artificial Intelligence 17
(2004) 111–122.
[27] L. S. Shapley, Notes on the n-person game—ii: The value of an n-person game.(1951),</p>
      <p>Lloyd S Shapley 7 (1951).
[28] C. Moreira, Y.-L. Chou, M. Velmurugan, C. Ouyang, R. Sindhgatta, P. Bruza, Linda-bn:
An interpretable probabilistic approach for demystifying black-box predictive models,
Decision Support Systems 150 (2021) 113561.
[29] X. Zhao, W. Huang, X. Huang, V. Robu, D. Flynn, Baylime: Bayesian local interpretable
model-agnostic explanations, in: Uncertainty in Artificial Intelligence, PMLR, 2021, pp.
887–896.
[30] J. M. Hickey, P. G. Di Stefano, V. Vasileiou, Fairness by explicability and adversarial
shap learning, in: Machine Learning and Knowledge Discovery in Databases: European
Conference, ECML PKDD 2020, Ghent, Belgium, September 14–18, 2020, Proceedings, Part
III, Springer, 2021, pp. 174–190.
[31] A. Goldsteen, G. Ezov, A. Farkash, Reducing risk of model inversion using privacy-guided
training, arXiv preprint arXiv:2006.15877 (2020).
[32] R. Shokri, M. Stronati, C. Song, V. Shmatikov, Membership inference attacks against
machine learning models, in: 2017 IEEE symposium on security and privacy (SP), IEEE,
2017, pp. 3–18.
[33] P. Angelov, E. Soares, Towards explainable deep neural networks (xdnn), Neural Networks
130 (2020) 185–194.
[34] X. Bai, X. Wang, X. Liu, Q. Liu, J. Song, N. Sebe, B. Kim, Explainable deep learning
for eficient and robust pattern recognition: A survey of recent developments, Pattern
Recognition 120 (2021) 108102.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>S.</given-names>
            <surname>Laato</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Tiainen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Najmul</given-names>
            <surname>Islam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mäntymäki</surname>
          </string-name>
          ,
          <article-title>How to explain ai systems to end users: a systematic literature review and research agenda</article-title>
          ,
          <source>Internet Research</source>
          <volume>32</volume>
          (
          <year>2022</year>
          )
          <fpage>1</fpage>
          -
          <lpage>31</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>K.</given-names>
            <surname>Shailaja</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Seetharamulu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Jabbar</surname>
          </string-name>
          ,
          <article-title>Machine learning in healthcare: A review, in: 2018 Second international conference on electronics, communication and aerospace technology (ICECA)</article-title>
          , IEEE,
          <year>2018</year>
          , pp.
          <fpage>910</fpage>
          -
          <lpage>914</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>A.</given-names>
            <surname>Mashrur</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Luo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. A.</given-names>
            <surname>Zaidi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Robles-Kelly</surname>
          </string-name>
          ,
          <article-title>Machine learning for financial risk management: a survey</article-title>
          ,
          <source>IEEE Access 8</source>
          (
          <year>2020</year>
          )
          <fpage>203203</fpage>
          -
          <lpage>203223</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>S.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Okun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A</given-names>
            .
            <surname>Jackson</surname>
          </string-name>
          ,
          <article-title>Learning to play go from scratch</article-title>
          ,
          <source>Nature</source>
          <volume>550</volume>
          (
          <year>2017</year>
          )
          <fpage>336</fpage>
          -
          <lpage>337</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>F. K.</given-names>
            <surname>Došilović</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Brčić</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Hlupić</surname>
          </string-name>
          ,
          <article-title>Explainable artificial intelligence: A survey, in: 2018 41st International convention on information and communication technology, electronics and microelectronics (MIPRO)</article-title>
          , IEEE,
          <year>2018</year>
          , pp.
          <fpage>0210</fpage>
          -
          <lpage>0215</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>W. J. von Eschenbach</surname>
          </string-name>
          ,
          <article-title>Transparency and the black box problem: Why we do not trust ai</article-title>
          ,
          <source>Philosophy &amp; Technology</source>
          <volume>34</volume>
          (
          <year>2021</year>
          )
          <fpage>1607</fpage>
          -
          <lpage>1622</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>O.</given-names>
            <surname>Dogan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Tiwari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Jabbar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Guggari</surname>
          </string-name>
          ,
          <article-title>A systematic review on ai/ml approaches against covid-19 outbreak</article-title>
          ,
          <source>Complex &amp; Intelligent Systems</source>
          <volume>7</volume>
          (
          <year>2021</year>
          )
          <fpage>2655</fpage>
          -
          <lpage>2678</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>A.</given-names>
            <surname>Wong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Otles</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. P.</given-names>
            <surname>Donnelly</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Krumm</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>McCullough</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>DeTroyer-Cooley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Pestrue</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Phillips</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Konye</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Penoza</surname>
          </string-name>
          , et al.,
          <article-title>External validation of a widely implemented proprietary sepsis prediction model in hospitalized patients</article-title>
          ,
          <source>JAMA Internal Medicine</source>
          <volume>181</volume>
          (
          <year>2021</year>
          )
          <fpage>1065</fpage>
          -
          <lpage>1070</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>K.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , X. Liu,
          <string-name>
            <given-names>J.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Sang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Liang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Wang</surname>
          </string-name>
          , et al.,
          <article-title>Clinically applicable ai system for accurate diagnosis, quantitative measurements, and prognosis of covid-19 pneumonia using computed tomography</article-title>
          ,
          <source>Cell</source>
          <volume>181</volume>
          (
          <year>2020</year>
          )
          <fpage>1423</fpage>
          -
          <lpage>1433</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>I. Y.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Pierson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Rose</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Joshi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Ferryman</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. Ghassemi,</surname>
          </string-name>
          <article-title>Ethical machine learning in healthcare, Annual review of biomedical data science 4 (</article-title>
          <year>2021</year>
          )
          <fpage>123</fpage>
          -
          <lpage>144</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>S. T.</given-names>
            <surname>Jan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Ishakian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Muthusamy</surname>
          </string-name>
          ,
          <article-title>Ai trust in business processes: The need for processaware explanations</article-title>
          ,
          <source>Proceedings of the AAAI Conference on Artificial Intelligence</source>
          <volume>34</volume>
          (
          <year>2020</year>
          )
          <fpage>13403</fpage>
          -
          <lpage>13404</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>GDPR</surname>
          </string-name>
          ,
          <article-title>General data protection regulation</article-title>
          ,
          <year>2018</year>
          . https://gdpr-info.eu/.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>CCPA</surname>
          </string-name>
          ,
          <article-title>The california consumer privacy act</article-title>
          ,
          <year>2018</year>
          . https://leginfo.legislature.ca.gov/faces/ billTextClient.xhtml?bill_id=201720180AB375.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>E.</given-names>
            <surname>Commission</surname>
          </string-name>
          , et al.,
          <source>White paper on artificial intelligence</source>
          .
          <article-title>a european approach to excellence and trust</article-title>
          ,
          <source>COM</source>
          (
          <year>2020</year>
          )
          <volume>65</volume>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>A.</given-names>
            <surname>Adadi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Berrada</surname>
          </string-name>
          ,
          <article-title>Peeking inside the black-box: a survey on explainable artificial intelligence (xai)</article-title>
          ,
          <source>IEEE access 6</source>
          (
          <year>2018</year>
          )
          <fpage>52138</fpage>
          -
          <lpage>52160</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>G.</given-names>
            <surname>Erion</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. D.</given-names>
            <surname>Janizek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Sturmfels</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. M.</given-names>
            <surname>Lundberg</surname>
          </string-name>
          ,
          <string-name>
            <surname>S.-I. Lee,</surname>
          </string-name>
          <article-title>Improving performance of deep learning models with axiomatic attribution priors and expected gradients</article-title>
          ,
          <source>Nature machine intelligence</source>
          <volume>3</volume>
          (
          <year>2021</year>
          )
          <fpage>620</fpage>
          -
          <lpage>631</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>M.</given-names>
            <surname>Hind</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Wei</surname>
          </string-name>
          , M. Campbell,
          <string-name>
            <given-names>N. C.</given-names>
            <surname>Codella</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Dhurandhar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mojsilović</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Natesan Ramamurthy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. R.</given-names>
            <surname>Varshney</surname>
          </string-name>
          ,
          <article-title>Ted: Teaching ai to explain its decisions</article-title>
          ,
          <source>in: Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>123</fpage>
          -
          <lpage>129</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>C.</given-names>
            <surname>Meske</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Bunde</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Schneider</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Gersch</surname>
          </string-name>
          , Explainable artificial intelligence: objectives,
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [35]
          <string-name>
            <given-names>F.</given-names>
            <surname>Busetti</surname>
          </string-name>
          ,
          <article-title>Genetic algorithms overview</article-title>
          ,
          <source>Retrieved on December</source>
          <volume>1</volume>
          (
          <year>2007</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [36]
          <string-name>
            <given-names>K.</given-names>
            <surname>Nakai</surname>
          </string-name>
          , Yeast,
          <source>UCI Machine Learning Repository</source>
          ,
          <year>1996</year>
          . DOI: https://doi.org/10.24432/C5KG68.
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [37]
          <string-name>
            <given-names>B.</given-names>
            <surname>Becker</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Kohavi</surname>
          </string-name>
          , Adult,
          <source>UCI Machine Learning Repository</source>
          ,
          <year>1996</year>
          . DOI: https://doi.org/10.24432/C5XW20.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [38]
          <string-name>
            <given-names>H.</given-names>
            <surname>Hofmann</surname>
          </string-name>
          ,
          <article-title>Statlog (German Credit Data)</article-title>
          ,
          <source>UCI Machine Learning Repository</source>
          ,
          <year>1994</year>
          . DOI: https://doi.org/10.24432/C5NC77.
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [39]
          <string-name>
            <given-names>V.</given-names>
            <surname>Rajkovic</surname>
          </string-name>
          , Nursery,
          <source>UCI Machine Learning Repository</source>
          ,
          <year>1997</year>
          . DOI: https://doi.org/10.24432/C5P88W.
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [40]
          <string-name>
            <surname>R. A</surname>
          </string-name>
          . Fisher, Iris,
          <source>UCI Machine Learning Repository</source>
          ,
          <year>1988</year>
          . DOI: https://doi.org/10.24432/C56C76.
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [41]
          <string-name>
            <given-names>W.</given-names>
            <surname>Wolberg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Mangasarian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Street</surname>
          </string-name>
          , W. Street,
          <article-title>Breast Cancer Wisconsin (Diagnostic)</article-title>
          ,
          <source>UCI Machine Learning Repository</source>
          ,
          <year>1995</year>
          . DOI: https://doi.org/10.24432/C5DW2B.
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [42]
          <string-name>
            <given-names>C.</given-names>
            <surname>Agarwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Krishna</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Saxena</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Pawelczyk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Johnson</surname>
          </string-name>
          , I. Puri,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zitnik</surname>
          </string-name>
          , H. Lakkaraju,
          <article-title>OpenXAI: Towards a transparent evaluation of model explanations</article-title>
          ,
          <source>in: Thirty-sixth Conference on Neural Information Processing Systems Datasets and Benchmarks Track</source>
          ,
          <year>2022</year>
          . URL: https://openreview.net/forum?id=
          <fpage>MU2495w47rz</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [43]
          <string-name>
            <given-names>M. T.</given-names>
            <surname>Keane</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Smyth</surname>
          </string-name>
          ,
          <article-title>Good counterfactuals and where to find them: A case-based technique for generating counterfactuals for explainable ai (xai)</article-title>
          ,
          <source>in: Case-Based Reasoning Research and Development: 28th International Conference, ICCBR</source>
          <year>2020</year>
          , Salamanca, Spain, June 8-12,
          <year>2020</year>
          , Proceedings 28, Springer,
          <year>2020</year>
          , pp.
          <fpage>163</fpage>
          -
          <lpage>178</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [44]
          <string-name>
            <given-names>O.</given-names>
            <surname>Sagi</surname>
          </string-name>
          , L. Rokach,
          <article-title>Explainable decision forest: Transforming a decision forest into an interpretable tree</article-title>
          ,
          <source>Information Fusion</source>
          <volume>61</volume>
          (
          <year>2020</year>
          )
          <fpage>124</fpage>
          -
          <lpage>138</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>