<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Decision Trees ⋆</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>ParisaJamshidi</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>SławomirNowaczyk</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mahmoud Rahat</string-name>
          <email>mahmoud.rahat@hh.se</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>ZahraTaghiyarrenan</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Center for Applied Intelligent Systems Research</institution>
          ,
          <addr-line>CAISR</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Halmstad University</institution>
          ,
          <addr-line>Halmstad</addr-line>
          ,
          <country country="SE">Sweden</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>eXplainable AI (XAI)</institution>
          ,
          <addr-line>Federated Learning, Incremental Decision Tree</addr-line>
          ,
          <country>Extremely Fast Decision Tree</country>
        </aff>
      </contrib-group>
      <fpage>0000</fpage>
      <lpage>0002</lpage>
      <abstract>
        <p>Explainable Artificial Intelligence (XAI) is crucial in ensuring transparency, accountability, and trust in machine learning models, especially in applications involving high-stakes decision-making. This paper focuses on addressing the research gap in federated learning (FL), specifically emphasizing the use of inherently interpretable underlying models. While most FL frameworks rely on complex, black-box models such as Artificial Neural Networks (ANNs), we propose using Decision Tree (DT) classifiers to maintain explainability. More specifically, we introduce a novel framework for horizontal federated learning using Extremely Fast Decision Trees (EFDTs) with streaming data on the client side. Our approach involves aggregating clients' EFDTs on the server side without centralizing raw data, and the training process occurs on the clients' sides. We outline three aggregation strategies and demonstrate that our methods outperform local models and achieve performance levels close to centralized models while retaining inherent explainability.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>CEUR
ceur-ws.org</p>
    </sec>
    <sec id="sec-2">
      <title>1. Introduction</title>
      <p>
        Explainable Artificial Intelligence (XAI) is essential in machine learning and artificial
intelligence as it improves transparency, accountability, trust, and the ability to enhance and debug
AI systems. As AI technologies expand to various sectors, from healthcare to finance, the
complexity and opacity of these models often prevent their deployment in critical decision-making
processes. XAI seeks to mitigate this issue by providing insights into how AI algorithms reason
and how they arrive at specific conclusions, enabling users to validate and comprehend model
behaviors. Most of XAI research today focuses on post-hoc explanations for black-box models.
These models have complex internal workings that are dificult for humans to interpret or
understand. Various forms of analysis, from surrogate models to gradient credit assignment, are often
used to explain these models and their decisions. However, these post-hoc explanations have
several drawbacks. For example, they often fail to capture the true inner workings of the model,
Lithunia
providing only surface-level insights that might not be trustw1o,r2]t.hFyu[rthermore, they
are vulnerable to adversarial attacks that can manipulate the explanations without changing
the model’s predictions3,[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. Such limitations make them a questionable choice in applications
where transparency is a priority. Conversely, white-box models are inherently explainable
models designed to be transparent and interpretable from the outset. These models allow users
to understand their decision-making processes without the need for additional interpretability
techniques 1[].
      </p>
      <p>Federated learning (FL) is a technique in machine learning in which multiple devices record
local data and share their model with a server. The server forms a global model and shares it
without exchanging raw data. This method is more eficient regarding bandwidth and computing
resources because it does not need to transfer large amounts of raw data to a central server.
Although this approach has gained much attention due to its capacity, most of the underlying
models for FL are Artificial Neural Networks (ANNs). Post-hoc explainability methods can
be applied to black-box models of FL. For example5,][used feature importance methods to
add explainability to their models, whi6l]ep[resented counterfactual explanation techniques.
However, there is a surprising lack of FL techniques suitable for inherently interpretable models.</p>
      <p>
        In this paper, we focus on Decision Tree (DT) classifiers. Shallow DTs, in contrast to NNs and
complex models, are designed in a way that their internal workings are easily understandable
to humans. This allows for straightforward tracing of how inputs are transformed into outputs,
facilitating immediate insight into the decision-making process. Such transparency is crucial for
validating the model’s logic, ensuring ethical compliance, and fostering trust among users. This
makes DTs particularly suitable for the identification and correction of biases and errors within
the model: their transparent nature enables users to pinpoint specific aspects of the model
that contribute to undesirable outcomes, facilitating targeted improvements. Moreover, some
research indicates that tree-based models outperform ANNs in some applications, specifically
tabular data7,[
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
      </p>
      <p>The typical approach in ML is batch learning, also known as ofline learning, which involves
training a model on a fixed dataset all at once. On the other hand, online learning, or incremental
learning, continuously updates the model as new data becomes available. This makes it suitable
for dynamic environments where data arrives sequentially. Much less research has been done
with incremental learning in FL fashion than batch learning.</p>
      <p>Therefore, there is a gap when clients have access to streaming data and aim to achieve an
inherently explainable global model with better performance than their local model through
collaboration in a federated mode. We proposed our method to aggregate the incremental
decision trees from clients’ sides in an FL framework, which can achieve higher performance
compared to local ones, and as the client’s models are DTs, they are inherently explainable. Our
contributions are as follows:
• We created a global model without centralizing raw data, using statistical information
stored in EFDTs.
• In our proposed method, EFDTs are trained locally and aggregated in each round.
• We have introduced three aggregation methods on the server to aggregate clients’ EFDTs.
• Furthermore, we maintain both local and global models as a single tree to make them
inherently explainable.</p>
      <p>The rest of the paper is organized as followResl:ated worksummarizes previous research
intersection of FL and tree-based methods. ThenPr,ienliminarie,swe introduce some basic
concepts of incremental decision trees and tree similarity methods. The proposed method is
described inMethodology. The experimental setup is found in SectiEoxnperimental Setup
followed by the result iRnesults. Finally, we draw some conclusions iCnonclusion</p>
    </sec>
    <sec id="sec-3">
      <title>2. Related work</title>
      <p>Within this distributed learning paradigm, horizontal federated learning (HFL) and vertical
federated learning (VFL) represent two approaches. Each client has diferent data samples in
HFL, but the feature set is the same. VFL, however, uses data samples shared among clients
with diferent features9[].</p>
      <p>
        SecureBoost1[0] is a VFL framework in which only one of the clients has the label of samples,
called active party, and the rest, without labels, called passive parties. This method works with
a gradient-tree boosting algorithm in which the active party shares the gradient and Hessian
values with the passive party. After that, the passive party categorizes the samples based on
local features into buckets and provides the total values of each bucket to the active party in
the form of a histogram. The active party then performs local calculations to find the best way
to split the nodes and directly coordinate updates with the passive parties. Other methods, like
OpBoost [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], were proposed to optimize SecureBoost. There are other methods categorized on
tree-based VFL, including FEDXGB1[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] and VF2Boost1[
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], which are Homomor Encryption
(HE)-based.
      </p>
      <p>
        There is also research on the conjugation of tree-based models and1H4]FLp.r[oposed a
method based on GBDT(Gradient-boosted decision trees), in which each client trains a decision
tree using their local dataset and adds it to the global model in turn. Boosting-based Federated
Random Forest (BOFRF)1[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] proposed a boosting framework for random forests in a federated
manner. Each client forms their RFs and, on the server side, considers those as weak classifiers
and proposes a method to calculate their weights. Our approach involves explicitly training a
single tree, which is explainable, whereas using the bagging approach in the server to form the
global model will result in a loss of explainability.
      </p>
      <p>Another approach is to train a tree-based model collaboratively in a server. Clients share
local statistical information with the server, and the server decides on node partitioning in each
round.</p>
      <p>There are few works of conjugation of incremental decision trees an1d6]FaLn.[d [17] are
two methods to train an IDT in VFL and HFL, respectively. The first method encrypts samples
from clients’ streaming data and passes them to the server. In the latter method, the Data
Collector collected the raw data from the client and aggregated them. In both, the server uses
that information to train a VFDT.</p>
      <p>Our proposed method falls under the category of HFL, and the features are consistent across
all clients. Additionally, we do not share raw data with the server. Instead, we train trees on the
client side, and the server is responsible for aggregating the trees.</p>
    </sec>
    <sec id="sec-4">
      <title>3. Preliminaries</title>
      <sec id="sec-4-1">
        <title>3.1. Incremental Decision tree</title>
        <p>The Hoefding Tree (HT) [ 18] is one of the fundamental research studies that proposed the
Incremental Decision Tree (IDT) for eficient data stream mining. The HT uses the Hoefding
bound to determine when a statistically significant decision about splitting a node can be made
based on the amount of data collected. The inequality for a random varwiiatbhlerange is
expressed as follows:</p>
        <p>2 ln (1/ )
 = , (1)</p>
        <p>√ 2
where  represents the margin of err oirs, the confidence level, and  is the number of
observations.</p>
        <p>In 2018, [19] proposed the Hoefding Anytime Tree, also known as the Extremely Fast Decision
Tree (EFDT). EFDT, in contrast, enhances the capabilities of HT by including an ”anytime”
aspect. This means it can continuously update and revise its model as new data becomes
available without revisiting old data. While HT makes decisions based on accumulated data up
to a certain point, EFDT continuously improves its decisions, resulting in better performance in
dynamic data streams.</p>
      </sec>
      <sec id="sec-4-2">
        <title>3.2. Decision Tree Similarity</title>
        <p>An important step in Federated Learning is an aggregation of models from diferent clients,
and a crucial component in that process is based on the similarity between them. Decision
tree similarity methods aim to quantify the similarity between diferent decision trees. Over
the years, several approaches have been proposed, typically divided into two main categories:
syntactic and semantic similarities.</p>
        <p>Syntactic similarity methods compare and analyze decision trees based on their structure.
Techniques such as tree edit distance, which calculates the minimal changes needed to transform
one tree into another, and structural similarity measures, which compare the arrangement and
features of nodes, are commonly use2d0[, 21, 22].</p>
        <p>Semantic similarity methods aim to capture the functional similarity between trees by
considering the decisions made by each tree, which means they look at the similarity in the
prediction distribution2s3,[22]. This can consider the similarity of decisions made by each
tree on a set of instances.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>4. Methodology</title>
      <p>Our methodology consists of two phases: training and aggregation. Training occurs on the
clients’ sides, while aggregation occurs on the server’s side. Our training process is incremental
since we are in the streaming setting. Our main contribution resides in the aggregation phase,
in which we put diferent aggregation strategies into contrast.</p>
      <p>Clinets’s sides
(training)</p>
      <p>Server’s side
(aggregation)
C1
C2
Ck
...</p>
      <sec id="sec-5-1">
        <title>4.1. Problem Formulation</title>
        <p>Assume we have clients 1 ⋯   where each one has access to a diferent streaming dataset
 1 ⋯   . Given that we are in a horizontal federated setting, the feature set is the same for all
datasets1 ⋯   .</p>
        <p>In particular, in order to maintain the inherent interpretability of the global model, we require
that the final result be limited to a single decision tree. While aggregating a number of trees
into a forest is a natural way to extend DT learning to a federated setting, our work is one of
the very few attempts to collaboratively create a single tree among all the clients.</p>
        <p>In our proposed method, the server only aggregates the clients’ trees without training them,
so there’s no need to share raw data with the server.</p>
      </sec>
      <sec id="sec-5-2">
        <title>4.2. Incremental training on the client side</title>
        <p>Each client ( ) trains its own decision tree on its local data)s.eAts( a new data point arrives,
they use incremental learning to update their tree. After a predefined number of data points are
received by each client, they send their trained decision trees to the server. Then, each one will
receive a new tree from the server and continue using and training this new tree by the next
part of their streaming data.</p>
        <p>Each node of EFDT stores a table, which can be interpreted as the number of instances
where the − ℎ attribute has th−e ℎ value, and the class label is. This table is used to
calculate metrics such as information gain or Gini index, which determine the best attribute for
splitting the data at a node. Each client will share their tree with this information instead of
raw data.</p>
      </sec>
      <sec id="sec-5-3">
        <title>4.3. Aggregation on the server side</title>
        <p>The aggregation process begins when the server receives decision trees from the clients. We
have proposed three diferent strategies for the server side. However, the aggregation process
remains consistent across all strategies, but each strategy difers in its approach to selecting
the base tree. Before delving into the specifics of each strategy, let’s first explore how the
aggregation process operates.</p>
        <p>As mentioned earlier, every node in EFDT contain s antable. As part of the aggregation
process, we merge the tables of nodes along the same path into two trees. Alg1ooruithlimnes
this method in detail. Specifically, when two nodes from two trees share common parents, they
can be efectively combined, even if they have opted for diferent features. This is possible
because the table contains aggregated information before the split.</p>
        <p>Algorithm 1 Aggregation Function</p>
        <p>As mentioned earlier, the strategies difer in terms of how the base tree is chosen to combine
the clients’ models. Below are the details of three proposed strategies.</p>
        <sec id="sec-5-3-1">
          <title>4.3.1. Syntactic strategy (Syn):</title>
          <p>In this approach, the base tree is the most representative syntactic tree among the clients’ trees,
meaning it is the tree that requires the fewest changes to match the other trees. By using the
edit distance on the bracket format of the trees, we can identify this tree, which we refer to as
”Syn” in Algorithm2. Then, the aggregation method is applied to ”Syn” along with the other
clients’ trees. Finally, the server broadcasts the Syn tree to all clients.</p>
        </sec>
        <sec id="sec-5-3-2">
          <title>4.3.2. Semantic strategy (Sem):</title>
          <p>The base tree in this strategy is that tree with minimum disagreement with other clients on
predictions. In Algorith3m, the server uses a proxy data set,,to identify this tree, referred
to as ”Sem”. Then, it iterates through all client trees, aggregating information from matching
paths in each tree into the Sem tree to update it. Finally, the server broadcasts the Sem tree to
all clients.</p>
          <p>Algorithm 2 Syntactic strategy (Syn)
1: Input: List of clients’ tre s =</p>
          <p>e
[ 1,  2, … ,   ]</p>
        </sec>
        <sec id="sec-5-3-3">
          <title>4.3.3. Individual strategy (Ind):</title>
          <p>In contrast to previous strategies, this strategy constructs one tree per client. It compares each
client’s tree with the trees of all other clients and aggregates information from matching paths
into the final tree for that client (Algori4t)h. mAfter updating all trees, the server will pass the
updated trees to their owners.</p>
          <p>Algorithm 4 Individual strategy (Ind)</p>
          <p>We used centralized learning and local learning to establish upper-bound and lower-bound,
respectively, for the mentioned strategies. In centralized learning, all the data exist in a single
location, and a single model is trained. In contrast, in local learning, each client trains a model
with its own data without interacting with other clients.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>5. Experimental Setup</title>
      <sec id="sec-6-1">
        <title>5.1. Dataset Description</title>
        <p>We use two datasets with categorical features, which lets us better control the experiment and
analyze the results. Below is a brief description of each.</p>
        <p>Mushroom dataset: This dataset contains 8124 descriptions of hypothetical samples
representing 23 species of gilled mushrooms in the Agaricus and Lepiota Family. Each species is
classified as edible or poisonous. There are 22 features and 8124 data po1.ints</p>
        <p>Chess (King-Rook vs. King-Pawn): This dataset contains two classes: white-can-win
(”won”) and white-cannot-win (”nowin”). The classes are mostly balanced at 52% and 48%,
respectively. There are 36 features and 3196 data p2o.ints</p>
      </sec>
      <sec id="sec-6-2">
        <title>5.2. Experiments Description</title>
        <p>To share the data across clients, we used stratified splitting so that the ratio of the classes was
the same for all clients. Moreover, the amount of data each client receives is the same as other
clients. To address the uncertainty, we randomly performed this data-sharing schema 10 times
with diferent seeds, and the result was averaged.</p>
        <p>It’s important to note that before dividing the data among clients, 1% of the data is set aside
to be stored on the server side. This data, referred toinasthis paper, is used to identify the
representative semantic tree.</p>
        <p>Our model choice is Extremely Fast Decision Tree (EFD1T9)][, which is the same for all
clients. For stable results, we resorted to the River library in P2y4t].hTohne[parameters for
EFDT are set to be the default parameters; only the binary_split is set to True. To construct
the representative syntactic tree used in SynRoot and SynPath, the APTED library is utilized
[20, 21].</p>
        <p>For the federated learning setting, we consider 4 clients and 200 rounds of communication
between the clients and the server. Accordingly, each client’s data is divided by 200 and fed
in a streaming manner to clients’ EFDT. This learning procedure is repeated for each strategy
mentioned in4.</p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>6. Results</title>
      <p>In this section, we present the results obtained using the proposed method. We compare the
result of the proposed method with diferent strategies against centralized and local learning.
We present two tables, Tabl1esand 2, showing the accuracy of five strategies after some
communication rounds for the mushroom and chess datasets, respectively. In almost all cases,
the performance of the proposed strategies falls between the performance of local and centralized
learning, as expected. These strategies benefit from aggregation, allowing them to access more
information at each point compared to local learning. However, their access to information is
still not as extensive as centralized learning.
1https://archive.ics.uci.edu/dataset/73/mushroom
2https://archive.ics.uci.edu/dataset/22/chess+king+rook+vs+king+pawn</p>
      <p>Then, we look deeper at the results on the mushroom dataset shown in 2F.igWuereobserve
that the speed of performance improvement in the three methods presented is initially higher
than that of local learning (refer to the dashed box in the lower left part of2F).igHuorweever,
this diference decreases after several stages, eventually reaching less than 5% in the last rounds
(see the dashed box in the upper right part of Fig2)u.rTehis suggests that a client with initially
limited data cannot efectively compete with aggregated methods. With access to more data,
the client may better understand the problem and achieve strong performance independently
close to aggregated methods.</p>
      <p>In all the experiments with the chess dataset, there are significant diferences between
the performance of the proposed methods and local learning performance. If you compare
the dashed boxes in Figur3e, you’ll notice that individual clients cannot achieve the same
performance as aggregated methods even after 200 rounds. However, similar to the results of
the mushroom dataset, the performance of the aggregated methods improves quickly in the
initial rounds in this dataset as well.
Type
Centralized
Local
Ind
Syn
Sem
Type
Centralized
Local
Ind
Syn
Sem
0
25
50
75
100
#round
125
150
175
200</p>
    </sec>
    <sec id="sec-8">
      <title>7. Conclusion</title>
      <p>This paper introduces a framework for horizontal federated learning using an Extremely Fast
Decision Tree (EFDT) as the underlying model and streaming data on the client side. We propose
three strategies to aggregate this type of underlying method. With our strategies, you can merge
the information on multiple EFDTs at each round, which helps to enhance the performance
at the client level. We also insist on having only one tree on the client side because a single
(shallow) tree is inherently explainable. We avoid using black-box models due to concerns about
using post-hoc explanation methods.</p>
      <p>We compared these three strategies with local and centralized learning. All three strategies
produce similar results, and their performance is higher than local learning and close (yet low)
to centralized learning. Experiments show that although, in some cases, the performance of
the proposed methods might not difer too much from local learning after some rounds, they
always grow very fast in the first rounds.</p>
    </sec>
    <sec id="sec-9">
      <title>Acknowledgments</title>
      <p>The work was carried out with support from The Knowledge Foundation and from Vinnova
(Sweden’s innovation agency) through the Vehicle Strategic Research and Innovation Programme,
FFI.
federated tree boosting framework based on order-preserving desensitization, arXiv
preprint arXiv:2210.01318 (2022).
[12] Y. Liu, Z. Ma, X. Liu, S. Ma, S. Nepal, R. H. Deng, K. Ren, Boosting privately: Federated
extreme gradient boosting for mobile crowdsensing, in: 2020 IEEE 40th international
conference on distributed computing systems (ICDCS), IEEE, 2020, pp. 1–11.
[13] F. Fu, Y. Shao, L. Yu, J. Jiang, H. Xue, Y. Tao, B. Cui, Vf2boost: Very fast vertical federated
gradient boosting for cross-enterprise learning, in: Proceedings of the 2021 International
Conference on Management of Data, 2021, pp. 563–576.
[14] L. Zhao, L. Ni, S. Hu, Y. Chen, P. Zhou, F. Xiao, L. Wu, Inprivate digging: Enabling
tree-based distributed data mining with diferential privacy, in: IEEE INFOCOM 2018-IEEE
Conference on Computer Communications, IEEE, 2018, pp. 2087–2095.
[15] M. Gencturk, A. A. Sinaci, N. K. Cicekli, Bofrf: A novel boosting-based federated random
forest algorithm on horizontally partitioned data, IEEE Access 10 (2022) 89835–89851.
[16] Z. Han, C. Ge, B. Wu, Z. Liu, Lightweight privacy-preserving federated incremental
decision trees, IEEE Transactions on Services Computing (2022).
[17] S. Sharma, K. Arora, P. S. Thakur, Horizontal federating decision tree learning from data
streams: Building intelligence in iot edge networks, in: 2022 IEEE 8th World Forum on
Internet of Things (WF-IoT), IEEE, 2022, pp. 1–6.
[18] P. Domingos, G. Hulten, Mining high-speed data streams, in: Proceedings of the sixth
ACM SIGKDD international conference on Knowledge discovery and data mining, 2000,
pp. 71–80.
[19] C. Manapragada, G. I. Webb, M. Salehi, Extremely fast decision tree, in: Proceedings of
the 24th ACM SIGKDD International Conference on Knowledge Discovery &amp; Data Mining,
2018, pp. 1953–1962.
[20] M. Pawlik, N. Augsten, Eficient computation of the tree edit distance, ACM Transactions
on Database Systems (TODS) 40 (2015) 1–40.
[21] M. Pawlik, N. Augsten, Tree edit distance: Robust and memory-eficient, Information</p>
      <p>Systems 56 (2016) 157–173.
[22] A. I. Weinberg, M. Last, Selecting a representative decision tree from an ensemble of
decision-tree models for fast big data classification, Journal of Big Data 6 (2019) 1–17.
[23] R. Miglio, G. Sofritti, The comparison between classification trees through proximity
measures, Computational statistics &amp; data analysis 45 (2004) 577–593.
[24] J. Montiel, M. Halford, S. M. Mastelini, G. Bolmier, R. Sourty, R. Vaysse, A. Zouitine, H. M.</p>
      <p>Gomes, J. Read, T. Abdessalem, et al., River: machine learning for streaming data in python
(2021).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>C.</given-names>
            <surname>Rudin</surname>
          </string-name>
          ,
          <article-title>Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead</article-title>
          ,
          <source>Nature machine intelligence</source>
          <volume>1</volume>
          (
          <year>2019</year>
          )
          <fpage>206</fpage>
          -
          <lpage>215</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>P.-J.</given-names>
            <surname>Kindermans</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hooker</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Adebayo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Alber</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. T.</given-names>
            <surname>Schütt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Dähne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Erhan</surname>
          </string-name>
          ,
          <string-name>
            <surname>B. Kim,</surname>
          </string-name>
          <article-title>The (un) reliability of saliency methods, Explainable AI: Interpreting, explaining and visualizing deep learning (</article-title>
          <year>2019</year>
          )
          <fpage>267</fpage>
          -
          <lpage>280</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>D.</given-names>
            <surname>Alvarez-Melis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. S.</given-names>
            <surname>Jaakkola</surname>
          </string-name>
          ,
          <article-title>On the robustness of interpretability methods</article-title>
          , arXiv preprint arXiv:
          <year>1806</year>
          .
          <volume>08049</volume>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>D.</given-names>
            <surname>Slack</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hilgard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Jia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Lakkaraju</surname>
          </string-name>
          ,
          <article-title>Fooling lime and shap: Adversarial attacks on post hoc explanation methods</article-title>
          ,
          <source>in: Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>180</fpage>
          -
          <lpage>186</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>G.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>Interpret federated learning with shapley values</article-title>
          , arXiv preprint arXiv:
          <year>1905</year>
          .
          <volume>04519</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>P.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Lu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. C.</given-names>
            <surname>Hung</surname>
          </string-name>
          ,
          <string-name>
            <surname>Evfl:</surname>
          </string-name>
          <article-title>An explainable vertical federated learning for data-oriented artificial intelligence systems</article-title>
          ,
          <source>Journal of Systems Architecture</source>
          <volume>126</volume>
          (
          <year>2022</year>
          )
          <fpage>102474</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>A.</given-names>
            <surname>Karlsson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Nowaczyk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Pashami</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Asadi</surname>
          </string-name>
          ,
          <article-title>Mind the data, measuring the performance gap between tree ensembles and deep learning on tabular data</article-title>
          ,
          <source>in: International Symposium on Intelligent Data Analysis</source>
          , Springer,
          <year>2024</year>
          , pp.
          <fpage>65</fpage>
          -
          <lpage>76</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>L.</given-names>
            <surname>Grinsztajn</surname>
          </string-name>
          , E. Oyallon, G. Varoquaux,
          <article-title>Why do tree-based models still outperform deep learning on typical tabular data?</article-title>
          ,
          <source>Advances in neural information processing systems</source>
          <volume>35</volume>
          (
          <year>2022</year>
          )
          <fpage>507</fpage>
          -
          <lpage>520</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Gai</surname>
          </string-name>
          ,
          <article-title>Decision tree-based federated learning: A survey</article-title>
          ,
          <source>Blockchains</source>
          <volume>2</volume>
          (
          <year>2024</year>
          )
          <fpage>40</fpage>
          -
          <lpage>60</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10] K. Cheng, T. Fan,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Jin</surname>
          </string-name>
          , Y. Liu,
          <string-name>
            <given-names>T.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Papadopoulos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <article-title>Secureboost: A lossless federated learning framework</article-title>
          ,
          <source>IEEE Intelligent Systems</source>
          <volume>36</volume>
          (
          <year>2021</year>
          )
          <fpage>87</fpage>
          -
          <lpage>98</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>X.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Hu</surname>
          </string-name>
          , W. Liu,
          <string-name>
            <given-names>H.</given-names>
            <surname>Feng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Peng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Hong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Ren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Qin</surname>
          </string-name>
          , Opboost: a vertical
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>