<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Infodeslib: Python Library for Dynamic Ensemble Learning using Late Fusion of Multimodal Data</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Firuz Juraev</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Shaker El-Sappagh</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Tamer Abuhmed</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>College of Computing and Informatics, Sungkyunkwan University</institution>
          ,
          <country country="KR">South Korea</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Faculty of Computer Science and Engineering, Galala University</institution>
          ,
          <country country="EG">Egypt</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>There has been a notable increase in research focusing on dynamic selection (DS) techniques within the field of ensemble learning. This leads to the development of various techniques for ensembling multiple classifiers for a specific instance or set of instances during the prediction phase. Despite this progress, the design and development of DS approaches with late fusion settings and their explainability remain unexplored. This work proposes an open-source Python library, Infodeslib, to address this gap. The library provides an implementation of several DS techniques, including four dynamic classifier selections and seven dynamic ensemble selection techniques, all of which are integrated with late data fusion settings and novel explainability features. Infodeslib ofers flexibility and customization options, making it a versatile tool for various complex applications that require the fusion of multimodal data and various explainability features. Multimodal data, which integrates information from diverse sources or sensor modalities, is a common and essential setting for real-world problems, enhancing the robustness and depth of data analysis. These data can be fused in two main ways: early fusion, where diferent modalities are combined at the feature level before model training, and late fusion, where each modality is processed separately and the results are combined at the decision level. The library is fully documented following the Read the Docs standards. The documentation, code, and examples are available anonymously on GitHub at https://github.com/InfoLab-SKKU/infodeslib.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Ensemble of classifiers</kwd>
        <kwd>Dynamic classifier selection</kwd>
        <kwd>Dynamic ensemble selection</kwd>
        <kwd>multimodal data fusion</kwd>
        <kwd>Late fusion</kwd>
        <kwd>Machine learning</kwd>
        <kwd>Explainable AI</kwd>
        <kwd>Python</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Ensemble learning is a thriving domain within the fields
of machine learning and pattern recognition [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ]. With
all the diverse ensemble classifiers available, each classifier
approaches the problem from a diferent perspective. The
main idea of ensemble learning is to leverage a group of
classifiers to provide comprehensive coverage of the learned
task [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. By utilizing diverse models that exhibit distinct
decision boundaries, ensemble learning seeks to maximize
the accuracy and efectiveness of the overall classification
process. As a result, the performance of ensemble
classiifers is better than any of its base classifiers [
        <xref ref-type="bibr" rid="ref4 ref5">4, 5</xref>
        ]. This
is because each base classifier concentrates on the specific
region of the error space and combining the decisions of
these classifiers improves the overall ensemble’s decisions.
Ensemble learning approaches can be broadly classified into
two categories: static and dynamic selection approaches
[
        <xref ref-type="bibr" rid="ref6 ref7">6, 7</xref>
        ]. In static selection [
        <xref ref-type="bibr" rid="ref8 ref9">8, 9</xref>
        ], a predetermined group of
classifiers is selected, and this group is utilized to make
decisions for each new test instance. In dynamic selection
[
        <xref ref-type="bibr" rid="ref10 ref11 ref12">10, 11, 12</xref>
        ], a new group of classifiers is selected for each
test instance, and this group is employed to make a decision
for that specific instance.
      </p>
      <p>
        Since real-world datasets are often complex and consist
of multiple feature groups or so-called ‘modalities’,
ensemble learning is a popular candidate to be used to combine
multiple models to improve the performance and robustness
of predictive models. One approach to ensemble learning
is early fusion, where all modalities are merged in a pool
for the classifiers to capture the potential interaction and
interdependencies among the modalities using either static
[
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] or dynamic selection [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ].
      </p>
      <p>
        Another approach to ensemble learning is the late
fusion or decision fusion, where each classifier in the pool
is trained with diferent feature groups or combinations of
feature groups to achieve greater diversity in the model pool.
This diversity is crucial for constructing a robust ensemble
that can efectively generalize to previously unseen data.
Moreover, late fusion provides more flexibility as classifiers
are assigned to diferent modalities considering that certain
classifiers are best to model certain modalities [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ].
      </p>
      <p>
        In current literature, late fusion-based ensemble learning
is solely available with static classifiers selection [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ], and
most of these studies show the superiority of late fusion
compared to early fusion for static ensemble [
        <xref ref-type="bibr" rid="ref17 ref18 ref19">17, 18, 19</xref>
        ].
This motivates us to explore the performance of late fusion
in dynamic selection compared to early fusion; however, to
the best of our knowledge, no study or implementation has
been conducted to examine the performance of late fusion
in dynamic selection settings. This work aims to implement
diferent types of dynamic selection techniques in the late
fusion setting. By doing so, we can explore the performance
of late fusion-based ensemble learning under dynamic
selection modeling, gaining a deeper understanding of its
potential advantages and limitations.
      </p>
      <p>
        Resulting late fusion-based dynamic ensembles are
expected to improve the performance of the resulting
classifiers. However, these models are black boxes and not
understandable. Trustworthy classifiers that are applicable
in the real world need to be interpretable. Explainable AI
(XAI) has gained significant attention in recent years [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ],
as it is crucial to provide insights into the decision-making
process of machine learning models. However, despite the
growing interest in this area, there is a lack of
explainability features for ensemble learning techniques, which are
increasingly used in complex real-world applications to
improve the trustworthiness of resulting models. To the best
of our knowledge, no study in the literature and no Python
packages are provided to implement XAI capabilities for
dynamic ensemble classifiers. This study aims to address
this research gap by developing a Python package that
offers novel explainability techniques for ensemble models,
making them accessible and informative for both domain
experts and developers.
• We extended the literature on dynamic ensemble
modeling by implementing four dynamic classifier
selection techniques and seven dynamic ensemble
selection techniques, incorporating a late fusion of
multiple modalities (see Table 1).
• We propose three types of novel explainability that
provide deep and suitable XAI for dynamic selection
techniques: Case-Based Reasoning, deep-based
classifiers contributions, and local feature importance.
• We compare the performance of the proposed
techniques with existing approaches on four well-known
and real-world multimodal datasets: Alzheimer’s
Disease Neuroimaging Initiative (ADNI), Credit
Card Clients, National Alzheimer’s Coordinating
Center, and Parkinson’s Progression Markers
Initiative (PPMI). We also tested the proposed techniques
in the Samarkand Neonatal Center dataset which is
collected by our team with the help of physicians.
• The implemented techniques have been included
in a standard public library called ‘Infodeslib’
following the industry-standard PEP 8 coding
guidelines, and Infodeslib is also clearly documented
in accordance with the Read the Docs standards:
https://infodeslib.readthedocs.io/en/latest/
• We ofer a wide range of valuable functions that
enable the assessment and evaluation of the excellence
and eficacy of the selected pool.
      </p>
      <p>The study is organized as follows. Section 1 highlights
the software framework of the proposed late fusion dynamic
ensemble learning. Section 2 presents Installation and Usage,
Section 4 discusses the performance analysis, and Section 5
introduces possible package extensions. Section 6 concludes
the paper.</p>
    </sec>
    <sec id="sec-2">
      <title>1. Late Fusion Dynamic Ensemble</title>
    </sec>
    <sec id="sec-3">
      <title>Framework</title>
      <p>In this section, we provide an overview of the late fusion
dynamic ensemble framework in algorithmic and visual
formats. This encompasses a thorough dissection of the
primary stages involved, along with step-by-step explanations
of the framework’s methodology.</p>
      <p>Since late fusion dynamic ensemble utilizes the decision
values obtained from each modality and fuses them using a
specific fusion mechanism  (such as averaging, weighted
averaging, majority voting, etc), let us assume that classifier
 is applied to modality . The final prediction can be
expressed as:
 =  (1(1), 2(2), ..., ())
(1)</p>
      <p>The proposed concept of dynamic selection with late
fusion is illustrated in Figure 1, which outlines a framework
consisting of three key stages: training, selection, and
prediction. Additionally, the concept is detailed algorithmically
in Algorithm 1.</p>
      <p>Training Phase. A pool of classifiers is selected and
assigned diferent feature sets. The classifiers within the
pool are selected based on their diversity, ensuring a wide
range of decision-making capabilities. Each feature set used
by selected classifiers is extracted from the same modality
to generate a homogeneous feature set. For example, in
the medical domain, demographic and MRI features are
diferent modalities that could be used to train two diferent
classifiers. Each classifier in the pool is then trained and
optimized with its designated feature set, resulting in a pool
of trained classifiers to be utilized in the next phases (1-4
lines in Algorithm 1).</p>
      <p>Selection Phase. During the selection phase (5-12 lines
of Algorithm 1), a region of competence (RoC) is determined
for a given new test instance by selecting the nearest
samples from the validation data (DSEL). Subsequently, each
classifier in the pool is evaluated on the samples within the
RoC, and a measure of competence is calculated for each
classifier. The specific method employed to compute the
competence varies depending on the chosen DS technique
(9-10 lines in Algorithm 1). Once the competencies of each
classifier in the pool are calculated, the DS techniques use
their own selection criteria to identify the most competent
classifiers. These criteria are specific to each DS technique.
If no competent classifier satisfies the criteria for a given
DS technique, all classifiers in the pool are selected to make
the final decision.</p>
      <p>Prediction Phase. During the final phase, the selected
classifiers are utilized to predict the class of a given test
instance, and their individual predictions are combined to
generate a final prediction. To provide more accurate
decisions, each of the selected classifiers could be weighed based
on its level of competence during the aggregation process
(line 13 in Algorithm 1).</p>
    </sec>
    <sec id="sec-4">
      <title>2. Installation and Usage</title>
      <p>Users can conveniently install the most recent version
of Infodeslib via pip, the Python package manager, by
executing the command pip install infodeslib.
Alternatively, the library can be installed via the
GitHub address, using the command pip install
git+https://github.com/InfoLab-SKKU/infodeslib.</p>
      <p>To use the implemented methods in Infodeslib, a list of
classifiers and feature sets must be provided as input. The
classifiers in the list can be of any type from the scikit-learn
library and should be trained on the corresponding feature
set before being used as input.</p>
      <p>Once the pool of classifiers and feature sets has been
initialized, the method fit(X_dsel, y_dsel) is applied to fit the
Dynamic Selection method, where (X_dsel, y_dsel) is the
validation dataset (DSEL) with true labels. Predictions for each
test instance x can be obtained using either the predict(x)
or predict_proba(x) methods. In the example provided
below, we demonstrate the steps involved in implementing
the KNORA-U technique.</p>
      <p>When utilizing the predict(X) method, an additional
parameter "plot" can be included to obtain explainability for
each test instance. By setting plot=True, explainability for
the given test instance can be visualized through a variety
of methods (see more details in Section 3).</p>
      <p>Infodeslib Methods. Figure 2 provides an overview
of the key methods of our library while other
supporting methods are available in the documentation of the
library. Some of these methods such as fit() , predict(),
predict_proba(), and score() are well-known and require
no detailed explanation; there are several other methods
that are particularly useful for pool generation and
obtaining information about new test samples. To facilitate
pool generation, we have implemented three additional
methods: get_average_accuracy(), get_pool_diversity(), and
get_coverage_score(). get_average_accuracy() method
computes the average performance of the classifiers in the pool
on the validation data. get_pool_diversity() method
calculates the diversity between classifiers in the pool and
requires the diversity measure type as a parameter. It supports
several diversity functions such as Q-statistic, Correlation
Coeficient, Disagreement Measure, Double Fault,
Negative Double Fault, and Ratio Errors. get_coverage_score()
method determines the number of samples in the DSEL data
that can be accurately predicted by any model in the given
pool. This information is particularly useful for evaluating
the coverage of the pool and ensuring that all samples are
accurately classified by at least one model. The prediction
process in machine learning often involves the use of
ensemble methods, where multiple classifiers are combined
to improve performance. Within these ensembles, three
methods play a crucial role: get_region_of_competence(x),
estimate_competence(roc), and select(competences).</p>
      <p>get_region_of_competence(x) method identifies the
region of competence for a given test sample by returning
the k nearest neighbors from the validation dataset. This
is achieved by applying the k-nearest neighbors algorithm.
The estimate_competence(roc) method calculates the
competence of each classifier in the ensemble on the region of
competence. The competence calculation difers depending
on the technique being used. For example, the k-Nearest
Oracle Union (KNORA-U) technique calculates the accuracy
of each classifier on the region of competence. The Dynamic
fit(X, y)</p>
      <p>Prepare the DS model by pre-processing
the information required to apply the DS
methods.
score(X, y)</p>
      <p>Return the mean accuracy on the given
data and labels.</p>
      <p>Pool
get_average_accuracy()</p>
      <p>Return the mean accuracy of
classifiers in the pool.
get_pool_diversity()</p>
      <p>Return the mean and list of
diversity scores between
classifiers in the pool.
get_coverage_score()</p>
      <p>Return the explainability how
the given pool of classifiers
can cover the task on validation
data.</p>
      <p>predict(X, plot=False)</p>
      <p>Return the class label for
each sample in X. plot=True
for getting the explainability.
predict proba(X)</p>
      <p>Return the probabilities for
each sample in X.</p>
      <p>Single instance
get_region_of_competence(x)</p>
      <p>Return k nearest samples of the
given test sample from validation
dataset.
estimate_competence(roc)</p>
      <p>Return the competences of each
base classifier on k nearest
samples from RoC.
select(competences)</p>
      <p>Return all base classifiers that are
competent enough.
get_rareness_score(x)</p>
      <p>Return the explainability how the
given test sample is rare on
trainingand validation data.</p>
      <p>Hyperparameters
k: int - number of neighbors used to estimate the competence of the base classifiers.
DFP: boolean - determines if the dynamic frienemy pruning is applied.
knn_metric: str or callable - distance metric utilized by the k-NN classifier.
dimensionality_reduction: boolean - determines if dimension reduction is applied.
reduction_technique: str or callable - technique utilized for dimension reduction.
n_components: int - number of components to keep.</p>
      <p>cbr_features: list - list of features to show in cased based reasoning XAI.
Ensemble Selection KNN (DESKNN) technique, on the other
hand, computes each classifier’s accuracy and diversity on
RoC and uses these metrics to assess its competence.</p>
      <p>
        select(competences) method selects the most competent
classifiers from the ensemble to make a prediction.
Diferent techniques may use diferent criteria for determining
the competence of a classifier, such as the number of
samples classified correctly within the region of competence.
For instance, the KNORA-U technique selects a classifier
if it has classified at least one sample within the region
of competence. Once the competent classifiers have been
identified, their competence values are used as weights in
aggregating their predictions. To evaluate a single test
instance, our library includes get_rareness_score(x) method,
which provides a detailed description of the instance. The
method evaluates whether there are many similar samples
to the given instance in the training and validation datasets,
allowing users to determine the rarity of the instance. If
the instance is an outlier, the method provides
information about how far it is from other classes. Furthermore,
get_rareness_score(x) method uses K-means clustering to
provide a potential class for the instance and generates
tables indicating which features of the instance make it similar
to this class. This approach provides valuable insights into
the characteristics of the instance and its potential
classification, aiding in the development of more accurate models.
Hyperparameters. Optimizing hyperparameters is a
critical step for improving the performance of ensemble learning
models. This can be achieved through various techniques,
including basic approaches such as grid search and random
search in Sklearn, as well as more advanced techniques like
Genetic algorithms, Bayesian optimization, and others. Our
library is designed to work seamlessly with other Python
packages such as TPOT [
        <xref ref-type="bibr" rid="ref28">28</xref>
        ], Scikit-Optimize [
        <xref ref-type="bibr" rid="ref29">29</xref>
        ], Optuna
[
        <xref ref-type="bibr" rid="ref30">30</xref>
        ], Hyperopt [
        <xref ref-type="bibr" rid="ref31">31</xref>
        ], BayesianOptimization [
        <xref ref-type="bibr" rid="ref32">32</xref>
        ], GPyOpt
[
        <xref ref-type="bibr" rid="ref33">33</xref>
        ], Optunity [
        <xref ref-type="bibr" rid="ref34">34</xref>
        ], and similar packages that implement
these advanced optimization techniques. This allows users
to leverage a variety of optimization methods to obtain the
best possible hyperparameters for their ensemble models.
      </p>
      <p>In our library, there are several key hyperparameters that
users can adjust to optimize the performance of ensemble
learning models. We present these key hyperparameters
along with their default values, which have been shown to
produce satisfactory results in the majority of cases. One
of the main hyperparameters is k (default: 7), which
represents the number of neighbors to be considered when
determining the region of competence. Another important
hyperparameter is DFP (default: False), which stands for
dynamic pruning technique and is particularly useful for
imbalanced datasets. In addition, users can also specify the
knn_metric (default: ’minkowski’), which determines the
distance metric used when computing distances between the
test sample and other samples in the validation dataset. Our
library provides several common metrics such as Minkowski,
cosine, Manhattan, and Euclidean, as well as the option for
users to define their own custom metric function. To
handle high-dimensional datasets, we also ofer a
dimensionality_reduction (default: False) hyperparameter, which
allows users to reduce the number of dimensions used in
calculating distances between samples. This can be achieved
using either Principal Component Analysis (PCA) or Kernel
PCA, or by specifying a custom dimensionality reduction
technique using the next reduction_technique (default:
’pca’). The n_component (default: 20) hyperparameter
determines the number of components to be retained if a
reduction technique is selected. Lastly, for those interested
in explainability, our library provides the cbr_features
(default: None) hyperparameter, which allows users to specify
a list of important features to be included in similar cases
data for Case-Based Reasoning.</p>
    </sec>
    <sec id="sec-5">
      <title>3. Model Explainability</title>
      <p>
        In the current version of our library, we ofer three main
XAI techniques: case-based reasoning, deep-based
classiifer contribution, and local feature importance [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ]. The
case-based reasoning technique aims to ofer domain
experts an explanation of the model’s prediction process for a
given test sample by presenting them with similar samples
and their corresponding labels found within the region of
x Test sample
0: AD
1: sMCI
2: CN
3: pMCI
      </p>
      <p>Region of
Competence: Φ
a) Estimating the region of competence (RoC) in validation dataset.</p>
      <p>Samples in the region of competence with selected features and labels
competence. This approach closely resembles how domain
experts make decisions in real-world situations, as they
frequently compare current cases with historical ones from
their experience. The deep-based classifier contribution
technique enables users to comprehend the contribution of
each selected classifier in the decision-making process for
a given test sample. Finally, the local feature importance
technique is a prevalent explainability method that
identiifes the most crucial features and their corresponding Shap
values for each selected classifier.</p>
      <p>Case-based reasoning. For example, in the case of the
KNORA-U technique, in the selection phase, the nearest
neighbors for each test instance are estimated in the
validation dataset based on their close similarity to the test sample.
The selected samples are used to generate the region of
competence for evaluating and selecting classifiers in the pool.
Figure 4 a) illustrates an example in which the given test
sample (light blue x) falls within the area of class 2, and
seven nearest samples are selected, six of which belong to
class 2 (blue dots), while one belongs to class 1 (green dot).
This finding suggests that, for the given test sample, the
chance of it being classified as class 2 is high. Moreover,
these samples can also be leveraged for conducting
casebased reasoning, which may be particularly valuable for
physicians, given that our dataset is in the medical domain.
Figure 4 b) provides comprehensive information about all
nearest samples within the region of competence, enabling
physicians to compare and contrast similar samples and
their corresponding labels or diagnoses.</p>
      <p>Deep-based classifiers contributions. After selecting
the group of classifiers for making the final decision, it may
be unclear how each classifier in the pool contributed to the
decision or what their individual predictions were for the
new test sample. In order to provide a more comprehensive
understanding of the decision-making process, an additional
Selected Classifier
[Classifier 1] XGB
[Classifier 2] XGB
[Classifier 3] MLP
[Classifier 4] SVC
[Classifier 5] XGB
[Classifier 6] KNN</p>
      <p>Contribution on decision
0.0 0.5 1.0 1.5 2.0
level of explainability can be utilized. This is illustrated in
Figure 5, which provides detailed information about each
classifier in the pool, including their competence level,
individual prediction on the new test sample, and confidence
level. This explanation provides valuable insight for the
development of an ensemble model, as it allows developers
to identify classifiers that may have a negative impact on
decision-making. For instance, as shown in Figure 5, it is
evident that most selected classifiers predict the label of the
given test sample as 2 with high confidence, while the SVC
classifier predicts it as 3. The SVC classifier demonstrates
a higher level of competence in the region of competence,
indicating that it has a more significant influence on the
decision. If this classifier consistently has a negative impact
on many test samples, it may be possible to remove it from
the pool of classifiers.</p>
      <p>Local feature importance. In addition to
understanding how the classifiers contributed to the decision-making
process, it is also important to identify which features were
particularly influential in making those decisions. For the
example mentioned earlier, we provide local feature
importance for each selected classifier, which can be visualized
through Figure 3.</p>
      <p>Furthermore, our proposed ensemble models have the
ability to provide interpretable explanations using two
approaches: surrogate model explainability and post-hoc
explainability methods. The surrogate model approach
involves creating a simplified model that roughly represents
the behavior of the original ensemble model and using this
model to explain the ensemble’s decisions. On the other
hand, post-hoc explainability techniques involve analyzing
the ensemble model’s decisions after they have been made
and providing explanations based on the input features that
contributed the most to the decision. Both methods treat
our ensemble model as a black box model.</p>
    </sec>
    <sec id="sec-6">
      <title>4. Performance Analysis</title>
      <p>Within this section, We compare the performance of the
proposed architecture with the existing approaches. we
provide an overview of the datasets that have been utilized
along with a detailed analysis of our proposed techniques.</p>
      <sec id="sec-6-1">
        <title>4.1. Evaluation Datasets</title>
        <p>In this section, we outline the five datasets utilized to
compare Infodeslib with existing models.</p>
        <p>
          Alzheimer’s Disease Neuroimaging Initiative
(ADNI) dataset [
          <xref ref-type="bibr" rid="ref35">35</xref>
          ]. The study includes a total of
1,371 subjects, with a male gender representation of
54.5%. Participants have been classified into four distinct
categories based on their clinical diagnosis, including
Cognitive Normal (CN), Stable Mild Cognitive Impairment
(sMCI), Progressive Mild Cognitive Impairment (pMCI), and
Alzheimer’s Disease (AD) [
          <xref ref-type="bibr" rid="ref36">36</xref>
          ]. The distribution of these
classes is as follows: 419 CN, 473 sMCI, 140 pMCI, and 339
AD individuals. The dataset has four distinct modalities
or feature groups, which contain demographics, cognitive
scores, assessment tests, and MRI features.
        </p>
        <p>
          Credit Card Clients dataset. The study includes a vast
participant cohort of 30,000 individuals, with the dataset
sourced from the UC Irvine Machine Learning Repository
[
          <xref ref-type="bibr" rid="ref37">37</xref>
          ]. This is a classification problem that involves
determining whether or not a client will make their next payment.
The two distinct classes are labeled as ’no’ and ’yes’, with
23,364 and 6,636 instances, respectively. The dataset has
four distinct modalities of features, including demographics,
ifnancial, and payment history features.
        </p>
        <p>
          National Alzheimer’s Coordinating Center (NACC)
dataset [
          <xref ref-type="bibr" rid="ref38">38</xref>
          ]. In this study, we examined a total of 37,547
patients focusing on the Global Clinical Dementia Rating
(CDRGLOB) as the primary task. CDRGLOB categorizes
patients into five classes based on dementia severity: no
impairment (8,253 patients), mild impairment (15,097
patients), moderate impairment (8,346 patients), and severe
impairment (5,851 patients). Our analysis included six
specific modalities for investigation: demographics, physical
health, medications, health history, neuropsychiatric
inventory questionnaire, and the geriatric depression scale. These
modalities were chosen to comprehensively assess various
aspects related to dementia and overall patient health [
          <xref ref-type="bibr" rid="ref39">39</xref>
          ].
        </p>
        <p>
          Parkinson’s Progression Markers Initiative (PPMI)
dataset [
          <xref ref-type="bibr" rid="ref40">40</xref>
          ]. Our study involves 952 patients and
focuses on a binary classification task to diferentiate between
healthy individuals and those diagnosed with Parkinson’s
disease (PD). Among these patients, 389 are categorized
as healthy, while 563 have been diagnosed with PD. The
dataset encompasses various information modalities,
including subject characteristics, biospecimen data, medical
history records, motor function assessments, and non-motor
features. This comprehensive dataset enables a thorough
analysis to identify potential diagnostic markers and factors
associated with PD, facilitating improved understanding
and diagnosis of the disease [
          <xref ref-type="bibr" rid="ref41">41</xref>
          ].
        </p>
        <p>Samarkand Neonatal Center dataset. Our study
involved 347 neonates from the intensive care unit at
Samarkand Neonatal Center. The dataset was collected by
our team by collaborating physicians in the hospital for a
binary classification task to predict whether a neonate
survives or passes away. Among these neonates, 303 survived
and 44 died during the study period. The dataset comprises
a comprehensive set of features categorized into multiple
modalities: demographic information, the mother’s medical
history and information, general notes on the neonate’s
condition, results from blood tests, and APGAR scores (a
standardized assessment of a neonate’s health at birth).</p>
      </sec>
      <sec id="sec-6-2">
        <title>4.2. Results</title>
        <p>
          This section contains a comprehensive analysis and
comparison of various machine-learning approaches against our
proposed late-fusion dynamic ensemble selection model.
We collect and present the testing results for each of the
considered models. To ensure greater consistency in the
results, we have applied the 10-holdout testing method [
          <xref ref-type="bibr" rid="ref42">42</xref>
          ].
The results are presented in the form of (mean ± standard
deviation). As a pool of classifiers, we utilized the following
heterogeneous baseline algorithms: XGboost (XGB),
LightGBM (LGBM), Random Forest (RF), Support Vector Classifier
(SVC), Multi-Layer Perceptron (MLP), Decision Tree (DT),
and k-Nearest Neighbors (KNN).
        </p>
        <p>Results based on ADNI dataset. Table 2 and Figure
6 a) show the top-performing results achieved by the
individual models, as well as the static ensemble with early
fusion, the dynamic ensemble with early fusion, the static
ensemble with late fusion, and our proposed technique
the dynamic ensemble with late fusion setting. From each
group, we selected the best-performed techniques and the
results show that our dynamic ensemble techniques,
KNORAU and KNORAU-W outperform all existing approaches
with 89.52% and 89.84% accuracy. In comparison, a static
ensemble with late fusion, voting classifier, achieves an
accuracy of 89.29%. This performance is close to the performance
of our model and surpasses that of early fusion techniques.
This result supports our claim for the significance of late
fusion in producing accurate ensemble models.</p>
        <p>Results based on Credit Card Clients dataset. Table 3
and Figure 6 b) present the results obtained from the analysis
of the Credit Card Clients dataset, following a similar format
to the previous dataset. Our proposed techniques have once
again outperformed the existing approaches in this instance.
Specifically, KNOP and KNORAU-W, utilizing the late fusion
setting, have achieved the highest accuracy scores of 86.65%
and 86.73%, respectively. In comparison, the static ensemble
methods that apply late fusion, specifically the voting and
stacking classifiers, demonstrate accuracies of 85.72% and
85.08%, respectively. In contrast, the ensemble methods
that employ early fusion achieve the highest accuracy of
84.16%, with the dynamic selection technique known as
KNORA-E. These results support our argument regarding
the importance of utilizing late fusion for the purpose of
producing highly accurate ensemble models.</p>
        <p>Results based on NACC dataset. Table 4 highlights
the results from the analysis of the National Alzheimer’s
Coordinating Center dataset, structured similarly to the
previous dataset. Among all existing techniques, the dynamic
ensemble models with late fusion demonstrate notably
superior performance. Specifically, the weighted KNORAU
(KNORAU-W) and DESP achieve the highest scores at 90.20%
and 91.16%, respectively. Given the substantial dataset size,
the results are well-balanced across various metrics.</p>
        <p>Results based on PPMI dataset. Table 5 presents the
results obtained from the analysis of the Parkinson’s
Progression Markers Initiative dataset, following a format similar
to the previous dataset. Within this dataset, the techniques
DESP and KNOP, utilizing late fusion settings, exhibit the
most robust performance among other algorithms,
achieving accuracies of 95% and 95.1%, respectively. Additionally,
static ensemble models with late fusion settings demonstrate
strong performance at 94.6% accuracy using a voting
technique. These results only marginally exceed those achieved
with LGBM alone, which achieved a performance of 93.9%.</p>
        <p>The fact that the LGBM achieved a high accuracy of 93.9%
suggests that the task at hand is not very complex.
Improving accuracy beyond this point becomes more challenging
when a basic technique like LGBM already performs well.
Essentially, reaching significantly higher accuracies with
more advanced methods might be dificult because the task
is relatively straightforward.</p>
        <p>Results based on Samarkand Neonatal Center
dataset. Table 6 presents the results obtained from
analyzing the Samarkand Neonatal Center ICU dataset,
following a similar structure to the previous datasets. Due to the
dataset’s small size, the results may not be consistent or
balanced across diferent metrics. Nonetheless, our proposed
late fusion-based dynamic ensemble models achieve notably
higher performance compared to other techniques, reaching
77.57% accuracy with the KNOP technique.</p>
        <p>Across all five datasets analyzed, the importance of late
fusion can be seen in the results. In each dataset, the
dynamic ensemble models with late fusion settings
outperform other existing models. Combining late fusion with
dynamic ensemble learning consistently delivers
promising and improved results. This highlights the efectiveness
and reliability of employing late fusion techniques within
dynamic ensemble models across various datasets.</p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>5. Library extension</title>
      <p>The primary focus of our paper is to introduce a novel
approach to dynamic ensemble selection (DES) that utilizes a
late fusion strategy for efectively fusing multi-modal data
and ofers a high degree of explainability for dynamic
selection techniques. Our current library ofers implementations
of four dynamic classifier selection and seven dynamic
ensemble selection techniques that use a late fusion strategy.
In addition, the library includes several features and options
that enhance its performance and capability. Furthermore,
the library provides three diferent types of explainability to
help users gain insights into the decision-making processes
of the models. Finally, the library has been designed to be
compatible with other important libraries, allowing users
to easily integrate it into their existing workflows.</p>
      <p>We plan to continue exploring the domain of
explainability in ensemble learning by proposing additional techniques
for providing comprehensive explanations to domain
experts. Our goal is to enhance our library’s ability to provide
context-based explanations that are tailored to the specific
needs of users. Additionally, we aim to incorporate what-if
explainability features that enable developers to gain deeper
insights into the behavior of their ensemble models. These
features will be included in future versions of our library.</p>
      <p>Through our experimental evaluations, we have
discovered that selecting an appropriate pool of classifiers with
matching feature groups is a critical aspect of successful
ensemble modeling. However, identifying the ideal
combination of classifiers for the pool remains a challenging task.
In future versions of our library, we plan to address this
issue by developing an automatic optimization process for
the selection of the optimal pool of classifiers. We believe
this to be a crucial task in the field of ensemble learning,
and we are committed to exploring ways to simplify this
process and make it more efective.</p>
    </sec>
    <sec id="sec-8">
      <title>6. Conclusion</title>
      <p>This paper presents a novel approach to dynamic selection
using a late fusion setting, which is implemented across four
dynamic classifier selection and seven dynamic ensemble
selection techniques. This late fusion-based approach is
particularly well-suited for complex tasks based on
multimodal datasets containing multiple feature groups, which
are common in real-world scenarios. As a result, the role of
late fusion is crucial in the context of ensemble learning for
ensuring diversity in the pool of classifiers. Furthermore,
we introduce a novel approach to explainability for dynamic
selection techniques. Our proposed approach goes beyond
the traditional methods and provides a more in-depth and
nuanced understanding of the dynamic selection process.
The efectiveness of our proposed techniques is evaluated
through a comprehensive comparison with existing
baseline approaches. The experimental results demonstrate the
superior performance of our proposed techniques over the
existing approaches, highlighting the potential of our
approach to improving the accuracy and reliability of ensemble
learning systems.</p>
    </sec>
    <sec id="sec-9">
      <title>Acknowledgments</title>
      <p>This work was supported by the National Research
Foundation of Korea(NRF) grant funded by the Korea
government(MSIT)(No. 2021R1A2C1011198), (Institute for
Information &amp; communications Technology Planning &amp;
Evaluation) (IITP) grant funded by the Korea government (MSIT)
under the ICT Creative Consilience Program
(IITP-20212020-0-01821), and AI Platform to Fully Adapt and Reflect
Privacy-Policy Changes (No.RS-2022-II220688).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>H.</given-names>
            <surname>Xiao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Xiao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>Ensemble classification based on supervised clustering for credit scoring</article-title>
          ,
          <source>Applied Soft Computing</source>
          <volume>43</volume>
          (
          <year>2016</year>
          )
          <fpage>73</fpage>
          -
          <lpage>86</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>D.</given-names>
            <surname>Di Nucci</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Palomba</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Oliveto</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. De Lucia</surname>
          </string-name>
          ,
          <article-title>Dynamic selection of classifiers in bug prediction: An adaptive method</article-title>
          ,
          <source>IEEE Transactions on Emerging Topics in Computational Intelligence</source>
          <volume>1</volume>
          (
          <year>2017</year>
          )
          <fpage>202</fpage>
          -
          <lpage>212</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>O.</given-names>
            <surname>Sagi</surname>
          </string-name>
          , L. Rokach,
          <article-title>Ensemble learning: A survey</article-title>
          ,
          <source>Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery</source>
          <volume>8</volume>
          (
          <year>2018</year>
          )
          <article-title>e1249</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>L. I.</given-names>
            <surname>Kuncheva</surname>
          </string-name>
          ,
          <article-title>A theoretical study on six classifier fusion strategies</article-title>
          ,
          <source>IEEE Transactions on pattern analysis and machine intelligence</source>
          <volume>24</volume>
          (
          <year>2002</year>
          )
          <fpage>281</fpage>
          -
          <lpage>286</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>M.</given-names>
            <surname>Fernández-Delgado</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Cernadas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Barro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Amorim</surname>
          </string-name>
          ,
          <article-title>Do we need hundreds of classifiers to solve real world classification problems?</article-title>
          ,
          <source>The journal of machine learning research 15</source>
          (
          <year>2014</year>
          )
          <fpage>3133</fpage>
          -
          <lpage>3181</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A. S.</given-names>
            <surname>Britto</surname>
          </string-name>
          <string-name>
            <surname>Jr</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Sabourin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. E.</given-names>
            <surname>Oliveira</surname>
          </string-name>
          ,
          <article-title>Dynamic selection of classifiers-a comprehensive review</article-title>
          ,
          <source>Pattern recognition 47</source>
          (
          <year>2014</year>
          )
          <fpage>3665</fpage>
          -
          <lpage>3680</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>X.</given-names>
            <surname>Dong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Cao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Shi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Ma</surname>
          </string-name>
          ,
          <article-title>A survey on ensemble learning</article-title>
          ,
          <source>Frontiers of Computer Science</source>
          <volume>14</volume>
          (
          <year>2020</year>
          )
          <fpage>241</fpage>
          -
          <lpage>258</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>L.</given-names>
            <surname>Breiman</surname>
          </string-name>
          , Bagging predictors,
          <source>Machine learning 24</source>
          (
          <year>1996</year>
          )
          <fpage>123</fpage>
          -
          <lpage>140</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>G.</given-names>
            <surname>Sakkis</surname>
          </string-name>
          , I. Androutsopoulos,
          <string-name>
            <given-names>G.</given-names>
            <surname>Paliouras</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Karkaletsis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. D.</given-names>
            <surname>Spyropoulos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Stamatopoulos</surname>
          </string-name>
          ,
          <article-title>Stacking classifiers for anti-spam filtering of e-mail, arXiv preprint cs/0106040 (</article-title>
          <year>2001</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>A. H.</given-names>
            <surname>Ko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Sabourin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. S.</given-names>
            <surname>Britto</surname>
          </string-name>
          <string-name>
            <surname>Jr</surname>
          </string-name>
          ,
          <article-title>From dynamic classifier selection to dynamic ensemble selection</article-title>
          ,
          <source>Pattern recognition 41</source>
          (
          <year>2008</year>
          )
          <fpage>1718</fpage>
          -
          <lpage>1731</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>R. M. Cruz</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Sabourin</surname>
            ,
            <given-names>G. D.</given-names>
          </string-name>
          <string-name>
            <surname>Cavalcanti</surname>
          </string-name>
          ,
          <article-title>Dynamic classifier selection: Recent advances and perspectives</article-title>
          ,
          <source>Information Fusion</source>
          <volume>41</volume>
          (
          <year>2018</year>
          )
          <fpage>195</fpage>
          -
          <lpage>216</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>F.</given-names>
            <surname>Juraev</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>El-Sappagh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Abdukhamidov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ali</surname>
          </string-name>
          , T. Abuhmed,
          <article-title>Multilayer dynamic ensemble model for intensive care unit mortality prediction of neonate patients</article-title>
          ,
          <source>Journal of Biomedical Informatics</source>
          <volume>135</volume>
          (
          <year>2022</year>
          )
          <fpage>104216</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>F.</given-names>
            <surname>Pedregosa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Varoquaux</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gramfort</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Michel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Thirion</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Grisel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Blondel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Prettenhofer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Weiss</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Dubourg</surname>
          </string-name>
          , et al.,
          <article-title>Scikit-learn: Machine learning in python</article-title>
          ,
          <source>the Journal of machine Learning research 12</source>
          (
          <year>2011</year>
          )
          <fpage>2825</fpage>
          -
          <lpage>2830</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>R. M. Cruz</surname>
            ,
            <given-names>L. G.</given-names>
          </string-name>
          <string-name>
            <surname>Hafemann</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Sabourin</surname>
            ,
            <given-names>G. D.</given-names>
          </string-name>
          <string-name>
            <surname>Cavalcanti</surname>
          </string-name>
          ,
          <article-title>Deslib: A dynamic ensemble selection library in python</article-title>
          ,
          <source>The Journal of Machine Learning Research</source>
          <volume>21</volume>
          (
          <year>2020</year>
          )
          <fpage>283</fpage>
          -
          <lpage>287</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>K.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Natarajan</surname>
          </string-name>
          ,
          <article-title>Learn to combine modalities in multimodal deep learning</article-title>
          , arXiv preprint arXiv:
          <year>1805</year>
          .
          <volume>11730</volume>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>S.</given-names>
            <surname>Raschka</surname>
          </string-name>
          ,
          <string-name>
            <surname>Mlxtend:</surname>
          </string-name>
          <article-title>Providing machine learning and data science utilities and extensions to python's scientific computing stack</article-title>
          ,
          <source>The Journal of Open Source Software</source>
          <volume>3</volume>
          (
          <year>2018</year>
          ). URL: https://joss.theoj.org/papers/ 10.21105/joss.00638. doi:
          <volume>10</volume>
          .21105/joss.00638.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>S.</given-names>
            <surname>El-Sappagh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ali</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Abuhmed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Alonso</surname>
          </string-name>
          ,
          <article-title>Automatic detection of alzheimer's disease progression: An eficient information fusion approach with heterogeneous ensemble classifiers</article-title>
          ,
          <source>Neurocomputing</source>
          <volume>512</volume>
          (
          <year>2022</year>
          )
          <fpage>203</fpage>
          -
          <lpage>224</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>C. G.</given-names>
            <surname>Snoek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Worring</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. W.</given-names>
            <surname>Smeulders</surname>
          </string-name>
          ,
          <article-title>Early versus late fusion in semantic video analysis</article-title>
          ,
          <source>in: Proceedings of the 13th annual ACM international conference on Multimedia</source>
          ,
          <year>2005</year>
          , pp.
          <fpage>399</fpage>
          -
          <lpage>402</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>F.</given-names>
            <surname>Juraev</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>El-Sappagh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Abuhmed</surname>
          </string-name>
          ,
          <article-title>Explainable dynamic ensemble framework for classification based on the late fusion of heterogeneous multimodal data</article-title>
          ,
          <source>in: Intelligent Systems Conference</source>
          , Springer,
          <year>2023</year>
          , pp.
          <fpage>555</fpage>
          -
          <lpage>570</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>S.</given-names>
            <surname>Ali</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Abuhmed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>El-Sappagh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Muhammad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Alonso-Moral</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Confalonieri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Guidotti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. Del</given-names>
            <surname>Ser</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Díaz-Rodríguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Herrera</surname>
          </string-name>
          ,
          <article-title>Explainable artificial intelligence (xai): What we know and what is left to attain trustworthy artificial intelligence</article-title>
          ,
          <source>Information Fusion</source>
          (
          <year>2023</year>
          )
          <fpage>101805</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>M.</given-names>
            <surname>Sabourin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mitiche</surname>
          </string-name>
          , D. Thomas,
          <string-name>
            <surname>G.</surname>
          </string-name>
          <article-title>Nagy, Classiifer combination for hand-printed digit recognition</article-title>
          ,
          <source>in: Proceedings of 2nd International Conference on Document Analysis and Recognition (ICDAR'93)</source>
          , IEEE,
          <year>1993</year>
          , pp.
          <fpage>163</fpage>
          -
          <lpage>166</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>K.</given-names>
            <surname>Woods</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. P.</given-names>
            <surname>Kegelmeyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Bowyer</surname>
          </string-name>
          ,
          <article-title>Combination of multiple classifiers using local accuracy estimates</article-title>
          ,
          <source>IEEE transactions on pattern analysis and machine intelligence</source>
          <volume>19</volume>
          (
          <year>1997</year>
          )
          <fpage>405</fpage>
          -
          <lpage>410</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>P. C.</given-names>
            <surname>Smits</surname>
          </string-name>
          ,
          <article-title>Multiple classifier systems for supervised remote sensing image classification based on dynamic classifier selection</article-title>
          ,
          <source>IEEE Transactions on Geoscience and Remote Sensing</source>
          <volume>40</volume>
          (
          <year>2002</year>
          )
          <fpage>801</fpage>
          -
          <lpage>813</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>R. G.</given-names>
            <surname>Soares</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Santana</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. M.</given-names>
            <surname>Canuto</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. C. P. de Souto</surname>
          </string-name>
          ,
          <article-title>Using accuracy and diversity to select classifiers to build ensembles</article-title>
          ,
          <source>in: The 2006 IEEE International Joint Conference on Neural Network Proceedings, IEEE</source>
          ,
          <year>2006</year>
          , pp.
          <fpage>1310</fpage>
          -
          <lpage>1316</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <surname>M. C. de Souto</surname>
            ,
            <given-names>R. G.</given-names>
          </string-name>
          <string-name>
            <surname>Soares</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Santana</surname>
            ,
            <given-names>A. M.</given-names>
          </string-name>
          <string-name>
            <surname>Canuto</surname>
          </string-name>
          ,
          <article-title>Empirical comparison of dynamic classifier selection methods based on diversity and accuracy for building ensembles</article-title>
          ,
          <source>in: 2008 IEEE international joint conference on neural networks (IEEE world congress on computational intelligence)</source>
          , IEEE,
          <year>2008</year>
          , pp.
          <fpage>1480</fpage>
          -
          <lpage>1487</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>T.</given-names>
            <surname>Woloszynski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kurzynski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Podsiadlo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. W.</given-names>
            <surname>Stachowiak</surname>
          </string-name>
          ,
          <article-title>A measure of competence based on random classification for dynamic ensemble selection</article-title>
          ,
          <source>Information Fusion</source>
          <volume>13</volume>
          (
          <year>2012</year>
          )
          <fpage>207</fpage>
          -
          <lpage>213</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>P. R.</given-names>
            <surname>Cavalin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Sabourin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. Y.</given-names>
            <surname>Suen</surname>
          </string-name>
          ,
          <article-title>Dynamic selection approaches for multiple classifier systems</article-title>
          ,
          <source>Neural computing and applications 22</source>
          (
          <year>2013</year>
          )
          <fpage>673</fpage>
          -
          <lpage>688</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>R. S.</given-names>
            <surname>Olson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. H.</given-names>
            <surname>Moore</surname>
          </string-name>
          ,
          <article-title>Tpot: A tree-based pipeline optimization tool for automating machine learning</article-title>
          ,
          <source>in: Workshop on automatic machine learning</source>
          ,
          <source>PMLR</source>
          ,
          <year>2016</year>
          , pp.
          <fpage>66</fpage>
          -
          <lpage>74</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>T.</given-names>
            <surname>Head</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Nahrstaedt</surname>
          </string-name>
          , G. Louppe, I. Shcherbatyi, scikit-optimize/scikit-optimize
          <source>: v0. 8</source>
          .1,
          <string-name>
            <surname>Zenodo</surname>
          </string-name>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>T.</given-names>
            <surname>Akiba</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Sano</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Yanase</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Ohta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Koyama</surname>
          </string-name>
          ,
          <article-title>Optuna: A next-generation hyperparameter optimization framework</article-title>
          ,
          <source>in: Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</source>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>J.</given-names>
            <surname>Bergstra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Yamins</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Cox</surname>
          </string-name>
          ,
          <article-title>Making a science of model search: Hyperparameter optimization in hundreds of dimensions for vision architectures</article-title>
          , in: International conference on machine learning,
          <source>PMLR</source>
          ,
          <year>2013</year>
          , pp.
          <fpage>115</fpage>
          -
          <lpage>123</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [32]
          <string-name>
            <given-names>F.</given-names>
            <surname>Nogueira</surname>
          </string-name>
          , et al.,
          <article-title>Bayesian optimization: Open source constrained global optimization tool for python</article-title>
          , URL https://github. com/fmfn/BayesianOptimization (
          <year>2014</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          [33]
          <string-name>
            <surname>T. G.</surname>
          </string-name>
          <article-title>authors, Gpyopt: A bayesian optimization framework in python</article-title>
          , http://github.com/ShefieldML/ GPyOpt,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          [34]
          <string-name>
            <given-names>M.</given-names>
            <surname>Claesen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Simm</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Popovic</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Moreau</surname>
          </string-name>
          , B. De Moor,
          <article-title>Easy hyperparameter search using optunity</article-title>
          ,
          <source>arXiv preprint arXiv:1412.1114</source>
          (
          <year>2014</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          [35]
          <string-name>
            <given-names>S. G.</given-names>
            <surname>Mueller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. W.</given-names>
            <surname>Weiner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. J.</given-names>
            <surname>Thal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. C.</given-names>
            <surname>Petersen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Jack</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Jagust</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. Q.</given-names>
            <surname>Trojanowski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. W.</given-names>
            <surname>Toga</surname>
          </string-name>
          ,
          <string-name>
            <surname>L. Beckett,</surname>
          </string-name>
          <article-title>The alzheimer's disease neuroimaging initiative</article-title>
          ,
          <source>Neuroimaging Clinics of North America</source>
          <volume>15</volume>
          (
          <year>2005</year>
          )
          <fpage>869</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          [36]
          <string-name>
            <given-names>N.</given-names>
            <surname>Rahim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Abuhmed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mirjalili</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>El-Sappagh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Muhammad</surname>
          </string-name>
          ,
          <article-title>Time-series visual explainability for alzheimer's disease progression detection for smart healthcare</article-title>
          ,
          <source>Alexandria Engineering Journal</source>
          <volume>82</volume>
          (
          <year>2023</year>
          )
          <fpage>484</fpage>
          -
          <lpage>502</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          [37]
          <string-name>
            <given-names>A.</given-names>
            <surname>Asuncion</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Newman</surname>
          </string-name>
          ,
          <source>Uci machine learning repository</source>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          [38]
          <string-name>
            <given-names>D. L.</given-names>
            <surname>Beekly</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. M.</given-names>
            <surname>Ramos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. W.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. D.</given-names>
            <surname>Deitrich</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. E.</given-names>
            <surname>Jacka</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. L.</given-names>
            <surname>Hubbard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. D.</given-names>
            <surname>Koepsell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. C.</given-names>
            <surname>Morris</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. A.</given-names>
            <surname>Kukull</surname>
          </string-name>
          , et al.,
          <article-title>The national alzheimer's coordinating center (nacc) database: the uniform data set</article-title>
          ,
          <source>Alzheimer Disease &amp; Associated Disorders</source>
          <volume>21</volume>
          (
          <year>2007</year>
          )
          <fpage>249</fpage>
          -
          <lpage>258</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          [39]
          <string-name>
            <given-names>N.</given-names>
            <surname>Rahim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>El-Sappagh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Rizk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O. A.</given-names>
            <surname>El-serafy</surname>
          </string-name>
          ,
          <source>T. Abuhmed</source>
          ,
          <article-title>Information fusion-based bayesian optimized heterogeneous deep ensemble model based on longitudinal neuroimaging data</article-title>
          ,
          <source>Applied Soft Computing</source>
          <volume>162</volume>
          (
          <year>2024</year>
          )
          <fpage>111749</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref40">
        <mixed-citation>
          [40]
          <string-name>
            <given-names>K.</given-names>
            <surname>Marek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Jennings</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Lasch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Siderowf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Tanner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Simuni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Cofey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kieburtz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Flagg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chowdhury</surname>
          </string-name>
          , et al.,
          <article-title>The parkinson progression marker initiative (ppmi</article-title>
          ),
          <source>Progress in neurobiology 95</source>
          (
          <year>2011</year>
          )
          <fpage>629</fpage>
          -
          <lpage>635</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref41">
        <mixed-citation>
          [41]
          <string-name>
            <given-names>M.</given-names>
            <surname>Junaid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ali</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Eid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>El-Sappagh</surname>
          </string-name>
          ,
          <string-name>
            <surname>T. Abuhmed,</surname>
          </string-name>
          <article-title>Explainable machine learning models based on multimodal time-series data for the early detection of parkinson's disease</article-title>
          ,
          <source>Computer Methods and Programs in Biomedicine</source>
          <volume>234</volume>
          (
          <year>2023</year>
          )
          <fpage>107495</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref42">
        <mixed-citation>
          [42]
          <string-name>
            <given-names>C.</given-names>
            <surname>Sammut</surname>
          </string-name>
          ,
          <string-name>
            <surname>G. I.</surname>
          </string-name>
          Webb (Eds.),
          <source>Holdout Evaluation</source>
          ,
          <string-name>
            <surname>Springer</surname>
            <given-names>US</given-names>
          </string-name>
          , Boston, MA,
          <year>2010</year>
          , pp.
          <fpage>506</fpage>
          -
          <lpage>507</lpage>
          . URL: https://doi.org/10.1007/978-0-
          <fpage>387</fpage>
          -30164-8_
          <fpage>369</fpage>
          . doi:
          <volume>10</volume>
          .1007/978-0-
          <fpage>387</fpage>
          -30164-8_
          <fpage>369</fpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>