<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Intelligence: July</journal-title>
      </journal-title-group>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Novel Approach for Benchmarking Local Binary Classification XAI Methods Using Synthetic Ground Truth Datasets</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Karim Moustafa</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Workshop</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>The Artificial Intelligence and Cognitive Load Research Lab, The Centre of Explainable Artificial Intelligence</institution>
          ,
          <addr-line>Technological</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>University Dublin</institution>
          ,
          <addr-line>Dublin</addr-line>
          ,
          <country>Republic of Ireland</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <volume>0</volume>
      <fpage>9</fpage>
      <lpage>11</lpage>
      <abstract>
        <p>Evaluating Explainable AI (XAI) methods, particularly local feature attributions on tabular data, is hindered by a lack of standardised benchmarks and objective metrics. This research introduces a novel technique to address this gap. We propose a methodology for generating synthetic ground-truth datasets with known featuretarget relationships, enabling objective evaluation. The proposed solution includes the Explainability Fidelity Assessor (XFA) to quantify interpretation faithfulness (completeness, conciseness, sensitivity) and the Optimum Explainability Score (OPS) to assess objective explanation quality (simplicity, completeness). Preliminary results using LIME and Anchor demonstrate the ability to compare XAI methods objectively. This work contributes to the creation of XAI groundtruth datasets and the development of novel evaluation metrics (XFA, OPS), aiming to enhance the reliability and comparability of local XAI methods.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Context and Motivation</title>
      <p>
        Machine learning (ML) models are increasingly deployed in critical domains, yet complex models often
function as black boxes, hindering understanding of their decision-making processes. Explainable
Artificial Intelligence (XAI) aims to address this by providing insights into model behaviour [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
Understanding how a model makes a prediction is crucial for debugging, building trust, ensuring fairness,
promoting accountability, and enabling user interaction [
        <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
        ]. For instance, in healthcare, knowing
which patient features drive a high-risk prediction is vital for clinical decision-making [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>
        Despite the proliferation of XAI methods, evaluating their efectiveness remains a significant challenge
[
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. A key problem is the lack of standardised benchmark datasets with known ground truth, particularly
for evaluating local explanations (explanations for individual predictions). Without such datasets,
comparing diferent XAI techniques objectively is dificult. Furthermore, there’s no universally accepted
definition or framework for assessing explanation quality, leading to inconsistency and concerns
about generalisability [
        <xref ref-type="bibr" rid="ref5 ref6">5, 6</xref>
        ]. The field often conflates inherent model
interpretability with post-hoc
explainability as in [
        <xref ref-type="bibr" rid="ref7 ref8">7, 8</xref>
        ]. It sometimes relies on subjective human evaluations [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], which, however, are
costly, dificult to scale, and prone to bias [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
      </p>
      <p>This research tackles these challenges by proposing a novel technique for evaluating local feature
attributions of XAI methods applied to tabular, cross-sectional data for binary classification. The core
contributions are: (1) The introduction of the first synthetic ground-truth datasets specifically designed
for local XAI evaluation, providing an objective basis for assessment. (2) A novel evaluation framework
comprising the Explainability Fidelity Assessor (XFA) for measuring faithfulness to the ground truth,
and the Optimum Explainability Score (OPS) for assessing objective explanation characteristics like
simplicity and completeness. By providing standardised datasets and objective metrics, this work aims
to improve the reliability, comparability, and trustworthiness of XAI methods.</p>
      <p>Late-breaking work, Demos and Doctoral Consortium, colocated with The 3rd World Conference on eXplainable Artificial</p>
      <p>CEUR</p>
      <p>ceur-ws.org</p>
    </sec>
    <sec id="sec-2">
      <title>2. Key Related Work</title>
      <p>
        The need for transparency in complex ML models has driven the development of numerous XAI
techniques [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. These methods can be categorised along several dimensions:
• Applicability: Model-specific (tailored to specific architectures, e.g., activation analysis in CNNs
[
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]) versus model-agnostic (treat model as black-box, e.g., LIME [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], SHAP [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], Anchor [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]).
      </p>
      <p>
        This work focuses on model-agnostic methods due to their flexibility.
• Scope: Global (overall model behaviour) versus local (individual predictions). This research
targets local explanations explicitly.
• Stage: Intrinsic (inherently simple models like decision trees [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]) versus post-hoc (techniques
applied after training complex models). Our focus is on evaluating post-hoc methods for black-box
models.
• Output Form: Explanations can be feature importance scores, rules, counterfactuals,
visualisations [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. This work primarily targets methods that produce feature importance scores or
similar quantifiable outputs.
      </p>
      <p>
        Evaluating XAI methods is crucial but challenging [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Existing evaluation approaches often fall into
categories proposed by Doshi-Velez and Kim [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]:
• Human-grounded: Uses user studies to assess understanding, trust, satisfaction [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Pro: Direct
human insight. Con: Subjective, costly, hard to scale/generalise.
• Application-grounded: Evaluates XAI impact on task performance in a specific domain [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ].
      </p>
      <p>
        Pro: Realistic context. Con: It is hard to isolate XAI impact; findings may not generalise.
• Function-grounded: Uses proxy metrics like fidelity, stability, robustness [
        <xref ref-type="bibr" rid="ref18 ref19">18, 19</xref>
        ]. Pro: Objective,
automatable. Con: Metrics may not reflect true human interpretability.
      </p>
      <p>
        Recent eforts have included developing benchmark datasets, primarily for NLP, based on human
annotations [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ], which inherit subjectivity and lack generalizability to other domains, such as tabular
data. Significant gaps remain in XAI evaluation [
        <xref ref-type="bibr" rid="ref21 ref22">21, 22</xref>
        ]. There is:
• Lack of Consensus and Generic Metrics: No agreed-upon definition of ”good” explanation or
standard metrics applicable across diferent XAI methods [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
• Confusion between Interpretability and Explainability: Over-reliance on subjective human
evaluation obscures the need for objective assessment of explanation fidelity.
• Limited Comparability: Dificulty in comparing diferent XAI algorithms (macro-level) or
tuning parameters within an algorithm (micro-level) due to a lack of standard benchmarks.
• Lack of Holistic Local Evaluation: No unified way to aggregate performance across diferent
local instances and evaluation aspects.
      </p>
      <p>This research directly addresses the need for objective, generic metrics and standardised ground-truth
benchmarks for local feature attributions of XAI methods on tabular data.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Research Questions and Objectives</title>
      <p>This research aims to develop an objective benchmarking technique for evaluating local feature
attribution explanations methods for binary classification on tabular cross-sectional data. The central research
question is:</p>
      <p>Can an XAI evaluation benchmarking technique—based on (1) a synthetic ground-truth dataset, (2)
explainability fidelity assessment, and (3) optimum local explainability assessment—efectively evaluate
the feature attribution performance of local XAI methods for binary classification?</p>
    </sec>
    <sec id="sec-4">
      <title>4. Research Approach, Methods, and Rationale</title>
      <p>To address the identified gaps and research questions, we propose a novel evaluation technique focused
on objective assessment using synthetic ground truth. The proposed solution workflow, contrasted
with typical literature approaches, is shown in Figure 1. It separates the generation of ground truth
from the evaluation of fidelity (XFA) and objective explanation characteristics (OPS).</p>
      <sec id="sec-4-1">
        <title>4.1. Synthetic Ground-truth Datasets Creation</title>
        <p>Rationale: The lack of objective ground truth is a major impediment to XAI evaluation. Synthetic
datasets allow us to precisely control the underlying feature-target relationships. Method: We generate
datasets where the binary target variable is determined by predefined ’Relationship Rules’ (RRs) applied
to specific subsets of features. These rules encompass linear and non-linear regression-based functions,
as well as distance-based clustering structures. Table 2 shows the summary of the RRs used in dataset
generation.</p>
        <p>The process involves:
1. Dataset Pool Creation: Generate datasets covering combinations of 2-12 RRs. Each dataset (10k
instances, 48 features) uses multivariate normal distribution features. Instances are stratified, and
labels are generated per stratum using the assigned RR. Crucially, instance-level metadata (the
ground truth specifying which features and rules generated the label) is stored.
2. Dataset Selection: Train a standard ML model (e.g., RandomForest) on each dataset of the 4082
datasets created in the previous step, as a (80/20 split). Select 11 datasets representing the 0th to
100th percentiles of model performance (e.g., AUC), ensuring a range of complexities (Table 1).</p>
        <p>These selected datasets form the benchmark suite.</p>
        <p>This provides the first benchmark suite with known instance-level ground truth for local XAI evaluation
on tabular data.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Explainability Fidelity Assessor (XFA)</title>
        <p>Rationale: To objectively measure how faithfully an XAI method’s explanation reflects the true
underlying reasons (ground truth) for a prediction. Method: XFA compares the features identified as important
by an XAI method (for a given instance) against the ground truth features from the dataset metadata
for that instance (Figure 2). It computes three instance-level metrics, which are then averaged across
the test set:
• Completeness (IF-Recall): Percentage of ground truth features correctly identified by the XAI
method. (High recall = finds most true features).
• Conciseness (1 - IF-FPR): Inverse of the False Positive Rate. Percentage of features identified by</p>
        <p>XAI that are actually in the ground truth. (High conciseness = identifies fewer incorrect features).
• Sensitivity (IF-Sen): Agreement between the rank order of feature importance given by XAI
and the rank order in the ground truth (if applicable, e.g., based on coeficients).</p>
        <p>The final XFA Score aggregates these metrics, weighted by dataset complexity, providing a single score
(0-100) for overall fidelity as shown in (Eq. 1).</p>
        <p>∗ 
 ],</p>
        <p>-
 ∗   
  
_  =
∑  ([1 −  
-  
 ,  
-  )
(1)</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Optimum Explainability Score (OPS)</title>
        <p>Rationale: To evaluate objective properties of the explanation itself, namely its simplicity and its
coverage (completeness of application). Aim is to find the ”sweet spot” between being informative and
being concise. Method: OPS considers two factors:
• Simplicity (  ): Average number of features (or explanation elements) included in the local
explanations across the dataset. Lower is simpler.</p>
      </sec>
      <sec id="sec-4-4">
        <title>4.4. Overall Score and Validation</title>
        <p>The Overall Score combines XFA and OPS (normalised) via a weighted average, providing a holistic
evaluation as shown in (Eq.3 ), with user-adjustable weights ( 1,  2) enabling prioritisation of specific
evaluation aspects.</p>
        <p>= 1 ( 1    + ( 2  −100 ∗ 100)) (3)</p>
        <p>2</p>
        <p>Validation Rationale: To ensure the proposed metrics (XFA, OPS) behave as expected and are sensitive
to changes in explanation quality. Validation Method:
• XFA Validation: Systematically introduce controlled errors (missing features, extra features,
incorrect ranking) into the ground truth metadata at varying levels (0-100%). Assess if the XFA
score and its components show a monotonic decrease with increasing error levels and a strong
association (correlation) with the error level.
• OPS Validation: Generate explanation sets with varying controlled levels of complexity (  )
and coverage (  ). Assess if the OPS score correctly ranks these explanation sets and correlates
with the deviation from the ideal (0 complexity, 100% coverage).</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Preliminary Results and Contributions to Date</title>
      <p>Progress has been achieved in developing the proposed XAI evaluation benchmarking framework:
• A comprehensive literature review identified critical gaps in current XAI evaluation approaches.
• The methodology for generating synthetic ground truth datasets was designed and implemented,
resulting in a benchmark suite of 11 datasets with varying, controllable complexity levels and
known instance-level ground truth.
• The Explainability Fidelity Assessor (XFA) component was designed and developed. Its core
properties (e.g., monotonicity, association with controlled error) were validated using the protocol
outlined in Section 4, confirming its suitability for objective fidelity assessment.
• The Optimum Explainability Score (OPS) component for evaluating objective explanation
characteristics (simplicity, coverage) was designed, and its underlying mathematical properties
(convexity) were established.</p>
      <p>
        Demonstrating Framework Utility: XFA Use Case To illustrate the practical application and
capabilities of our framework, we conducted an initial case study using the XFA component and the
synthetic datasets. This involved comparing two widely recognised local, model-agnostic explanation
methods: LIME [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] and Anchor [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. The process involved generating explanations from both methods
for instances in our benchmark datasets and evaluating their fidelity using XFA against the known
ground truth (see Figure 4 for the workflow).
      </p>
      <p>Qualitative Insights from the Use Case: This preliminary application demonstrated that the XFA
framework successfully enables:
• Objective Quantitative Comparison: The framework provided distinct quantitative fidelity
scores for LIME and Anchor based on metrics assessing completeness (finding true features),
conciseness (avoiding false positives), and sensitivity (ranking features correctly) versus ground
truth.
• Analysis of Fidelity Profiles: The comparison highlighted nuanced diferences in how these
methods perform. For example, it allows for the analysis of trade-ofs, such as potential diferences
in achieving high recall versus maintaining high conciseness in the generated explanations.
• Assessment of Complexity Impact: The experiments using datasets of varying complexity
indicated that the fidelity of explanations can be influenced by the complexity of the underlying
data-generating process, and the framework allows for observing these efects.
• Highlighting Areas for Improvement: The overall assessment suggested that while current
methods provide valuable insights, there is measurable variance in their fidelity and potential
scope for enhancing the faithfulness of local explanations produced by these methods.</p>
      <p>These initial results validate the core components of our framework and demonstrate its potential to
provide objective, nuanced, and comparable evaluations of local XAI methods, addressing a significant
gap identified in the literature. Further experiments involving the OPS component and additional XAI
methods are planned as outlined in Section 6.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Expected Next Research Steps and Final Contribution</title>
      <p>The next steps focus on completing the evaluation of the benchmarking framework and expanding its
application:
• OPS Module Evaluation: Rigorously evaluate the OPS metric using the validation protocol (Sec
4), assessing its sensitivity and ranking ability using simulated explanation sets and applying it to
outputs from LIME, Anchor, and potentially SHAP.
• Extended XAI Method Experiments: Apply the full framework (Datasets, XFA, OPS, Overall
Score) to evaluate additional state-of-the-art local XAI methods (e.g., SHAP) to provide a broader
comparative analysis.
• Publications: Disseminate the findings related to the synthetic datasets, XFA, and OPS through
publications in relevant peer-reviewed journals and conferences.
• Thesis Completion: Structure and write the PhD thesis, incorporating the benchmarking
framework design, evaluations, experiments, use cases, and conclusions.</p>
      <p>The expected final contributions of this research are significant:
1. A Clear Distinction between Explanation Fidelity and Presentation: By focusing on
objective fidelity assessment separate from user perception, we provide a foundation for evaluating
the core accuracy of XAI methods.
2. Novel Evaluation benchmarking technique for Tabular Data: The first benchmarking
technique specifically for local, feature attribution XAI on tabular binary classification, addressing
key limitations of prior work.
3. Standardised Synthetic Ground Truth Datasets: A publicly available benchmark suite
enabling objective, reproducible evaluation and comparison of local XAI methods.
4. Explainability Fidelity Assessor (XFA): A novel, quantitative, multi-faceted metric
(completeness, conciseness, sensitivity) for assessing explanation faithfulness against ground truth.
5. Optimum Explainability Score (OPS): A novel metric quantifying objective explanation
characteristics (simplicity, coverage) to guide tuning towards optimal explainability.
6. Advancement of XAI Evaluation Methodology: Providing a systematic approach that
incorporates objective metrics, enables comparisons, and addresses the challenges of evaluating local
explanations, ultimately fostering the development of more robust, reliable, and comparable XAI
methods.</p>
      <p>While this work focuses on tabular binary classification, the underlying principles of using synthetic
ground truth and objective metrics could potentially be adapted for other tasks (e.g., regression,
multiclass) in future research.</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author used Gemeni in order to: Grammar and spelling check.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>D.</given-names>
            <surname>Gunning</surname>
          </string-name>
          ,
          <source>Explainable artificial intelligence (xai)</source>
          ,
          <source>Defense Advanced Research Projects Agency (DARPA)</source>
          ,
          <source>nd Web 2</source>
          (
          <year>2017</year>
          )
          <article-title>2</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>S.</given-names>
            <surname>Mohseni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Zarei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. D.</given-names>
            <surname>Ragan</surname>
          </string-name>
          ,
          <article-title>Multifaceted interpretability of machine learning models: A survey, ACM Computing Surveys (CSUR) 55 (</article-title>
          <year>2023</year>
          )
          <fpage>1</fpage>
          -
          <lpage>38</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>F.</given-names>
            <surname>Doshi-Velez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <article-title>Towards a rigorous science of interpretable machine learning</article-title>
          ,
          <source>arXiv preprint arXiv:1702.08608</source>
          (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>C.</given-names>
            <surname>Rudin</surname>
          </string-name>
          ,
          <article-title>Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead</article-title>
          ,
          <source>Nature Machine Intelligence</source>
          <volume>1</volume>
          (
          <year>2019</year>
          )
          <fpage>206</fpage>
          -
          <lpage>215</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Z. C.</given-names>
            <surname>Lipton</surname>
          </string-name>
          ,
          <article-title>The mythos of model interpretability</article-title>
          ,
          <source>Queue</source>
          <volume>16</volume>
          (
          <year>2018</year>
          )
          <fpage>30</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>T.</given-names>
            <surname>Miller</surname>
          </string-name>
          ,
          <article-title>Explanation in artificial intelligence: Insights from the social sciences</article-title>
          ,
          <source>Artificial intelligence 267</source>
          (
          <year>2019</year>
          )
          <fpage>1</fpage>
          -
          <lpage>38</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>L.</given-names>
            <surname>Rizzo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Longo</surname>
          </string-name>
          ,
          <article-title>A qualitative investigation of the explainability of defeasible argumentation and non-monotonic fuzzy reasoning</article-title>
          ,
          <source>in: Proc. of the 26th AIAI Irish Conference on Artificial Intelligence and Cognitive Science</source>
          , Dublin, Ireland,
          <year>2018</year>
          , pp.
          <fpage>138</fpage>
          -
          <lpage>149</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>G.</given-names>
            <surname>Vilone</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Rizzo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Longo</surname>
          </string-name>
          ,
          <article-title>A comparative analysis of rule-based, model-agnostic methods for explainable artificial intelligence</article-title>
          ,
          <source>in: Proc. of The 28th Irish Conference on Artificial Intelligence and Cognitive Science</source>
          , Dublin, Republic of Ireland, volume
          <volume>2771</volume>
          , CEUR-WS.org,
          <year>2020</year>
          , pp.
          <fpage>85</fpage>
          -
          <lpage>96</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>G.</given-names>
            <surname>Vilone</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Longo</surname>
          </string-name>
          ,
          <article-title>Development of a human-centred psychometric test for the evaluation of explanations produced by xai methods</article-title>
          , in: L.
          <string-name>
            <surname>Longo</surname>
          </string-name>
          (Ed.),
          <source>Explainable Artificial Intelligence</source>
          , Springer Nature Switzerland, Cham,
          <year>2023</year>
          , pp.
          <fpage>205</fpage>
          -
          <lpage>232</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>W.</given-names>
            <surname>Samek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Montavon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Vedaldi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. K.</given-names>
            <surname>Hansen</surname>
          </string-name>
          ,
          <string-name>
            <surname>K.-R. Müller</surname>
          </string-name>
          ,
          <article-title>Explainable ai: interpreting, explaining and visualizing deep learning</article-title>
          , Springer (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>M. D. Zeiler</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Fergus</surname>
          </string-name>
          ,
          <article-title>Visualizing and understanding convolutional networks</article-title>
          ,
          <source>European conference on computer vision</source>
          (
          <year>2014</year>
          )
          <fpage>818</fpage>
          -
          <lpage>833</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>M. T.</given-names>
            <surname>Ribeiro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Guestrin</surname>
          </string-name>
          ,
          <article-title>Why should i trust you? explaining the predictions of any classifier</article-title>
          ,
          <source>in: Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining</source>
          ,
          <year>2016</year>
          , pp.
          <fpage>1135</fpage>
          -
          <lpage>1144</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>S. M.</given-names>
            <surname>Lundberg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.-I.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <article-title>A unified approach to interpreting model predictions</article-title>
          ,
          <source>in: Advances in neural information processing systems</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>4765</fpage>
          -
          <lpage>4774</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>M. T.</given-names>
            <surname>Ribeiro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Guestrin</surname>
          </string-name>
          , Anchors: High-Precision
          <string-name>
            <surname>Model-Agnostic</surname>
            <given-names>Explanations</given-names>
          </string-name>
          ,
          <source>Proc. of 32nd Conference on Artificial Intelligence (AAAI)</source>
          (
          <year>2018</year>
          )
          <fpage>1527</fpage>
          -
          <lpage>1535</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>C.</given-names>
            <surname>Rudin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Semenova</surname>
          </string-name>
          ,
          <string-name>
            <surname>C. Zhong,</surname>
          </string-name>
          <article-title>Interpretable machine learning: Fundamental principles and 10 grand challenges</article-title>
          ,
          <source>Statistic Surveys</source>
          <volume>16</volume>
          (
          <year>2022</year>
          )
          <fpage>1</fpage>
          -
          <lpage>85</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>G.</given-names>
            <surname>Vilone</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Longo</surname>
          </string-name>
          ,
          <article-title>Classification of explainable artificial intelligence methods through their output formats</article-title>
          ,
          <source>Machine Learning and Knowledge Extraction</source>
          <volume>3</volume>
          (
          <year>2021</year>
          )
          <fpage>615</fpage>
          -
          <lpage>661</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>D. V.</given-names>
            <surname>Carvalho</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. M.</given-names>
            <surname>Pereira</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. S.</given-names>
            <surname>Cardoso</surname>
          </string-name>
          ,
          <article-title>Machine learning interpretability: A survey on methods and metrics</article-title>
          ,
          <source>Electronics</source>
          <volume>8</volume>
          (
          <year>2019</year>
          )
          <fpage>832</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>S.</given-names>
            <surname>Hooker</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Erhan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.-J.</given-names>
            <surname>Kindermans</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <article-title>A benchmark for interpretability methods in deep neural networks</article-title>
          ,
          <source>Advances in Neural Information Processing Systems</source>
          <volume>32</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>W.</given-names>
            <surname>Samek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Binder</surname>
          </string-name>
          , G. Montavon,
          <string-name>
            <given-names>S.</given-names>
            <surname>Lapuschkin</surname>
          </string-name>
          ,
          <string-name>
            <surname>K.-R. Müller</surname>
          </string-name>
          ,
          <article-title>Evaluating the visualization of what a deep neural network has learned</article-title>
          ,
          <source>IEEE transactions on neural networks and learning systems 28</source>
          (
          <year>2016</year>
          )
          <fpage>2660</fpage>
          -
          <lpage>2673</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>J. DeYoung</surname>
            , S. Jain,
            <given-names>N. F.</given-names>
          </string-name>
          <string-name>
            <surname>Rajani</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <string-name>
            <surname>Lehman</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Xiong</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Socher</surname>
            ,
            <given-names>B. C.</given-names>
          </string-name>
          <string-name>
            <surname>Wallace</surname>
          </string-name>
          ,
          <article-title>Eraser: A benchmark to evaluate rationalized nlp models</article-title>
          ,
          <source>in: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>7114</fpage>
          -
          <lpage>7129</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>A.</given-names>
            <surname>Adadi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Berrada</surname>
          </string-name>
          ,
          <article-title>Peeking inside the black-box: A survey on explainable artificial intelligence (xai)</article-title>
          ,
          <source>IEEE access 6</source>
          (
          <year>2018</year>
          )
          <fpage>52138</fpage>
          -
          <lpage>52160</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>W.</given-names>
            <surname>Saeed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Omlin</surname>
          </string-name>
          ,
          <article-title>Explainable ai (xai): A systematic meta-survey of current challenges and future opportunities</article-title>
          ,
          <source>Knowledge-Based Systems</source>
          <volume>263</volume>
          (
          <year>2023</year>
          )
          <fpage>110273</fpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>