<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Journal of Machine Learning Research 9
(2008) 371-421.
[18] A. N. Angelopoulos</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="doi">10.1080/01621459</article-id>
      <title-group>
        <article-title>Uncertainty Considerations of Explainable AI in Data-Driven Systems</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Fatima Rabia Yapicioglu</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science and Engineering, University of Bologna</institution>
          ,
          <addr-line>Bologna</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Marketing and Sales, Automobili Lamborghini S.p.A., Sant'Agata Bolognese</institution>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2016</year>
      </pub-date>
      <volume>105</volume>
      <fpage>1050</fpage>
      <lpage>1059</lpage>
      <abstract>
        <p>Artificial Intelligence (AI) systems are increasingly relied upon in high-stakes domains such as healthcare, finance, and autonomous driving, as well as in high-value commercial applications like luxury automotive design and exclusive financial services, where decision-making must be both accurate and trustworthy. However, the opaque nature of many AI models raises concerns about transparency and accountability, driving the development of Explainable AI (XAI) techniques to foster trust. While these methods aim to improve interpretability, questions persist regarding the reliability and certainty of these explanations, particularly under varying conditions and sources of uncertainty. This underscores the need for robust trust measures to assess the validity and consistency of AI-generated explanations across diferent contexts. Consequently, the question shifts from "Can I trust this model?" to "To what extent can I trust the explanations and the reasoning behind the model's decisions?"-emphasizing the importance of reliable frameworks for explainability. To reliably quantify uncertainty in AI-generated predictions, we integrate conformal prediction, a distribution-free, model-agnostic framework that constructs prediction sets with statistically valid coverage guarantees, ensuring that the true outcome is included with a userspecified probability. By adapting to diferent tasks and data distributions, conformal prediction provides a robust foundation for uncertainty measurement and enables the generation of consistent, uncertainty-aware explanations across varying conditions. We term this approach “uncertainty-aware explanations”, providing systematic methods to assess the trustworthiness of AI insights in diverse contexts, including time series forecasting, classification, and other data-driven tasks. By addressing the relationship between uncertainty and explainability, this work aims to enhance the reliability of AI-driven decision-making in high-stakes environments.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Explainable Artificial Intelligence</kwd>
        <kwd>Uncertainty-Awareness</kwd>
        <kwd>Certainty in Explanations</kwd>
        <kwd>Trustworthy AI</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Context and Motivation</title>
      <p>
        As Artificial Intelligence (AI) systems become increasingly complex, explaining their decisions becomes
more challenging, leading to concerns about trust and reliability. This has prompted the development
of Explainable AI (XAI) to enhance transparency by providing human-interpretable explanations for
AI-driven decisions. XAI methods produce diferent types of explanations depending on the task,
afecting their applicability across domains. For example, a practitioner analyzing ECG signals to assess
a patient’s risk of developing cardiac disease requires retrospective analysis to identify key time intervals
contributing to the prediction (time series forecasting) and to rank other influential factors such as diet,
weight, and physical activity (time series classification). Despite advancements in Explainable AI (XAI),
there is a lack of robust methods to quantify the uncertainty in AI-generated explanations, leading
to potential overreliance on explanations provided in critical domains. As a result, a critical question
arises: Can we trust these explanations, and if so, to what extent? This highlights the need for rigorous
evaluation frameworks to assess the reliability, consistency, and validity of AI-generated explanations
across various contexts [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
      </p>
      <p>
        Uncertainty quantification (UQ) plays a critical role in AI systems, as it provides a measure of
confidence in the model’s predictions, helping practitioners assess the reliability of outputs, especially
in high-stakes applications [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. A promising approach to enhancing the reliability of AI-generated
explanations is conformal prediction, a statistical framework that ofers valid uncertainty quantification
with formal guarantees. In a classification task, conformal prediction provides a prediction set—a
collection of possible labels for a new instance—accompanied by a confidence level. Instead of assigning
a single label, the model ofers a set of labels guaranteed to contain the true label with a specified
probability, such as 90%. This enables practitioners to understand both the model’s most likely prediction
and the associated uncertainty, leading to more informed and reliable decision-making. By producing
prediction sets that adapt to the uncertainty present in the data, conformal prediction ofers
wellcalibrated measures of confidence and paves the way for reliable uncertainty-aware explanations [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
      </p>
      <p>Uncertainty-aware explanations in XAI enhance interpretability by revealing both the reasoning
behind a model’s outputs and the confidence in those outputs. In this work, they are defined as
explanations that capture how variations in uncertainty influence the trustworthiness of predictions.
For example, in financial risk assessment, such an explanation might show that a loan applicant’s risk
score is less reliable due to missing income data, with uncertainty rising by 30% compared to complete
cases. This approach supports practitioners in evaluating the robustness of AI insights and making
informed decisions under uncertainty.</p>
      <p>
        The lack of uncertainty-aware explanations in AI can cause major issues, especially in high-stakes
areas like healthcare and autonomous driving [
        <xref ref-type="bibr" rid="ref4 ref5">4, 5</xref>
        ]. For instance, without uncertainty quantification,
a healthcare model may overestimate a patient’s risk, leading to incorrect diagnoses due to incorrect
reasoning, or an autonomous vehicle may make decisions without knowing the level of uncertainty,
increasing accident risk. By integrating conformal prediction, we can generate explanations that not
only identify key features but also quantify the confidence in these attributions, ensuring more reliable
and transparent AI decision-making across applications.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Background</title>
      <sec id="sec-2-1">
        <title>2.1. Model-Agnostic and Post-hoc Explainability</title>
        <p>
          Post-hoc explainability refers to techniques applied after a machine learning model has been trained,
aiming to interpret its predictions without altering the underlying model structure [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ]. These methods
are particularly valuable in high-stakes domains such as healthcare and finance, where understanding
the rationale behind predictions is crucial for trust and accountability [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]. Post-hoc approaches include
feature importance methods, surrogate models, and visualization techniques, which help uncover the
decision-making process of complex models like deep neural networks or ensemble methods [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ].
        </p>
        <p>
          Model-agnostic explainability, a subset of post-hoc methods, is designed to be applicable to any
machine learning model, regardless of its architecture or complexity [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ]. There are two approaches:
global explainability for overall patterns and local explainability for individual predictions. Techniques
such as LIME (Local Interpretable Model-agnostic Explanations) [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] and SHAP (SHapley Additive
exPlanations) [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] are prominent examples. LIME approximates the behavior of a model locally by
creating interpretable surrogate models, while SHAP leverages game theory to attribute prediction
outcomes to individual features. These methods provide flexibility and transparency, making them
widely adopted in practice.
        </p>
        <p>
          The growing demand for explainability stems from regulatory requirements, such as the European
Union’s General Data Protection Regulation (GDPR), which emphasizes the right to explanation [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ].
Additionally, model-agnostic methods enable practitioners to maintain high predictive performance
while ensuring interpretability, bridging the gap between accuracy and transparency [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ].
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Uncertainty Quantification and Conformal Prediction</title>
        <p>
          Uncertainty quantification (UQ) is a fundamental aspect of machine learning that focuses on measuring
and interpreting the uncertainty associated with model predictions. This is particularly critical in
highstakes applications such as healthcare, autonomous systems, and financial forecasting, where decisions
based on overconfident predictions can lead to severe consequences [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ]. UQ methods aim to provide
probabilistic estimates, confidence intervals, or prediction intervals to convey the reliability of model
outputs. These techniques can be broadly categorized into Bayesian approaches, ensemble methods,
and evidential deep learning [
          <xref ref-type="bibr" rid="ref13 ref14">13, 14</xref>
          ]. For example, Bayesian neural networks quantify uncertainty by
modeling distributions over model parameters, while ensemble methods leverage multiple models to
estimate predictive variance [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ].
        </p>
        <p>Conformal prediction is a model-agnostic, non-parametric framework for uncertainty quantification
(UQ) that provides valid confidence intervals without strong distributional assumptions [ 16]. Relying
on the weaker exchangeability assumption rather than i.i.d., it calibrates prediction sets or intervals
using a hold-out validation set to guarantee user-specified coverage (e.g., 95%) [ 17]. It applies to any
model, including black-box architectures, and has recently been extended to time-series forecasting and
high-dimensional data [18].</p>
        <sec id="sec-2-2-1">
          <title>2.2.1. Mathematical Formulation of Conformal Prediction</title>
          <p>Let  = {(, )}=1 denote a dataset, where  represents the input features and  represents the
corresponding true label or value. Conformal prediction works as follows:
1. Nonconformity Measure: A nonconformity measure (, ) quantifies how unusual a pair
(, ) is with respect to the model’s predictions. For example, in regression, (, ) could be the
absolute residual | − ˆ|, where ˆ is the model’s prediction.
2. Calibration Set: A hold-out calibration set cal = {(, )}=1 is used to compute
nonconformity scores  = (, ) for each point in the calibration set.
3. Prediction Interval Construction: For a new input new, the conformal prediction framework
constructs a prediction interval (new) such that:</p>
          <p>(new) = { : (new, ) ≤  },
where  is the (1 −  )-th quantile of the nonconformity scores {}=1. This ensures that the
interval (new) covers the true label new with probability 1 −  (  is user-specified error rate).</p>
        </sec>
        <sec id="sec-2-2-2">
          <title>2.2.2. Key Metrics in Conformal Prediction</title>
          <p>1. Coverage: Coverage measures the proportion of true labels that fall within the prediction
intervals or sets. For a dataset , the empirical coverage is defined as:</p>
          <p>Coverage = 1 ∑︁ I( ∈ ()),
 =1
(1)
where I(· ) is the indicator function. A well-calibrated conformal prediction framework ensures
that the empirical coverage is approximately 1 −  .
2. Set Size: Set size measures the size of the prediction sets or intervals. For classification tasks,
it is the number of labels in the prediction set, and for regression tasks, it is the width of the
prediction interval. Smaller set sizes indicate more precise predictions, while larger set sizes
reflect higher uncertainty.</p>
        </sec>
        <sec id="sec-2-2-3">
          <title>2.2.3. Integration and Advantages of Conformal Prediction</title>
          <p>Integrating UQ and conformal prediction into machine learning pipelines improves decision-making by
ofering insights into prediction reliability, crucial in safety-critical applications where overconfidence
can lead to catastrophic outcomes [19]. In healthcare, conformal prediction provides condfience intervals
for patient outcomes, aiding clinicians in making informed decisions. In autonomous systems, UQ
assesses prediction reliability in dynamic environments.</p>
          <p>These methods align with the growing focus on transparency and robustness in AI, supported
by regulatory frameworks and industry standards [20]. Conformal prediction ofers valid coverage
guarantees without distribution assumptions, is model-agnostic, and is applicable to various tasks,
including time-series forecasting and high-dimensional data, with recent extensions like split and
adaptive conformal prediction [18].</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Related Work</title>
      <p>Related work linking explainability and conformal prediction remains limited. [21] propose
CONFIDERAI, refining rule-based classifiers by combining conformal prediction with explainable ML for
improved reliability. [22] introduces CONFINE, a framework for interpretable neural networks with
robust uncertainty estimates. [23] explores oracle coaching to generate valid, eficient conformal
classiifers optimized for specific test sets. [ 24] compares frequentist, Bayesian, and conformal uncertainty
estimation, highlighting conformal methods for trustworthy confidence sets in model explanations.
[25] apply XAI to cardiovascular risk prediction in COPD patients, comparing counterfactual methods
and proposing counterfactual conformity for validation. [26] presents a conformal prediction-based
framework for interpreting unsupervised node representations in graphs. Most relevant to this PhD
is Calibrated Explanations (CE) [27], which provides stable, model-agnostic local feature importance
maps with uncertainty quantification via Venn-Abers predictors [ 28]. In contrast, this work uses a
perturbation-based, post-hoc, model-agnostic approach with classical conformal prediction, tailored
to specific tasks and analyzing how predictive uncertainty shifts under varying calibration sets and
systematic noise.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Research Questions and Objectives</title>
      <p>Following research questions (RQ) have been proposed for this research:
1. How can uncertainty in AI-predictions be efectively quantified to enhance trust and reliability
in decision-making?
2. Which evaluation measures are needed to assess the validity/performance of conformal prediction,
and how can we leverage them to produce uncertainty-aware explanations?
3. How do uncertainty-aware explanations generated by our proposed frameworks enhance
decisionmaking and compare to conventional explainers across diverse real-world scenarios?
To address RQ1, we begin by quantifying uncertainty in AI predictions using the flexible framework
of conformal prediction. In classification tasks, this involves prediction sets with varying confidence levels ,
while in regression, it involves prediction intervals around outputs. Additionally, scalar uncertainty
measures—such as variance from ensembles, dropout, input perturbations, or adversarial
modifications—provide adaptable, task-specific metrics [ 29]. This phase identifies the most suitable uncertainty
quantification methods across diferent tasks.</p>
      <p>For RQ2, we extend conformal prediction to evaluate and communicate uncertainty in AI-generated
explanations across tasks like classification , time series forecasting, and clustering. A key property of
conformal prediction is that its coverage, while guaranteed on average (Equation 1), varies with
calibration sets. We will explore how perturbing training or calibration data afects uncertainty and model
performance, aiming to integrate these efects into reliable and transparent explanation frameworks.</p>
      <p>Uncertainty metrics will be task-specific: in classification, we analyze prediction set size and coverage
(Section 2.2.2); in forecasting, we assess changes in confidence interval bounds (Equation 3). Perturbing
input features helps recalibrate models and track shifts in uncertainty. We also explore both local and
global explainability.</p>
      <p>In RQ3, we compare our uncertainty-aware explanation frameworks with SHAP, LIME, Saliency
Maps, and Integrated Gradients. Efectiveness is tested via ablation of the top-ranked feature or segment;
robustness by varying conformal prediction confidence levels; and faithfulness by comparison with
inherently explainable models. These evaluations integrate uncertainty to enhance transparency and
reliability.
This research aims to develop task-specific uncertainty-aware explanations within a conformal
prediction framework. Using a modular approach for classification, forecasting, and other tasks, it
systematically measures and explains feature or segment contributions to predictive uncertainty.</p>
      <sec id="sec-4-1">
        <title>5.1. Approach and Methods</title>
        <p>For each task (e.g., classification, regression, time-series forecasting), the conformal prediction
framework is tailored to produce prediction sets or intervals that quantify uncertainty, ensuring task-specific
validity and interpretability [17].</p>
        <sec id="sec-4-1-1">
          <title>Classification Tasks</title>
          <p>Given an input  and error rate  ∈ (0, 1), the prediction set is:
(2)
(3)
() = { ∈  | (, ) ≥   },
where  is the label set, (, ) is a conformity score, and   is a threshold ensuring coverage of 1 −  .
Regression Tasks For a predicted value ˆ(), the prediction interval is:</p>
          <p>() = [ˆ() −   , ˆ() +   ],
ensuring the interval captures the true value with probability 1 −  .</p>
          <p>We assess feature or segment contributions to uncertainty by applying systematic perturbations and
measuring changes in Coverage and Set-size (Section 2.2.2). For classification , individual features are
perturbed and models recalibrated; for forecasting, PELT segmentation [30] is used, and perturbing
each segment reveals its impact on interval bounds (Equation 3). Comparisons with the unmodified
baseline identify the main sources of predictive uncertainty.</p>
        </sec>
      </sec>
      <sec id="sec-4-2">
        <title>5.2. Rationale and Testing</title>
        <p>The rationale for using conformal prediction for producing uncertainty-aware explanations is its ability
to provide reliable uncertainty estimates independent of data distribution. However, despite its robust
uncertainty quantification, conformal prediction often lacks systematic methods for explaining the
sources of uncertainty and adapting to various tasks.</p>
        <p>We hypothesize that perturbing features or segments and examining their efect on uncertainty
metrics can yield meaningful, interpretable uncertainty-aware explanations. This is validated on diverse
datasets to ensure robustness and generalizability. Efectiveness is assessed through ablation studies,
where significant features or intervals are removed and performance is remeasured. We further evaluate
robustness by varying confidence levels (1 −  ), as defined in Section 5.1, and faithfulness by comparing
our results to explanations from intrinsically interpretable models, verifying that they reflect genuine
uncertainty sources rather than artifacts of the method.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>6. Results and Contributions to Date</title>
      <p>
        In our research, we developed the global explainability framework ConformaSight [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], based on
conformal prediction, to address prediction set-type outputs across classifiers. The framework generates a
global feature importance table, making it easy for non-experts to identify factors afecting conformal
metrics like coverage and set-size. In efectiveness tests, selecting the top 7 features by each explainer
and retraining models resulted in a 0.7% improvement over SHAP and Permutation [31].
      </p>
      <p>
        We contributed to Fast Calibrated Explanations (FCE) [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], which combines ConformaSight’s
perturbation techniques with Calibrated Explanations (CE) [27] to deliver rapid, uncertainty-aware explanations.
Applicable to classification and thresholded regression, FCE provides probabilistic outputs while
preserving uncertainty quantification, achieving up to 19× faster regression, over 75× faster than calibrated
LIME, and 200× faster than calibrated SHAP. Having developed uncertainty-aware explanations for
classification and regression tasks, we are now extending our research to time-series forecasting. We
present ConformaSegment, a segment-based explanation framework that identifies, segments, and
weights time-series intervals by their decision importance. In ablation studies on the most influential
segment, ConformaSegment outperformed Saliency Maps [32] with a 42% average 2 gain and 25.73%
higher prediction interval coverage, and Integrated Gradients [33] with an 18% 2 gain and 40.15%
higher coverage.
      </p>
      <p>In summary, the contributions up to date are as follows:
1. Leveraged conformal prediction to generate uncertainty-aware explanations for tabular
data classification (ConformaSight): We designed a framework to identify which features
contribute most to predictive uncertainty when subjected to significant perturbations, potentially
causing the model to make incorrect predictions. The framework shows how calibration set
perturbations influence prediction set outcomes, highlighting their impact on model performance.
2. Contributed FCE for rapid uncertainty-aware explanations for tabular data classification
and regression: We proposed a method designed for generating faster, uncertainty-aware
explanations by incorporating perturbation techniques from ConformaSight into the core elements
of CE. This method boosts computational eficiency for real-time use while preserving uncertainty
quantification in classification and probabilistic regression.
3. Extended conformal prediction to generate uncertainty-aware explanations for
timeseries forecasting (ConformaSegment): We adapted our framework to time-series forecasting
tasks, focusing on identifying the most critical time segments that contribute to predictive
uncertainty, thereby influencing the accuracy of the forecasted values.</p>
    </sec>
    <sec id="sec-6">
      <title>7. Expected Next Steps and Final Contribution to Knowledge</title>
      <p>This PhD research aims to enhance trust in model decision-making by identifying key factors driving
significant changes in model uncertainty. We explore how conformal prediction, which provides
statistically guaranteed prediction sets with user-specified coverage, can be leveraged to generate
uncertainty-aware explanations. Our goal is to develop a family of post-hoc, model-agnostic frameworks
designed to produce reliable and interpretable explanations while advancing the transparency of
conformal prediction-based explainability methods. Next, we aim to extend these frameworks to
anomaly detection, synthetic data generation, and clustering, advancing transparent and generalizable
uncertainty-aware explainability.</p>
    </sec>
    <sec id="sec-7">
      <title>8. Acknowledgments</title>
      <p>Fatima Rabia is a PhD student at DISI, University of Bologna, funded by PNRR (n. 9990) and Automobili
Lamborghini S.p.A., Italy.</p>
    </sec>
    <sec id="sec-8">
      <title>9. Declaration on Generative AI</title>
      <p>The author has used ChatGPT-4o exclusively for grammar checking and rephrasing; the author originally
produced all content.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>S.</given-names>
            <surname>Ali</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Abuhmed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>El-Sappagh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Muhammad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Alonso-Moral</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Confalonieri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Guidotti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. Del</given-names>
            <surname>Ser</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Díaz-Rodríguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Herrera</surname>
          </string-name>
          ,
          <article-title>Explainable artificial intelligence (xai): What we know and what is left to attain trustworthy artificial intelligence</article-title>
          ,
          <source>Information fusion 99</source>
          (
          <year>2023</year>
          )
          <fpage>101805</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>M.</given-names>
            <surname>Abdar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Pourpanah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hussain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Rezazadegan</surname>
          </string-name>
          , L. Liu,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ghavamzadeh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Fieguth</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Cao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Khosravi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>U. R.</given-names>
            <surname>Acharya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Makarenkov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Nahavandi</surname>
          </string-name>
          ,
          <article-title>A review of uncertainty quantification in deep learning: Techniques, applications</article-title>
          and challenges,
          <source>Information Fusion</source>
          <volume>76</volume>
          (
          <year>2021</year>
          )
          <fpage>243</fpage>
          -
          <lpage>297</lpage>
          . URL: https://www.sciencedirect.com/science/article/pii/S1566253521001081. doi:https://doi. org/10.1016/j.inffus.
          <year>2021</year>
          .
          <volume>05</volume>
          .008.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q. V.</given-names>
            <surname>Liao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. K.</given-names>
            <surname>Bellamy</surname>
          </string-name>
          ,
          <article-title>Efect of confidence and explanation on accuracy and trust calibration in ai-assisted decision making</article-title>
          ,
          <source>in: Proceedings of the 2020 conference on fairness, accountability, and transparency</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>295</fpage>
          -
          <lpage>305</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>F. R.</given-names>
            <surname>Yapicioglu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Stramiglio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Vitali</surname>
          </string-name>
          , Conformasight:
          <article-title>Conformal prediction-based global and model-agnostic explainability framework</article-title>
          ,
          <source>in: World Conference on Explainable Artificial Intelligence</source>
          , Springer,
          <year>2024</year>
          , pp.
          <fpage>270</fpage>
          -
          <lpage>293</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>T.</given-names>
            <surname>Löfström</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. R.</given-names>
            <surname>Yapicioglu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Stramiglio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Löfström</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Vitali</surname>
          </string-name>
          ,
          <article-title>Fast calibrated explanations: Eficient and uncertainty-aware explanations for machine learning models</article-title>
          ,
          <source>arXiv preprint arXiv:2410.21129</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>C.</given-names>
            <surname>Molnar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Casalicchio</surname>
          </string-name>
          ,
          <string-name>
            <surname>B. Bischl,</surname>
          </string-name>
          <article-title>Interpretable machine learning-a brief history, state-of-the-art and challenges</article-title>
          ,
          <source>in: Joint European conference on machine learning and knowledge discovery in databases</source>
          , Springer,
          <year>2020</year>
          , pp.
          <fpage>417</fpage>
          -
          <lpage>431</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>R.</given-names>
            <surname>Guidotti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Monreale</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ruggieri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Turini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Giannotti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Pedreschi</surname>
          </string-name>
          ,
          <article-title>A survey of methods for explaining black box models, ACM computing surveys (CSUR) 51 (</article-title>
          <year>2018</year>
          )
          <fpage>1</fpage>
          -
          <lpage>42</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>M. T.</given-names>
            <surname>Ribeiro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Guestrin</surname>
          </string-name>
          ,
          <article-title>" why should i trust you?" explaining the predictions of any classifier</article-title>
          ,
          <source>in: Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining</source>
          ,
          <year>2016</year>
          , pp.
          <fpage>1135</fpage>
          -
          <lpage>1144</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>S.</given-names>
            <surname>Lundberg</surname>
          </string-name>
          ,
          <article-title>A unified approach to interpreting model predictions</article-title>
          ,
          <source>arXiv preprint arXiv:1705.07874</source>
          (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>B.</given-names>
            <surname>Goodman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Flaxman</surname>
          </string-name>
          ,
          <article-title>European union regulations on algorithmic decision-making and a “right to explanation”</article-title>
          ,
          <source>AI</source>
          magazine
          <volume>38</volume>
          (
          <year>2017</year>
          )
          <fpage>50</fpage>
          -
          <lpage>57</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>A. B. Arrieta</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Díaz-Rodríguez</surname>
            ,
            <given-names>J. Del</given-names>
          </string-name>
          <string-name>
            <surname>Ser</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Bennetot</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Tabik</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Barbado</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>García</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Gil-López</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Molina</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Benjamins</surname>
          </string-name>
          , et al.,
          <article-title>Explainable artificial intelligence (xai): Concepts, taxonomies, opportunities and challenges toward responsible ai</article-title>
          ,
          <source>Information fusion 58</source>
          (
          <year>2020</year>
          )
          <fpage>82</fpage>
          -
          <lpage>115</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>R. C.</given-names>
            <surname>Smith</surname>
          </string-name>
          , Uncertainty Quantification: Theory, Implementation, and
          <string-name>
            <surname>Applications</surname>
          </string-name>
          , SIAM,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>A.</given-names>
            <surname>Kendall</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Gal</surname>
          </string-name>
          ,
          <article-title>What uncertainties do we need in bayesian deep learning for computer vision</article-title>
          ?,
          <source>Advances in Neural Information Processing Systems</source>
          <volume>30</volume>
          (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>B.</given-names>
            <surname>Lakshminarayanan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Pritzel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Blundell</surname>
          </string-name>
          ,
          <article-title>Simple and scalable predictive uncertainty estimation using deep ensembles</article-title>
          ,
          <source>Advances in Neural Information Processing Systems</source>
          <volume>30</volume>
          (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Gal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Ghahramani</surname>
          </string-name>
          ,
          <article-title>Dropout as a bayesian approximation: Representing model uncertainty in</article-title>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>