<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>T h1e Dynamics of Explainability: Diverse Insights from SHAP Explanations using Neighbourhoods</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Urja Pawar</string-name>
          <email>urja.pawar@mycit.ie</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ruairi O'Reilly</string-name>
          <email>Ruairi.OReilly@mtu.ie</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Christian Beder</string-name>
          <email>Christian.Beder@mtu.ie</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Donna O'Shea</string-name>
          <email>Donna.OShea@mtu.ie</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Munster Technological University</institution>
          ,
          <addr-line>Cork</addr-line>
          ,
          <country country="IE">Ireland</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>This paper presents a discussion on utilising a dashboard tool to enhance the interpretability of SHapley Additive exPlanations (SHAP) in healthcare Artificial Intelligence(AI)-based applications. Despite SHAP's potential to demystify an AI model's decisions, interpreting SHAP values remains challenging, especially when considering diferent data neighbourhoods [ 1]. This issue is particularly critical in healthcare, where decision-making requires high precision and clarity. We demonstrate three use cases that can efectively demonstrate the utility of interactive neighborhood exploration. The first compares SHAP explanations in two similar patient neighbourhoods with diferent classifications, ofering unique insights into features that influence classification changes. The second use case focuses on “feature freezing” which isolates certain features to better understand their impact. This can enable highlighting diagnostic tests considered important by a Machine Learning (ML) model for a specific population of patients (e.g., patients of the same age). The final use case demonstrates the relationship between suficient features for a given classification and the importance ranking by SHAP.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Explainable AI</kwd>
        <kwd>Neighbourhoods</kwd>
        <kwd>SHAP explanations</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        The field of Explainable AI (XAI) is dedicated to enhancing the transparency and trustworthiness
of AI systems [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. One state-of-the-art XAI framework is SHAP, which highlights the influence
of dataset features on AI model predictions by assigning importance scores(SHAP values)
and is based on a solid theoretical foundation [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. However, interpreting SHAP values is
challenging due to their dependence on the sample sets or “neighbourhoods" used during
analysis, complicating their application in the precision-critical field of healthcare [
        <xref ref-type="bibr" rid="ref1 ref4 ref5">4, 5, 1</xref>
        ].
      </p>
      <p>In SHAP’s analysis, diferent samples are constructed by mixing the feature values from the
input sample(to be explained) and the neighbourhood samples. Since diferent neighbourhoods
produce varying explanations, a layer of complexity and ambiguity is introduced when
interpreting SHAP values. When only one of these varying explanations is presented in a standalone
manner, it can lead to an inconsistent interpretation of the model’s behaviour. Furthermore,
the assigned importance scores by SHAP are relative to feature values in the training data,
complicating their interpretation against a fixed baseline. Presenting multiple explanations
collectively to users enables understanding how distinct feature values influence the model’s
decisions, enhancing insights into the model’s behaviour.</p>
      <p>
        Previous studies have explored the impacts of neighbourhoods on SHAP’s explanations [
        <xref ref-type="bibr" rid="ref4 ref6">4, 6</xref>
        ].
However, fewer studies have focused on providing varying explanations by exploring diferent
neighbourhoods and interpreting these explanations in the context of a use case. There is an
emerging need for an interactive exploration of these neighbourhoods, enabling users to adjust
various parameters such as distance metrics, handle data imbalance, and allow feature value
ranges in a neighbourhood. Additionally, fundamental explanation blocks, such as suficient
feature sets, provide specific insight regarding an ML model. Suficient sets of features can
preserve a classification regardless of the values of other features.
      </p>
      <p>To address the challenges of clarity and consistency in interpreting SHAP values across
neighbourhoods, this work presents a discussion for an interactive analysis of AI
decisionmaking processes. Our tool enables users to explore how diferent types of samples impact
the explanations and their interpretation. This holistic interpretation will facilitate a deeper
understanding of model decisions, enhancing SHAP-driven insights. To demonstrate the utility
of the proposed approach, we focused on the following three insights:
• Similar vs diferent patient records: SHAP values using similar and diferently
classiifed patient records are compared to highlight key influences on classification confidence.
• Feature Freezing: SHAP values are analysed when certain features are held constant in
a neighbourhood for understanding feature significance within specific patient groups.
• Suficient Sets: The relationship between suficient feature sets and their importance
rankings across various neighbourhoods is analysed.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        SHAP [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] calculates the impact of features on model predictions by averaging the changes when
a feature is added or removed, using training set values for replacing “removed" features. This
approach forms a ‘neighbourhood’ around an input, , where SHAP analyses the perturbed
samples formed by combining attributes from  and training data.
      </p>
      <p>
        Authors in [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] demonstrated how XAI frameworks - SHAP and LIME can be fooled to generate
FI scores that do not reflect the actual learned representation of AI/ML models. The ML models
might be biased, but if specific neighbourhoods based on adversarial examples are used, the
explanations generated can be used to hide the biases in models.
      </p>
      <p>
        Various frameworks have been designed to improve the transparency of ML models via
interactive exploration. For instance, “modelStudio"[
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], “explAIner" [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], and “InterpretML"[
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]
provided an interactive environment that allows users to explore and understand predictions
through various explanations. Complementing these tools, “The what-if tool" [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] focused on
interactive probing of models, providing users with the means to test hypothetical scenarios
and observe how model predictions change with varying input data for understanding model
sensitivity and decision boundaries. However, none of the dashboards explores the
neighbourhoods used by XAI frameworks to understand the context of model predictions, which is crucial
for evaluating the robustness and applicability of explanations.
      </p>
      <p>
        To explore the impact of various neighbourhoods on the explanations produced by surrogate
XAI frameworks such as LIME, [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] developed an interactive dashboard - Explain-Explore.
However, no exploratory tool has been proposed in the literature to interpret SHAP’s explanations
with respect to distinct neighbourhoods. In [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], a comprehensive analysis of the impact of
neighbourhoods on SHAP and LIME was conducted. This work presents an extension on [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] by
enabling an interactive analysis of neighbourhoods with SHAP.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <p>This proposed work is implemented with the Dash1 and Plotly2 libraries in Python. The
subsequent sections outline the methodology for building this tool, detailing the types of
neighbourhoods and the use cases for utility analysis.</p>
      <p>The workflow of our dashboard creation starts with the selection and pre-processing of the
required dataset, followed by model training. SHAP values for the model’s predictions are
generated by analysing an input sample  and the trained classifier. Initially, the neighbourhood
for calculating the values is derived from the training dataset. The dashboard also presents
the classification output and the associated confidence score. A critical step is identifying
minimal subsets of features that are suficient for determining the model’s predictions. A
detailed description of these sets is provided in Section 3.3.</p>
      <p>
        As shown in Figure 1, the proposed tool plots SHAP values and displays confidence scores
and suficient feature sets. Users can modify input sample values, observe feature importance
and confidence score changes, and choose neighbourhood settings. Updates in neighbourhood
settings re-calibrate SHAP calculations with updated scores. Neighbourhood information (e.g.,
mean feature values) is accessible via hover-over tooltips on the bar charts. The time complexity
for generating SHAP explanations is directly proportional to the number of features in the
dataset. In future work, we will explore optimisation techniques that can be used with SHAP
for generating explanations faster[
        <xref ref-type="bibr" rid="ref11">11</xref>
        ].
      </p>
      <sec id="sec-3-1">
        <title>3.1. Dataset and Classifier</title>
        <p>
          In this study, we used the Heart Disease dataset from the UCI repository [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] for heart disease
prediction using individual health records. This dataset was selected to demonstrate diferent
1https://dash.plotly.com/
2https://plotly.com/python/
use cases about the medical features in tabular datasets and includes 14 attributes related
to heart disease, spanning initial and detailed medical tests, with 297 records where 160 are
classified as negative and 137 as positive for heart disease. Features such as age, sex, types
of angina, blood sugar, blood pressure, and cholesterol levels, which are standard initial tests
before more advanced examinations, are included [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]. In one of our use cases, we freeze these
standard features in a neighbourhood to showcase how SHAP’s revised rankings enhance model
understanding. The presented approach is extensible to any other tabular medical dataset.
        </p>
        <p>For the model, a Support Vector Machine (SVM) classifier was employed due to its robustness
in handling both linear and non-linear classifications suited for the complex patterns in heart
disease data. The classifier achieved an accuracy of 86.67% using 5-fold cross-validation.
However, the primary focus of this work is not on model accuracy but on the application of SHAP, a
model-agnostic framework. This emphasis allows us to concentrate on the interpretability and
utility of SHAP explanations, regardless of the model’s accuracy metrics.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Neighbourhoods</title>
        <p>As outlined earlier, SHAP constructs contexts by generating perturbed samples by randomly
replacing original input feature values with those from neighbourhood samples. The
standard version of SHAP uses training data and is referred to as “standard". Further explored
neighbourhoods are described below, based on the sample’s classifications:
1. Balanced/Skewed: Balanced/skewed distributions of classified perturbed samples
2. Similar versus Diferent Classifications : Composing samples that either share the
same classification as the input (referred to as ) or difer from it ( )
3. Using Mahalanobis Distance for Locality: Neighborhood samples based on
Mahalanobis distance metric
4. Freezing the Features: Specific features can be fixed across samples to have the same
value as the input sample to isolate and understand the efects of other varying features.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Utility Analysis</title>
        <p>This section describes use cases to demonstrate how multiple insights can enhance clinicians’
understanding of an AI/ML model.</p>
        <p>Similar vs Diferent Patients: SHAP explanations for patient records under two
neighbourhoods - outside and inside, are compared to the standard and balanced settings. We discuss a
use-case using a specific patient record and then provide Kendall correlation between feature
rankings in standard SHAP versus diferent neighbourhoods to demonstrate the dissimilarity.
Presenting diferent rankings, we discuss how diferent types of samples afect SHAP values
and how to interpret them with respect to the classification results.</p>
        <p>
          Feature Freezing: Specific neighbourhoods are constructed, keeping some features constant
across all samples. This enables examining how particular features impact groups of patients
who share characteristics like age or gender. By keeping the basic set of features constant,
clinicians can better understand how specific advanced tests influence the outcome. Kendall
correlation between the two rankings of variable features - with frozen neighbourhood and
with standard SHAP - are provided to demonstrate the dissimilarity in their rankings.
Suficient Sets: The relationship between suficient feature sets and high-ranked features by
SHAP is studied. The suficient sets of features are subsets that keep the classification as it is in
a neighbourhood even when other features were varied [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ]. The percentage of times, when the
top-3 ranked features by SHAP under diferent neighbourhoods, are also the suficient subset
of features is presented. As SHAP values can be positive/negative, the results are categorised
based on whether these top-3 features are negatively scored, positively scored, or overall top-3
features. The code to reproduce results is available on Github3.
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results</title>
      <p>To demonstrate the utility of our dashboard, specific patient records are considered and the
SHAP plots are shown for discussions.</p>
      <sec id="sec-4-1">
        <title>4.1. Similar vs Diferent Patient Records</title>
        <p>We analysed SHAP values for a patient record (detailed in Table 1) under four settings:
standard, balanced, inside, and outside. Figure 2a shows the standard SHAP values for classifying
the patient as healthy ( () = 0). Figure 2b illustrates the values in a balanced neighbourhood,
and Figures 2c and 2d display the values using the outside and inside settings, respectively.</p>
        <p>The figures highlight variations in feature rankings across diferent settings, each providing
a distinct interpretation. This analysis focuses on the impact of two specific features, major
vessels and fasting blood sugar, on classification probabilities. Changing the number of major
vessels from 1 to 0 increases classification probability from 0.72 to 0.90 for class 0, whereas
a change from 1 to 2 alters the classification. This shows how the number of major vessels
influences classification. The mean value of major vessels in training data is 1. Still, the standard
SHAP values in Figure 2a do not explain why there is a positive contribution associated with
major vessels equal to 1. Conversely, the mean value of major vessels in “outside" is 2, making
it easy to interpret that an increase in the value alters (negative impact shown in Figure 2c) the
classification. This is also evident in Figure 2b, however, with less strength due to the combined
efect from samples of both classes. For fasting blood sugar, altering the value from 0 to 1
improves the prediction probability to 0.87 for class 0. This significant efect is captured in
“inside" (Figure 2d) as the mean value of fasting blood sugar in this neighborhood is 1.</p>
        <p>Kendall correlation coeficients reveal discrepancies in SHAP rankings across diferent
neighbourhood settings. Specifically, the correlation between standard and balanced neighbourhood
settings is 0.669, between standard and outside settings is 0.694, and between standard and
inside settings is 0.671. These variations highlight how diferent neighbourhood contexts yield
diferent insights that can be used for clearer interpretations. In the context of healthcare, these
insights can significantly enhance patient care. By understanding how specific changes in
3Link to the code - https://github.com/UrjaPawar/shap-dash
(a) Standard SHAP scores</p>
        <p>(b) SHAP scores with “balanced"
(c) SHAP scores with “outside"
(d) SHAP scores with “inside"
5A0g.0e 0S.e0x A0T.yn0pgiicnaal A0A.n0tygpinicaal 1BR2Pe0s.0ting 0SB.u0logoadr 0ER.C0esGt t2Ce1hr9oo.0lles - A1PN.an0oignni-nal
Table 3
Patient record selected for use-case: feature freezing</p>
        <p>A patient record, which includes basic medical history and basic diagnostic tests like age,
sex, presence of typical or atypical angina, fasting blood sugar, blood pressure, and cholesterol
levels, is selected for analysis as described in Table 3. We keep these values constant across a
local neighbourhood to assess the impact on the rankings of the remaining features.</p>
        <p>Figure 3a shows the standard SHAP values, while Figure 3b displays the values when basic
features are held constant (“frozen"). There’s a noticeable shift: Thalassemia, an advanced
medical test, gains more importance when basic features are frozen compared to its lesser
significance in standard settings. The complex interactions between features may obscure
Thalassemia’s impact in standard settings compared to when the basic medical features are
frozen, highlighting the significance of the test.</p>
        <p>The Kendall correlation between the rankings of “unfrozen" features in standard and frozen
settings was found to be 70%, indicating notable dissimilarity. This use case can enable clinicians
to clearly understand and amplify the relevance of diagnostic tests like Thalassemia with respect
to specific patients.</p>
        <p>(a) SHAP Scores with standard settings
(b) SHAP Scores with Frozen Features</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.3. Suficient Feature Sets</title>
        <p>An experiment was conducted to observe and associate the suficient feature sets with SHAP
feature rankings under diferent neighbourhood settings to note how many of the top-3 features
form a suficient set to maintain a given classification. As shown in Table 4, the standard SHAP
scores could identify suficient sets in only 46% of the patient records. However, with 
settings, SHAP’s analysis focuses more on patients with a diferent classification and identifies
suficient sets in 86% of the patients when the top-3 overall features are considered. This shows
that to identify features suficient to maintain a classification, SHAP should be provided with
more examples from diferent classifications to highlight impactful features with respect to the
current classification.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusions</title>
      <p>This paper presented an interactive approach to improve the interpretability of SHAP in
healthcare by exploring the impact of diferent neighbourhoods. The utility of this dashboard is
demonstrated through three specific use cases, showing its efectiveness in providing detailed
insights into SHAP explanations. Future work will focus on expanding the dashboard’s
functionalities and improving its user interface. It will include adding more diverse XAI techniques
and exploring further neighbourhood definitions to encompass a more comprehensive array of
clinical scenarios. We also plan to make the tool available as a pip package and conduct user
studies with healthcare professionals to refine and enhance the dashboard’s features based on
real-world feedback.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>U.</given-names>
            <surname>Pawar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Beder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Ruairi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Donna</surname>
          </string-name>
          , et al.,
          <article-title>On the impact of neighbourhood sampling to satisfy suficiency and necessity criteria in explainable ai</article-title>
          ,
          <source>in: Causal Learning and Reasoning</source>
          , PMLR,
          <year>2024</year>
          , pp.
          <fpage>570</fpage>
          -
          <lpage>586</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>H.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Lundberg</surname>
          </string-name>
          ,
          <string-name>
            <surname>S.-I. Lee</surname>
          </string-name>
          ,
          <article-title>Explaining models by propagating shapley values of local components</article-title>
          , arXiv preprint arXiv:
          <year>1911</year>
          .
          <volume>11888</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>S. M.</given-names>
            <surname>Lundberg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.-I.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <article-title>A unified approach to interpreting model predictions</article-title>
          , in: I. Guyon,
          <string-name>
            <given-names>U. V.</given-names>
            <surname>Luxburg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Bengio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wallach</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Fergus</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Vishwanathan</surname>
          </string-name>
          , R. Garnett (Eds.),
          <source>Advances in Neural Information Processing Systems</source>
          <volume>30</volume>
          ,
          <string-name>
            <surname>Curran</surname>
            <given-names>Associates</given-names>
          </string-name>
          , Inc.,
          <year>2017</year>
          , pp.
          <fpage>4765</fpage>
          -
          <lpage>4774</lpage>
          . URL: http://papers.nips.cc/paper/ 7062-a
          <article-title>-unified-approach-to-interpreting-model-predictions</article-title>
          .pdf.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>D.</given-names>
            <surname>Slack</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hilgard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Jia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Lakkaraju</surname>
          </string-name>
          ,
          <article-title>Fooling lime and shap: Adversarial attacks on post hoc explanation methods</article-title>
          ,
          <source>in: Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>180</fpage>
          -
          <lpage>186</lpage>
          . URL: https://doi.org/10.1145/3375627.3375830. doi:
          <volume>10</volume>
          .1145/3375627.3375830.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>I. E.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Venkatasubramanian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Scheidegger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Friedler</surname>
          </string-name>
          ,
          <article-title>Problems with shapleyvalue-based explanations as feature importance measures</article-title>
          ,
          <source>in: International Conference on Machine Learning, PMLR</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>5491</fpage>
          -
          <lpage>5500</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>D.</given-names>
            <surname>Collaris</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. J. van Wijk</surname>
          </string-name>
          ,
          <article-title>Explainexplore: Visual exploration of machine learning explanations</article-title>
          ,
          <source>in: 2020 IEEE Pacific Visualization Symposium (PacificVis)</source>
          , IEEE,
          <year>2020</year>
          , pp.
          <fpage>26</fpage>
          -
          <lpage>35</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>H.</given-names>
            <surname>Baniecki</surname>
          </string-name>
          , P. Biecek,
          <article-title>modelstudio: Interactive studio with explanations for ml predictive models</article-title>
          ,
          <source>Journal of Open Source Software</source>
          <volume>4</volume>
          (
          <year>2019</year>
          )
          <fpage>1798</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>T.</given-names>
            <surname>Spinner</surname>
          </string-name>
          , U. Schlegel,
          <string-name>
            <given-names>H.</given-names>
            <surname>Schäfer</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>El-Assady, explainer: A visual analytics framework for interactive and explainable machine learning</article-title>
          ,
          <source>IEEE transactions on visualization and computer graphics 26</source>
          (
          <year>2019</year>
          )
          <fpage>1064</fpage>
          -
          <lpage>1074</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>H.</given-names>
            <surname>Nori</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Jenkins</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Koch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Caruana</surname>
          </string-name>
          ,
          <article-title>Interpretml: A unified framework for machine learning interpretability</article-title>
          , arXiv e-prints (
          <year>2019</year>
          ) arXiv-
          <fpage>1909</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>J.</given-names>
            <surname>Wexler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Pushkarna</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Bolukbasi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wattenberg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Viégas</surname>
          </string-name>
          , J. Wilson,
          <article-title>The what-if tool: Interactive probing of machine learning models</article-title>
          ,
          <source>IEEE transactions on visualization and computer graphics 26</source>
          (
          <year>2019</year>
          )
          <fpage>56</fpage>
          -
          <lpage>65</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>N.</given-names>
            <surname>Jethani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sudarshan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I. C.</given-names>
            <surname>Covert</surname>
          </string-name>
          ,
          <string-name>
            <surname>S.-I. Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ranganath</surname>
          </string-name>
          , Fastshap:
          <article-title>Real-time shapley value estimation</article-title>
          ,
          <source>in: International Conference on Learning Representations</source>
          ,
          <year>2021</year>
          . URL: https://doi.org/10.48550/arXiv.2107.07436.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>P.</given-names>
            <surname>Robert Detrano</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.D.</given-names>
            ,
            <surname>Uci</surname>
          </string-name>
          .
          <year>2010</year>
          .
          <article-title>v. a. medical center, long beach and cleveland clinic foundation (</article-title>
          <year>1988</year>
          ). URL: https://archive.ics.uci.edu/ml/datasets/heart+disease.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>U.</given-names>
            <surname>Pawar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. T.</given-names>
            <surname>Culbert</surname>
          </string-name>
          ,
          <string-name>
            <surname>R.</surname>
          </string-name>
          <article-title>O'Reilly, Evaluating hierarchical medical workflows using feature importance</article-title>
          ,
          <source>in: 2021 IEEE 34th International Symposium on Computer-Based Medical Systems (CBMS)</source>
          , IEEE,
          <year>2021</year>
          , pp.
          <fpage>265</fpage>
          -
          <lpage>270</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>S.</given-names>
            <surname>Galhotra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Pradhan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Salimi</surname>
          </string-name>
          ,
          <article-title>Explaining black-box algorithms using probabilistic contrastive counterfactuals</article-title>
          ,
          <source>in: Proceedings of the 2021 International Conference on Management of Data</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>577</fpage>
          -
          <lpage>590</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>