<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>From XAI to XEE: eXplainable End-to-End using Influence and Provenance</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Paolo Missier</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Riccardo Torlone</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Pasquale Leonardo Lazzaro</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Luca Gregori</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Clyde Fernandes</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>TianYang Xie</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>RomaTre University</institution>
          ,
          <addr-line>Rome</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>University of Birmingham</institution>
          ,
          <addr-line>Edgbaston</addr-line>
          ,
          <country country="UK">UK</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2026</year>
      </pub-date>
      <abstract>
        <p>It is generally accepted that Machine Learning and AI models can be made “trustworthy” by providing explanations that justify their predictions, and multiple techniques have been proposed to achieve this. These are collectively referred to as model-based explanations, and the kind of models that can be explained using those are referred to as XAI. There is, however, also a complementary need to provide data explanations that account for the dataset selection and data engineering steps that precede model training. Data explanations can be obtained by capturing and then examining the provenance of the data in a training set. In this position paper, we suggest that there is value in combining these two types of explanations, producing an “eXplainable End-to-End” (XEE) backward path that originates from model predictions and traces all the way back to raw data, using a combination of XAI and provenance methods. We demonstrate how this can be achieved in practice using Influence Functions, a well-known XAI technique, in combination with PROLIT, a provenance-based tool that we have been developing to provide data explanations. We provide a simple but representative example and suggest the next steps for more in-depth research into XEE.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Data-Centric AI</kwd>
        <kwd>Explainable AI (XAI)</kwd>
        <kwd>Data Provenance</kwd>
        <kwd>Influence Functions</kwd>
        <kwd>End-to-End Explainability</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Multiple methods are available to achieve transparency and explainability of predictions made
by Machine Learning and AI models [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], including highly nonlinear, complex models such as
Deep Neural Networks, which are generally regarded as “black boxes” from an interpretability
perspective. In such complex settings, a specific model outcome can be “explained” by providing
examples of data points in the training set, which have been the most influential contributors to
the outcome. We rely on Influence Functions [
        <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
        ] to retrieve training points that contribute to
specific predictions. These are summarized in Section 2.
      </p>
      <p>In this paper, we explore and illustrate the idea of combining model explanations with data
explanations to provide an end-to-end account of a model’s prediction. More precisely, consider
a training set that was constructed from a collection of initial datasets that are processed using
a workflow consisting of data transformation operators.</p>
      <p>The data explanation of data point  in the training set is ’s provenance, for which we adopt
the W3C PROV model (http://www.w3.org/TR/prov-dm/) and its standard graph representation</p>
      <p>Such end-to-end explanations are useful in many practical settings. For instance, a data
preparation pipeline may include data transformations that may afect model performance
by accidentally injecting out-of-distribution data points or by altering the set through data
augmentation or other forms of synthetic data injection. In such cases, it is important to identify
the influential data points and trace them back through the pipeline to understand where they
come from.</p>
      <p>Given a model ℎ : R → R trained on , model explanations and data explanations can be
described in abstract terms using two functions. The first is a function  that maps the outcome
ˆ = ℎ(x) for input x to a set of influential training data points in the training set :
ℐ(ˆ) = {1 . . . },  ∈</p>
      <sec id="sec-1-1">
        <title>The second returns the provenance prov () of a training point .</title>
        <p>Conceptually, combining model and data explanations involves composing these two
functions to return a unified explanation as a set of provenance graphs. We denote such composition
as XEE ℎ (eXplanations End to End relative to model ℎ):</p>
        <p>XEE ℎ(x) = {prov ()| ∈ ℐ(ℎ(x))}</p>
        <p>
          The main contribution of this paper is a demonstration of how such composition may be
realised in a practical setting and a discussion of how the idea can be extended to more general
settings. Specifically: (i) we have adapted the Tracin algorithm [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ], available through the
Influenciae python library https://github.com/deel-ai/influenciae and originally designed
to operate on images and Keras-based models, to work on training sets that consist of tabular
records; (ii) we use the PROLIT provenance generator [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] to track, retrieve, and describe the
provenance of the influential training records; (iii) we demonstrate the connection between
the two using the well-known Titanic dataset in combination with a simple data preparation
pipeline and a binary classifier implemented as a simple feed-forward neural network.
        </p>
        <p>The rest of the paper is organized as follows. In Section 2 we provide an overview of model
explanations based on influence functions. In Section 3 we discuss data explanation describing
PROLIT, a system we have developed for this purpose. In Section 4 we illustrate how model
and data explanations can be combined and finally, in Section 5 we draw some conclusions and
discuss future research on XEE.</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>2. Model explanations using influence functions</title>
      <p>
        Research into methods that make “black box models” interpretable is not new, with notable
model-agnostic algorithms like LIME (Local Interpretable Model-Agnostic Explanations) dating
back from 2016 [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. LIME is a local explanation technique that approximates the behavior of a
complex model around a specific instance by locally fitting a simpler model around that instance.
It generates explanations by perturbing the input data and observing changes in predictions.
The resulting surrogate model highlights the contribution of each feature to the prediction for
that instance. LIME is particularly useful for understanding individual predictions, but it does
not provide global insights into the behavior of the model [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>
        Similarly, SHapley Additive exPlanations (SHAP) uses concepts from cooperative game theory
to assign Shapley values to features, quantifying their contribution to a model’s predictions.
Unlike LIME, SHAP provides both local and global explanations. It is particularly efective for
tree-based models and neural networks but can be computationally expensive for large datasets
or highly complex models [
        <xref ref-type="bibr" rid="ref7 ref8">7, 8</xref>
        ].
      </p>
      <p>
        The idea of using Influence Functions to support black-box explanations originates around
the same time [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], but uses a distinctly diferent approach, aiming to trace a model’s prediction
through its learning algorithm and back to the training data. In practice, this entails answering
the counterfactual question: how would the model’s predictions change if a particular point
was not included in the training set? This could be tested in theory by training the model of
the entire training set, then selectively removing a single training point, retraining the model
using the remainder of the set, and finally comparing the vectors of (optimised) weights in
the two scenarios. The computational cost of this naive approach is prohibitive. However,
it has been shown [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] that Influence Functions (IF) can be used to achieve practically good
approximations of the actual relevance of each training data point while removing the need
for retraining altogether. The key idea is that IF makes it possible to observe changes in the
model’s parameters as one single training point is “upweighted” by an infinitesimal amount. In
practice, this amounts to “diferentiating through the training” to estimate, in closed form, the
efect of training perturbations.
      </p>
      <p>A complete formalisation of the problem is beyond the scope of this description. Intuitively,
the idea is to estimate the change in the parameters  ∈ Θ of the model due to removing a
single point  from the training set. As we know, the training process calculates the optimal
values for :
1 ∑︁ (, )
ˆ= arg min ∈Θ  =1
where (, ) is the loss function and  1 . . .  are the training points.</p>
      <p>The change in parameters can be expressed as ˆ− − ,ˆwhere
ˆ 1 ∑︁ (, ).
 − = arg min∈Θ 
̸=</p>
      <p>Influence functions remove the need to repeatedly retrain the model for each  in order to
calculate ˆ− . The idea is to slightly upweight  in the loss function by a small perturbation  ,
giving new parameters:
ˆ 1 ∑︁ (, ) + (, ).</p>
      <p>
        , = arg min∈Θ  =1
Results from influence functions theory show that this can be achieved by calculating the Hessian
of (, ) , which does not require retraining but is itself computationally expensive. However,
the computational challenge of this calculation can be addressed using known approximation
techniques, making the method usable in practice (see [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] for full details).
      </p>
      <p>
        These findings paved the way for a number of variants of the methods, for instance [
        <xref ref-type="bibr" rid="ref10 ref11">10, 11, 12</xref>
        ],
many of which are described in a recent survey [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. For the purpose of our experiments we
have used one of these, the Tracin algorithm [13], which is available as a Python library.1
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Data Explanations using provenance</title>
      <p>The provenance of tabular data that flows through a data science pipeline in preparation for use
by machine learning algorithms is a graph representation of the sequence of transformations that
alter both their schema and values. The PROLIT system [14] is designed to collect, store, manage,
and query provenance information such as data flows through data preparation pipelines. It
also leverages a Large Language Model (LLM) to allow users to query the provenance with
questions expressed in natural language and to return short provenance narratives describing
in words the operations performed on the data.</p>
      <p>PROLIT relies on a provenance model that extends PROV [15] to capture provenance at
diferent levels of granularity within a table, including atomic value, row, column, and whole
table. This model, illustrated in Figure 1, represents provenance data with a graph structure
involving Entity nodes (representing data values), Column nodes (representing entire columns
of a dataset), and Activity nodes (representing operations on the data). Relationships between
these nodes capture actions like data consumption (used), creation (GeneratedBy), removal
(InvalidatedBy), and derivation (DerivedFrom) during data transformations. The belongsTo
relationship links individual cells to their respective columns whereas the next relationship
explicitly sequences activities.</p>
      <p>PROLIT relies on a LLM in two ways. Firstly, it recognises individual operations performed
on the data from an arbitrary Pandas program and creates a structured representation of
their transformations, rewriting user-defined pipelines into a standardised format. This code
segmentation allows PROLIT to precisely associate provenance to each code snippet. Secondly, it
generates human-readable narratives of both the code snippets and of elements of the provenance
graph in response to a user question.</p>
      <p>More in detail, PROLIT takes a Python script that manipulates Pandas dataframes as input.
It first rewrites the script, using a LLM to identify and document the activities. The PROLIT
1https://deel-ai.github.io/influenciae/
provenance generator then observes the execution of the rewritten script, producing a
provenance graph fragment that captures input/output data dependencies for every activity. The
complete provenance graph at the end of execution is written into a graph database (Neo4j),
where it can be analysed using either native Neo4j access or through a LLM-based component
that translates natural language questions into Cypher queries and then renders the resulting
graph as a natural language narrative.</p>
      <p>An example of provenance captured by PROLIT for a pipeline that first eliminates from a
dataset the Purchased column, which acts as the target of a classification model, and then
imputes null values occurring in the Age column is reported in Figure 2.</p>
      <p>PROLIT addresses the problem of the excessive volume of provenance data that can be
generated when tracking individual elements of a large dataset by enabling customization of
the level of granularity to which the provenance is collected and queried. In particular, it can
provide a sketch of the collected provenance, involving a few instances for each operation in the
pipeline, and a view of the provenance at schema-level, showing only operations and involved
columns, hiding their efect on individual data.</p>
    </sec>
    <sec id="sec-4">
      <title>4. XEE: a proof-of-concept end-to-end example</title>
      <p>We now illustrate how model and data explanations can be combined, with a simple example
based using the well-known Titanic tabular dataset, here denoted simply . Firstly we describe
 and the script  used to prepare  for model training. The model itself is a simple Multilayer
Perceptron and is not shown in the detail. Then we perform a model prediction for a random
test data point, and show the ranked list of influential records used in the prediction produced by
Tracin. Finally, we query PROLIT with the top-ranking such record and display its provenance.</p>
      <p>
        More in detail:
1.  is processed using script . This results in training set ′ = (), derived from 
through the operators in ;
2. PROLIT is used to collect the provenance graph ProvDS ′ () of each element  in ′ by
observing the processing;
3. ′ is split into training and test sets ′, ′ and ′ is used to train a simple
neural network, producing a logistic regression classifier ℎ where ˆ = ℎ() ∈ {0, 1}
predicts whether or not a Titanic passengers survives.
4. The Tracin library, which implements the XAI approach described in [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], is used to
generate explanations ℐ(ℎ()) = {1 . . . }, where  ∈ ′, for a specific instance
 ∈ ′. Note that for the purpose of this exercise, we have adapted the reference
Tracin implementation referenced above, which is designed to work on image data, to
return records in a table as explanations.
5. Finally, PROLIT is used to retrieve the provenance ProvDS ′ () for each  ∈ ℐ(ℎ()).
      </p>
      <p>As discussed above, each of the provenance records is rendered using natural language
narratives.</p>
      <sec id="sec-4-1">
        <title>The Titanic dataset  comprises 12 attributes, shown in Fig.3.</title>
        <p>The script , reported in Figure 4, operates on  by performing three operations.
1. Drops unnecessary columns such as “Name”, “Ticket”, and “Cabin” which do not yield
any useful information for the logistic regression model.
2. Imputes the mode for the “Embarked” column and the median for the “Age” column to
handle missing values.
3. Performs one-hot encoding on categorical features “Pclass”, “Sex”, and “Embarked”.</p>
        <p>
          This pipeline prepares the data to allow us to train our model, and as we are tracking
provenance, we are able to investigate the data before it has been altered. Tracin runs alongside
the training process as described in [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]. After training, when we run model inference on a
random test data point (passenger ID 436), Tracin returns a ranked list of most influential
training points (IDs 234, 775, and 262).
        </p>
        <p>We should emphasise that validating influence, that is, understanding why these records
have been ranked as highly influential, is an interesting question that is, however, beyond the
scope of this example. In general, it would be interesting to study consistency across diferent
influence-based methods, which are themeselves non-deterministic. In this example, one may
speculate that the three survivors are “similar” to passenger 436 based on either their age, or
sex.</p>
        <p>We can now use PROLIT to retrieve the provenance of those passenger records, which will
have been recorded when  is executed, focusing on one of the influential passengers (ID
234). PROLIT records the provenance of a record or single attribute, only when changes occur
anywhere in the record. In this example, the state of the record for passenger 234 changes
twice, firstly due to imputation, and then due to the one-hot encoding. These state changes are
reflected in the provenance graph for this record, as shown in Fig.6. The top part of the figure
shows the state of the record before and after each of the changes, while the bottom part shows
the corresponding provenance derivations (read from the right “back” to the past on the left).
Note that no other entities appear in the provenance record, because no other attribute values
have changed.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusions and future works</title>
      <p>In this paper, we have argued for the importance of combining model-based explanations (XAI)
with data explanations to create an eXplainable End-to-End (XEE) framework. By integrating
techniques such as Influence Functions for model interpretability and a system that collects
data provenance for pre-processing pipelines, we have demonstrated the feasibility of tracing
predictions back to their raw data origins. This approach not only enhances the transparency
Record state
Record Provenance</p>
      <p>234
Age:NaN
Sex: f
…
Age: NaN
usage</p>
      <p>gen</p>
      <p>Fillna()
{ arg: median(age) }
234
Age:5
Sex: f
…
Age:5
Sex: F
usage</p>
      <p>Encode gen
Age:5
Sex_m:0</p>
      <p>…
Age:5
Sex_f: 1
Sex_m: 0
of machine learning models but also provides an improved understanding of the entire
data-toprediction pipeline, fostering a better explainability of AI systems.</p>
      <p>We have presented a practical example of how “black box” explanations of model inference,
obtained using Influence Functions, can be efectively combined with data explanations obtained
by capturing and then querying the provenance of training data points through a pre-processing
pipeline. The supporting example relies on the assumption that model explanations consist of
training data points. This assumption is necessary so that the two explanations can be combined,
however, it rules out alternative model explanation methods like SHAP, as this returns a ranking
of attributes in order of relevance, for the predicted data point itself. This assumption is not
too strict, given the demonstrated applicability of Influence Functions, however further insight
into model explainability may be required to properly generalise the method and ensure that
provenance can be safely composed with XAI methods, leading to full XEE.</p>
      <p>There are indeed several possible directions of future research aimed at developing the XEE
framework, including the following: (i) the current idea relies on Influence Functions and
PROLIT, which may face scalability challenges with large datasets or complex models — future
work could explore more eficient algorithms or approximations to make XEE applicable to
real-world, large-scale systems; (ii) while Influence Functions are efective, they represent only
one of many XAI methods — investigating how other XAI techniques, such as SHAP, LIME,
or counterfactual explanations, can be integrated into the XEE framework would broaden
its applicability and robustness; (iii) the efectiveness of XEE should be evaluated not only
from a technical perspective but also in terms of its usability and interpretability by end-users,
including domain experts and non-technical stakeholders — user studies could provide insights
into how XEE explanations can be exploited in practice and tailored to diferent audiences;
(iv) the current framework focuses on structured, tabular data — extending XEE to handle
unstructured data, such as text, images, or time-series data, would expand its scope and utility.</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <sec id="sec-6-1">
        <title>The authors have not employed any Generative AI tools.</title>
        <p>39th International Conference on Machine Learning, PMLR, 2022, pp. 13468–13504. URL:
https://proceedings.mlr.press/v162/lin22h.html, iSSN: 2640-3498.
[12] D. Liu, P. Cheng, H. Zhu, X. Tang, Y. Chen, X. Wang, W. Pan, Z. Ming, X. He, Diwift:
Discovering instance-wise influential features for tabular data, in: Proceedings of the
ACM Web Conference 2023, WWW ’23, Association for Computing Machinery, New York,
NY, USA, 2023, p. 1673–1682. URL: https://doi.org/10.1145/3543507.3583382. doi:10.1145/
3543507.3583382.
[13] A. Picard, L. Hervier, T. Fel, D. Vigouroux, Influenciae: A library for tracing the influence
back to the data-points, 2024.
[14] P. L. Lazzaro, M. Lazzaro, P. Missier, R. Torlone, PROLIT: Supporting the Transparency
of Data Preparation Pipelines through Narratives over Data Provenance, in:
Proceedings of the 28th International Conference on Extending Database Technology (EDBT),
OpenProceedings.org, Barcelona, Spain, 2025.
[15] P. Missier, K. Belhajjame, J. Cheney, The W3C PROV family of specifications for modelling
provenance metadata, in: Procs. EDBT’13 (Tutorial), ACM, Genova, Italy, 2013. URL:
http://www.edbt.org/Proceedings/2013-Genova/papers/edbt/a80-missier.pdf.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>G.</given-names>
            <surname>Schwalbe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Finzel</surname>
          </string-name>
          ,
          <article-title>A comprehensive taxonomy for explainable artificial intelligence: a systematic survey of surveys on methods and concepts</article-title>
          ,
          <source>Data Mining and Knowledge Discovery</source>
          <volume>38</volume>
          (
          <year>2024</year>
          )
          <fpage>3043</fpage>
          -
          <lpage>3101</lpage>
          . URL: https://doi.org/10.1007/s10618-022-00867-8. doi:
          <volume>10</volume>
          . 1007/s10618-022-00867-8.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>R. D.</given-names>
            <surname>Cook</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Weisberg</surname>
          </string-name>
          ,
          <article-title>Characterizations of an empirical influence function for detecting influential cases in regression</article-title>
          ,
          <source>Technometrics</source>
          <volume>22</volume>
          (
          <year>1980</year>
          )
          <fpage>495</fpage>
          -
          <lpage>508</lpage>
          . doi:
          <volume>10</volume>
          .1080/00401706.
          <year>1980</year>
          .
          <volume>10486199</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Hammoudeh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Lowd</surname>
          </string-name>
          ,
          <article-title>Training data influence analysis and estimation: a survey</article-title>
          ,
          <source>Machine Learning</source>
          <volume>113</volume>
          (
          <year>2024</year>
          )
          <fpage>2351</fpage>
          -
          <lpage>2403</lpage>
          . URL: https://doi.org/10.1007/s10994-023-06495-7. doi:
          <volume>10</volume>
          .1007/s10994-023-06495-7.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>G.</given-names>
            <surname>Pruthi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kale</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sundararajan</surname>
          </string-name>
          ,
          <article-title>Estimating Training Data Influence by Tracing Gradient Descent</article-title>
          ,
          <source>in: Advances in Neural Information Processing Systems</source>
          , volume
          <volume>33</volume>
          ,
          <string-name>
            <surname>Curran</surname>
            <given-names>Associates</given-names>
          </string-name>
          , Inc.,
          <year>2020</year>
          , pp.
          <fpage>19920</fpage>
          -
          <lpage>19930</lpage>
          . URL: https://proceedings.neurips.cc/paper_ ifles/paper/2020/hash/e6385d39ec9394f2f3a354d9d2b88eec-Abstract.html.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>A.</given-names>
            <surname>Chapman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Lauro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Missier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Torlone</surname>
          </string-name>
          ,
          <article-title>Supporting better insights of data science pipelines with fine-grained provenance</article-title>
          ,
          <source>ACM Trans. Database Syst</source>
          .
          <volume>49</volume>
          (
          <year>2024</year>
          ). URL: https://doi.org/10.1145/3644385. doi:
          <volume>10</volume>
          .1145/3644385.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>M. T.</given-names>
            <surname>Ribeiro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Guestrin</surname>
          </string-name>
          ,
          <article-title>"Why Should I Trust You?" : Explaining the Predictions of Any Classifier</article-title>
          ,
          <source>in: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining - KDD '16</source>
          , ACM Press, New York, New York, USA,
          <year>2016</year>
          , pp.
          <fpage>1135</fpage>
          -
          <lpage>1144</lpage>
          . URL: http://dl.acm.org/citation.cfm?doid=
          <volume>2939672</volume>
          .2939778. doi:
          <volume>10</volume>
          . 1145/2939672.2939778.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>A.</given-names>
            <surname>Ghorbani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zou</surname>
          </string-name>
          , Data Shapley:
          <article-title>Equitable Valuation of Data for Machine Learning</article-title>
          ,
          <source>in: Proceedings of the 36th International Conference on Machine Learning, PMLR</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>2242</fpage>
          -
          <lpage>2251</lpage>
          . URL: https://proceedings.mlr.press/v97/ghorbani19c.html, iSSN:
          <fpage>2640</fpage>
          -
          <lpage>3498</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>R.</given-names>
            <surname>Jia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Dao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. A.</given-names>
            <surname>Hubis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Hynes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. M.</given-names>
            <surname>Gürel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Song</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. J.</given-names>
            <surname>Spanos</surname>
          </string-name>
          ,
          <article-title>Towards eficient data valuation based on the shapley value</article-title>
          ,
          <source>in: The 22nd International Conference on Artificial Intelligence and Statistics</source>
          , PMLR,
          <year>2019</year>
          , pp.
          <fpage>1167</fpage>
          -
          <lpage>1176</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>P. W.</given-names>
            <surname>Koh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Liang</surname>
          </string-name>
          ,
          <article-title>Understanding black-box predictions via influence functions</article-title>
          ,
          <source>in: International conference on machine learning, PMLR</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>1885</fpage>
          -
          <lpage>1894</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>H.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. F.</given-names>
            <surname>Rajani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Hase</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bansal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Xiong</surname>
          </string-name>
          , Fastif:
          <article-title>Scalable influence functions for eficient model interpretation and debugging, 2021</article-title>
          . URL: https://arxiv.org/abs/
          <year>2012</year>
          .15781. arXiv:
          <year>2012</year>
          .15781.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>J.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lécuyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Panda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Sen</surname>
          </string-name>
          ,
          <article-title>Measuring the Efect of Training Data on Deep Learning Predictions via Randomized Experiments</article-title>
          , in: Proceedings of the
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>