<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Udine, Italy
* Corresponding author.
$ albarelli@unive.it (A. Albarelli); claudio.lucchese@unive.it (C. Lucchese); matteo.rizzo@unive.it (M. Rizzo);
alberto.veneri@unive.it (A. Veneri)</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>On the Application of a Common Theoretical Explainability Framework in Information Retrieval</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Andrea Albarelli</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Claudio Lucchese</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Matteo Rizzo</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Alberto Veneri</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Ca' Foscari University of Venice</institution>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2024</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0002</lpage>
      <abstract>
        <p>Most of the current state-of-the-art models used to solve the search and ranking tasks in Information Retrieval (IR) are considered “black boxes” due to the enormous number of parameters employed, which makes it dificult for humans to understand the relation between input and output. Thus, in the current literature, several approaches are proposed to explain their outputs, trying to make the models more explainable while maintaining the high level of efectiveness achieved. Even though many methods have been developed, there is still a lack of a common way of describing and evaluating the models and methods of the Explainabile IR (ExIR) field. This work shows how a common theoretical framework for explainability (previously presented in the biomedical field) can be applied to IR. We first describe the general framework and then focus on specific explanation techniques in the IR field, focusing on core IR tasks: search and ranking. We show how well-known methods in ExIR fit into the framework and how specific IR explainability evaluation metrics can be described using this new setting.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Explainable Information Retrieval</kwd>
        <kwd>Theoretical Explainability Framework</kwd>
        <kwd>Search and Ranking</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        The emergence of Deep Learning (DL), and in particular the application of Pretrained Language
Models (PLMs), have drastically changed the Information Retrieval (IR) landscape. Previously
skeptically considered by part of the community [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], the advent of the first publicly available
PLM (BERT [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]) has completely changed the adoption of DL models in the IR field, especially in
the core IR tasks, i.e. search and ranking, given their high efectiveness. Even though PLM-based
approaches are highly efective, they are also opaque and way more challenging to analyze,
debug, and understand than the traditional IR methods, such as the well-known BM25 [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
Therefore, in pursuit of ensuring more reliable and trustworthy IR systems, recent years have
witnessed a growing interest in the field of Explainable Information Retrieval ( ExIR) [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. The
motivation to go beyond the opacity of the current state-of-the-art method is not purely technical
(e.g., to create more robust and simple to debug IR systems) or ethical (e.g., to easily investigate
possible unfair behavior in the model), but it also has a compliance nature, given the current or
upcoming international regulation for Artificial Intelligence ( AI) systems, such as the AI Act, in
which the concept of explainability is a crucial one [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], even though most of the state-of-the-art
eXplainable Artificial Intelligence ( XAI) methods can provide only a limited answer to the
compliance requirements [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>
        As commonly happens in a fast-growing field, one of the problems faced by ExIR is the
dificulty of relating diferent approaches since diferent terminology is used and diferent
evaluation metrics have been proposed. For example, some methods are called “interpretable”
while others “explainable“ and some works present evaluation metrics related to the same
concept. However, it is unclear how they should be compared or which characteristic of
the model they try to underline. Some recent surveys [
        <xref ref-type="bibr" rid="ref4 ref7">4, 7</xref>
        ] tried to organize the relevant
literature; however, they did not create a common framework useful to compare and evaluate
the explainability of diferent IR models. This is problematic, especially if we want to evaluate
various methods that belong to diferent categories, which usually have diferent evaluation
metrics.
      </p>
      <p>
        With this work, we aim to show how one of our recent contributions presented in the
biomedical domain, a theoretical framework for explainability [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], can be suitably employed
also in the IR domain, and thus describe all the methods present in the ExIR literature through
the same lens. By applying the framework to IR, we show how the current most popular
explanation methods in IR fit the framework. Even though we highlight that the framework
is not a novel contribution per se, we claim that starting to apply it to the new explanation
techniques and model presented in IR is an important step towards building more rigorous
explanation methods that take into account all the aspects related to this interdisciplinary
and complex subfield. We present how three so-called post-hoc explanation methods fit the
framework: LIRME[
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], MULTIPLEX [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], and the explainability techniques based on IR axioms,
while also analyzing two so-called “intrinsically interpretable” (or “intrinsically explainable”)
models such as ColBERT[
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] (not explicitly interpretable but considered interpretable) and
Interpretable LambdaMART (ILMART)[
        <xref ref-type="bibr" rid="ref12">12</xref>
        ].
      </p>
      <p>The paper is organized as follows: in Sec. 2, we present the most related works; in Sec. 3,
we briefly summarize the framework; in Sec. 4 we highlight the peculiarities of applying the
framework in the IR presenting 5 case studies, and, finally, in Sec. 5 we present our conclusions.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Works</title>
      <p>Even though the focus on the explainability aspects in IR is relatively new, numerous works
have been published, but no general framework has been proposed. To the author’s knowledge,
this work presents the application of a general framework to explainability techniques in IR
for the first time in the literature. Nonetheless, other works have been presented in the form
of surveys, categorizing ExIR works and trying to create a common taxonomy. However, they
do not directly try to create a common framework to describe and compare ExIR methods but
mainly focus on categorizing them.</p>
      <p>
        The more relevant survey is the one presented by Anand et al. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], in which the authors have
nicely categorized 32 ExIR approaches into three general categories: i) post hoc, 2) grounding to
IR properties, and 3) interpretable by design. The survey presents various explanation aspects,
including the diference between local and global explanations and the diference between
pointwise, pairwise, and listwise explanations. One of the main conceptual diferences between
their survey and our work is that they diferentiate between “interpretability" (interpretability
by design) and “explainability" (mainly post hoc explainability) while we approach the problem
diferently; we disagree with this categorization, and we claim that every model has some
explainability degree, and diferentiating between interpretable by design and post hoc explainable
by design models is not well defined and sometimes misleading.
      </p>
      <p>
        Saha et al. [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], in another survey, provide an overview of ExIR methods similar to [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] but
also add some methods from the related Natural Language Processing (NLP) field. Similarly
to Anand et al., they diferentiate between interpretable and explainable models and further
categorize the approaches into categories, including embeddings, sequence models, attention,
transformers, and BERT. Similarly to [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], their work is designed to solve the categorization
problem of the various explanation techniques, and they do not provide a common framework
to describe and compare the ExIR methods available.
      </p>
      <p>
        More broadly, some attempts have been made to outline the diferences in the terminology
used in XAI related field, such as in [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] and in [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. Still, they lack a definition of an explanation’s
inner structure and meaning.
      </p>
      <p>
        In brief, as we have common frameworks to describe the learning phase of various Machine
Learning (ML) tasks, including supervised learning, unsupervised learning, and reinforcement
learning, among others [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ], we aim to present a common framework to create an explanation
for a ML model decision. This work shows that the framework is also suitable for all the existing
ExIR techniques.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Theoretical Framework for Explainability</title>
      <p>
        In this section, we briefly recap each framework component, accompanying the descriptions
using, as an example, the well-known explanation technique based on analyzing the attention
weights used in a transformer model, e.g., [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. Since the transformer-based models are really
popular and the worthiness of the explanations based on the attention weights is debated [
        <xref ref-type="bibr" rid="ref17 ref18">17, 18</xref>
        ],
we aim to show how to analyze explanations techniques for which the consensus on their
usefulness is not shared across the community using a common framework. A schema of the
whole framework and the relation between the components taken from the original paper is
presented in Fig. 1.
      </p>
      <sec id="sec-3-1">
        <title>3.1. Explanation Framework components</title>
        <p>In the following paragraphs, we present the fundamental components of the framework: evidence,
interpretation, explanation, and explanation interface.</p>
        <p>Evidence The evidence, used to create explanations for an AI system, is any information that
we can retrieve from a model and that can give some understanding of its inner workings (e.g.,
model parameters, gradients, input/output values, etc.). Two related concepts of evidence are
the evidence extractor and the explanatory potential. The former is the process of retrieving the
evidence from the model and/or its input/output data, while the latter is described as “how much
of a model the selected type of evidence can explain”. In particular, the concept of explanatory
potential is fundamental in the framework since it helps the developer of the explanation
technique to understand the maximum expected faithfulness level of the explanation. The
measure of the explanatory potential can change case by case, just as we use diferent metrics
for evaluating diferent tasks. However, if we retrieve the evidence directly from the model, a
good metric for the explanatory potential can be the ratio between the number of parameters
analyzed and the total number of parameters of the model. We identify the evidence with the
symbol , and with (· ) the function computing the explanatory potential, and thus with
(), we identify the explanatory potential of the evidence.</p>
        <p>Example. In the case of attention-based explanation, the evidences are the attention weights,
which can be extracted just by retrieving the weights of the attention modules in the model. The
explanatory potential of this explanation can be measured as the ratio between the number of
parameters in the attention modules and the total number of parameters of the net. In the case
of BERT with only 12 transformer blocks1, we have that the parameters used in the attention
modules related to the “weight” and “bias” for queries and keys are 14,174,2082 while the total
number of parameters of the model is 109,482,240, and thus, we can say that the explanatory
potential is approximately 13%. In this way, the explanation technique’s explanatory potential
is between 0% and 100%, and an explanatory potential of 13% seems insuficient to provide a
faithful explanation.</p>
        <p>Interpretation An interpretation is a function applied to some evidence and mapping its
instances into explanations. Interpretation can be any function applied to the evidence. Still, it
is usually based on some social attribution, which makes it easier for humans to understand
the content and lowers the cognitive load. Interpretation can also be as simple as the identity
1https://huggingface.co/google-bert/bert-base-uncased
2We only consider the weights directly involved in computing the attention weights.
function, as in the so-called "white box" models, where no actual interpretation is needed; only
the evidence is suficient. We identify the interpretation function with the symbol (· ).</p>
        <p>Example. In the case of attention-based explanation, the common interpretation is that the
weights of the model’s attention module can summarize the importance of the input token to
the final prediction. There are two main caveats to this interpretation. First, each attention head
at each transformer block gives a diferent weight, and it is tricky to combine this information.
Second, each transformer block’s input and output tokens do not necessarily relate to the same
token. It is common to assume that the token in the -th position always refers to the same
concept, but this is not necessarily true; furthermore, assuming that each token is contextualized
after each block of the transformer, we could have that they replace the concept of the token
itself.</p>
        <p>Explanation The concept of “explanation” is defined as “the output of an interpretation
function applied to some evidence, providing the answer to a “why question” posed by the user”.
In mathematical terms, the explanation  results from applying the interpretation function to
the retrieved evidence,  = ().</p>
        <p>Example. In our attention-based example, the explanation is the triples formed by the input
token, the output token, and the associated attention weight for each head for each transformer
block.</p>
        <p>Explanation interface Last but not least, the explanation has to be presented to the end
user with an adequate user interface characterized by three main properties: (i) human
understandability, (ii) informativeness, and (iii) completeness. The human-understandability measures
how easily the user can understand the explanation provided. The informativeness (i.e., depth)
measures how much information is given to the user to understand the behavior of the AI
system for a specific user need. Finally, the completeness (i.e., width) describes how well the
explanation pictures the entire inner workings of the model.</p>
        <p>
          Example. In the case of attention-based explanation, every attention weight is presented by
transformer block and attention head interactively; see [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ] for an example.
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Explanation Framework Evaluation</title>
        <p>The framework presents two main aspects to take into account during the evaluation of an AI
system, namely the plausibility and the faithfulness of the explanation.</p>
        <p>Faithfulness We define the property of faithfulness of an interpretation as the degree to which
an interpretation accurately reflects the behavior of the transformation function applied by a ML
model. Various measures of faithfulness can be associated with diferent types of explanations,
analogous to the metrics used to evaluate an ML model’s performance on a given task.</p>
        <p>
          When designing a faithful explanation method, we can opt for two approaches. Faithfulness
can be achieved by design incorporating this property into pre-selected interpretations during
the model design process (white box models), or alternatively, we can ignore the explanation
during the design and propose an explanation after the creation of the model (post hoc
explanation). Although formal proofs are currently lacking in the literature, several tests for faithfulness
have been recently proposed [
          <xref ref-type="bibr" rid="ref17 ref18 ref19 ref20">19, 17, 18, 20</xref>
          ].
        </p>
        <p>Plausibility Plausibility is the degree to which an explanation aligns with the user’s
understanding of the model’s partial or overall inner workings. It is worth highlighting that
plausibility is mainly a property influenced by users. Unlike faithfulness, the plausibility of
explanations can mainly be assessed via user studies. We highlight that plausibility and
faithfulness can be competing properties of an explanation. Interestingly, an unfaithful but plausible
explanation may deceive a user into believing that a model behaves according to a rationale
when this is not the case. Given the attention weights’ low explanatory potential and tricky
interpretation, we claim that they are unfaithful but plausible. Fig. 2 provides a simplified
problem overview.</p>
        <p>
          Summing up, in this section, we presented the main components and evaluation criteria of
the framework proposed in [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ]. The components are general enough to include any explanation
technique, and it is worth highlighting that the framework does not distinguish between post
hoc explanation techniques and explainable by design models. All the models have to be
interpreted given some evidence. The models usually called “intrinsically interpretable” are
simply models in which the interpretation of the evidence is trivial.
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Applying the Framework to Search and Ranking Applications</title>
      <p>
        The proposed framework has been presented in the realm of bioinformatics, focusing on the
most common explanation techniques and explainable models in the literature. This section
shows how the framework can be successfully adapted to the particular case of ExIR. We
describe five explanation techniques and associated evaluation metrics from five diferent
works in the literature. In particular, we chose three so-called post hoc explanation methods:
LIRME [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], MULTIPLEX [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], and the explanation technique based on the adherence to IR axioms
presented in [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ]. In addition, we analyze two considered white-box models: ColBERT [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]
and ILMART [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. For each explanation method proposed, we identify i) the evidence used, 2)
the interpretation, 3) how the evaluation strategies relate to the concept of faithfulness and
plausibility. In this case, we leave the analysis of the explanation interface for future work.
4.1. LIRME
In [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], the authors present a method to locally approximate the function of a complex text
ranking model with a simple local surrogate function, similar to LIME [22] but considering
sampling strategies more suitable for the ranking task. The local surrogate model is created as a
scoring function (, ) = ∑︀∈∩ (, ) in which  represents a document,  the query,
 a term in the document and query and (, ) the term weight, learned with an optimization
function to approximate the real and complex function  (, ) of the text ranker.
Evidence After having fixed the query , the optimization function (Eq. 1 in [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]) to
deifne (· , · ) uses as evidence  the predictions made on a sample of the documents, thus
 = {(, )}1 where  is the total number of sampled documents,  is a sampled document
and  =  (, ) is the score of the complex model. The explanatory potential of this evidence
() can be estimated by the ratio of the documents sampled ( ) over the total number of
possible sampled documents in the (possibly infinite) neighborhood of the document. In [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], it
is assumed that the sampled documents are only those obtained by removing terms from the
document, thus limiting the neighborhood.
      </p>
      <p>Interpretation The interpretation function (· ), in this case, takes in input the whole evidence
 ︀( {(, )}1</p>
      <p>
        )︀ a return the model (· , · ). In other words, the assumption is that we can explain
the model’s behavior just by looking at its output and using a simple model that mimics the
behavior of  (· , · ). (· , · ), as defined in [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] gives an estimate of the term importance for the
model decision for each term, so if (, ) is relatively high, it is important for the scoring
function, and the opposite otherwise.
      </p>
      <p>Evaluation The authors assessed the quality of their explanation with two metrics, the
explanation consistency and the explanation correctness. On the one hand, evaluating the explanation
consistency measures how a “particular choice of samples around the pivot document, D, should
not result in considerable diferences in the predicted explanation vector.”. This metric measures
faithfulness since it measures how much the sampling can change the explanation provided.
Thus, this is a proxy for how dificult the function is in that particular subspace; the higher the
consistency, the higher the faithfulness of the explanation. On the other hand, the explanation
correctness measures how many terms of high contributions in the surrogate models correspond
to terms occurring in documents that are judged relevant by assessors. Since its formulation
does not consider the original model and is explicitly linked with the relevances given by the
assessors, it can be considered a sort of user study and, thus, a measure of plausibility.</p>
      <sec id="sec-4-1">
        <title>4.2. MULTIPLEX</title>
        <p>
          In [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ], the authors present an explanation method designed to find the subset of terms most
impacting in the prediction of a text-ranker. The problem statement defined the term subset
to be identified as “small,” and that can explain most of the preference pairs { ≻  } from
the original ranking  produced by a complex model, where  is the -th document in the
ranking  . To find the subset of terms, multiple simple ranking models are used to rank the
most important features.
        </p>
        <p>Evidence Similarly to LIRME, the evidence  can be defined as  = {(,  , )}1 where
 is the number of preference pairs are sampled from  ,  and  are document in position
 and  and  represents the preference of the complex model for the pair (either positive or
negative). The explanatory potential of this evidence (), as for LIRME, can be estimated by
the ratio of the sampled preference  with respect to the total number of possible preference
pairs. We have full explanatory potential for this particular task if all the pairs are sampled.
Interpretation The interpretation function (· ) for MULTIPLEX takes as input the whole
evidence and returns a subset of terms that explain most of the preference pairs. During the
interpretation, various assumptions and heuristics were taken into account to find the subset of
terms, including using only a limited subset of the terms used by the documents, using three
simple ranking models (term matching, position-aware, and semantic similarity) to identify the
utility of each term, and using the approximation introduced by the optimization algorithm to
combine the found utility of each term.</p>
        <p>
          Evaluation In [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ], the evaluation is mainly measured by the “fidelity” of the explanation.
The fidelity is computed with “the fraction of the maintained preference pairs by the explainers
given the explanation terms.” Naturally, in our framework, this is a measure of faithfulness
in the explanation. Besides an anecdotal example, no evaluation has been performed using
information from the end-user, and thus, no plausibility evaluation has been performed.
        </p>
      </sec>
      <sec id="sec-4-2">
        <title>4.3. Explanation by IR axioms</title>
        <p>
          Since other works, as [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ], consider the explanation through IR axioms a completely diferent
category, as a last example of so-called “post-hoc” explanation, we analyze the work by Câmara
and Hauf [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ]. In the aforementioned work, the authors used the concept of diagnostic datasets
to analyze if BERT fulfills the retrieval axioms proposed by [ 23]. In particular, they created one
diagnostic dataset, one for each axiom, and checked if the rank produced by the model was
aligned with the one artificially created using the heuristic. The explanation aimed to explain
the model predictions using one or more heuristics.
        </p>
        <p>Evidence As in the case of LIRME and MULTIPLEX, the evidence is only based on the model
score attributed to a document-query pair, i.e.,  = {(, )}1 for a fixed query, where 
is the -th document of a diagnosing dataset, and  is the associated score. The explanatory
potential () can be, therefore, estimated with the ratio between the number of documents
taken into account  and the number of documents in the countable (in general case, possibly
infinite) document space that can be created for the particular diagnosing dataset.
Interpretation The interpretation of the evidence ( = {(, )}1 ) says that if the
document order produced by BERT is aligned with the order of the diagnosing dataset, BERT follows
the particular axiom with which the dataset was created.</p>
        <p>Evaluation The evaluation is based only on the agreement between BERT ranking and
the order of the documents in the diagnosing dataset. Therefore, there is no measure of the
explanation’s faithfulness but only a quantitative measurement of its plausibility since measuring
the agreement with a diagnosing dataset is a proxy for measuring the agreement with a fictitious
user.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.4. ColBERT</title>
        <p>
          Even though not explicitly presented as an explainable model, ColBERT is usually considered
an “intrinsically interpretable” model [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] in which the weights on the so-called late interaction
between tokens are considered term importance [24]. The late interaction is implemented
using the MaxSim operator, in which for each query token representation after the last
transformer block , with 1 &lt;  &lt;  , where  is the number of query tokens, the maximum
similarity to all the other document tokens is computed and then summed up. The MaxSim
between a query  and a document  is therefore defined as: Φ( , ) = ∑︀
=1 , with
 = max∈{1,...,} (,  ), and where  is the number of document tokens,  is the -th
token of , and (· , · ) is a similarity function.
        </p>
        <p>Evidence The evidence used during the explanation is the subset of similarity values resulting
from the MaxSim operator. Thus, the evidence is the set of similarity values  = {}1 . Since
we can consider the set of similarities as a part of the model weights, the explanatory potential
is equal to () = / , where  is the total number of model weights. Since  in the
small BERT version is 512 and  &gt; 100, 000, 000, we have () &lt; 5.12 · 10− 6, where ()
is theoretically bounded between 0 and 1.</p>
        <p>Interpretation In the common interpretation, the similarity of the query-document token
pairs contributing to the summation {max∈{1,...,} (,  )}, represents the importance of
the term association between  and  . We can, therefore, rank the most important query and
document token for each query.</p>
        <p>Evaluation The original paper does not evaluate the explainability of those scores. However,
other papers have explored their properties, e.g. [24]. In this case, we highlight that the
explanatory potential of those weights is minimal and that the “contextualization” brought after
each transformer block might result in a token at the end of the transformer in which there is
more “context” than the token itself, as mentioned in the previous section.</p>
      </sec>
      <sec id="sec-4-4">
        <title>4.5. ILMART</title>
        <p>
          In [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ], an explainable by-design model for ranking with hand-crafted features has been
presented. The authors presented a simple additive model by constraining the well-known
LambdaMART algorithm [25] using only one or two features per tree, creating a scoring function
that is a sum of univariate or bivariate functions. Formally, the score of a query-document
pair (, ) is defined as (, ) = ∑︀∈ℳ  () + ∑︀{,}∈ℐ  (), where ℳ is the set of
feature, ℐ is the set of all the possible pair of features, and  () and  () are respectively
the univariate and bivariate functions. In addition, to limit the complexity of the function, the
author presents a greedy way to limit the number of univariate and bivariate functions.
Evidence The evidence in this type of model considered explainable by design (as others in
the literature as NeuralRankGAM [26] or BM25) is the entire model itself, and thus this type of
model has full explanatory potential.
        </p>
        <p>Interpretation Given the explainable design, interpreting the evidence is trivial and can be
formalized as the identity function since the output of () is the model itself. We highlight that
even though the interpretation is considered trivial, the final explanation provided might not be
plausible for the user.</p>
        <p>Evaluation The authors claim that the model does not need explanation and thus does not
provide any evaluation that can be mapped in our framework, but just a measure of the model’s
efectiveness (similarly to [ 26]). The plausibility aspect is only mentioned with an anecdotal
example of model visualization. In this case, both faithfulness and plausibility evaluation are
missing.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>In this paper, we presented how a theoretical explainability framework presented to be applied
in the biomedical domain can also be suitable for the IR field, with a particular focus on the
core IR tasks, i.e., search and ranking. We first summarize the framework and then show
how a selection of explanation techniques presented in the IR literature can easily fit in the
framework. We selected three so-called “post hoc" techniques (including one in the category
of explanation based on IR axioms) and two “white box" models and highlighted that the
explanation procedure is the same for all the methods considered. All explanations start from
evidence and are provided to the user through an interpretation; the main diference is that the
interpretation can be convoluted or trivial. We also showed that the evaluation performed in the
papers analyzed can always be mapped to one of the two evaluation categories we identified,
namely, faithfulness or plausibility. We claim that this unified view can be a starting point
to create a common vocabulary for the ExIR field and to allow a better comparison between
explanation techniques previously thought to be diametrically opposed, helping to pave the
path to a more structured and robust field development.
[22] M. T. Ribeiro, S. Singh, C. Guestrin, "Why Should I Trust You?": Explaining the Predictions
of Any Classifier, in: Proceedings of the 22nd ACM SIGKDD International Conference
on Knowledge Discovery and Data Mining, ACM, San Francisco California USA, 2016, pp.
1135–1144. URL: https://dl.acm.org/doi/10.1145/2939672.2939778. doi:10.1145/2939672.
2939778.
[23] H. Fang, C. Zhai, An exploration of axiomatic approaches to information retrieval, in:
Proceedings of the 28th annual international ACM SIGIR conference on Research and
development in information retrieval, ACM, Salvador Brazil, 2005, pp. 480–487. URL:
https://dl.acm.org/doi/10.1145/1076034.1076116. doi:10.1145/1076034.1076116.
[24] T. Formal, B. Piwowarski, S. Clinchant, A White Box Analysis of ColBERT, in: D. Hiemstra,
M.-F. Moens, J. Mothe, R. Perego, M. Potthast, F. Sebastiani (Eds.), Advances in Information
Retrieval, Springer International Publishing, Cham, 2021, pp. 257–263. doi:10.1007/
978-3-030-72240-1_23.
[25] C. J. Burges, From ranknet to lambdarank to lambdamart: An overview, Learning 11 (2010)
81.
[26] H. Zhuang, X. Wang, M. Bendersky, A. Grushetsky, Y. Wu, P. Mitrichev, E. Sterling, N. Bell,
W. Ravina, H. Qian, Interpretable Learning-to-Rank with Generalized Additive Models,
arXiv:2005.02553 [cs, stat] (2020). URL: http://arxiv.org/abs/2005.02553, arXiv: 2005.02553.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <source>The Neural Hype and Comparisons Against Weak Baselines, ACM SIGIR Forum</source>
          <volume>52</volume>
          (
          <year>2019</year>
          )
          <fpage>40</fpage>
          -
          <lpage>51</lpage>
          . URL: https://dl.acm.org/doi/10.1145/3308774.3308781. doi:
          <volume>10</volume>
          .1145/ 3308774.3308781.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          , BERT:
          <article-title>Pre-training of Deep Bidirectional Transformers for Language Understanding</article-title>
          , in: J.
          <string-name>
            <surname>Burstein</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Doran</surname>
          </string-name>
          , T. Solorio (Eds.),
          <source>Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          , Volume
          <volume>1</volume>
          (Long and Short Papers),
          <source>Association for Computational Linguistics</source>
          , Minneapolis, Minnesota,
          <year>2019</year>
          , pp.
          <fpage>4171</fpage>
          -
          <lpage>4186</lpage>
          . URL: https://aclanthology.org/N19-1423. doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>N19</fpage>
          -1423.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>S.</given-names>
            <surname>Robertson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Zaragoza</surname>
          </string-name>
          ,
          <article-title>The Probabilistic Relevance Framework: BM25 and Beyond</article-title>
          ,
          <source>Foundations and Trends® in Information Retrieval</source>
          <volume>3</volume>
          (
          <year>2009</year>
          )
          <fpage>333</fpage>
          -
          <lpage>389</lpage>
          . URL: https://www. nowpublishers.com/article/Details/INR-019. doi:
          <volume>10</volume>
          .1561/1500000019, publisher: Now Publishers, Inc.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>A.</given-names>
            <surname>Anand</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Lyu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Idahl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wallat</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , Explainable Information Retrieval: A Survey,
          <year>2022</year>
          . URL: http://arxiv.org/abs/2211.02405. doi:
          <volume>10</volume>
          .48550/arXiv. 2211.02405, arXiv:
          <fpage>2211</fpage>
          .02405 [cs].
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>L.</given-names>
            <surname>Nannini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Balayn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. L.</given-names>
            <surname>Smith</surname>
          </string-name>
          , Explainability in AI Policies:
          <article-title>A Critical Review of Communications, Reports, Regulations, and Standards in the EU, US, and UK</article-title>
          ,
          <source>in: Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency</source>
          , FAccT '23,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2023</year>
          , pp.
          <fpage>1198</fpage>
          -
          <lpage>1212</lpage>
          . URL: https://doi.org/10.1145/3593013.3594074. doi:
          <volume>10</volume>
          .1145/3593013.3594074.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>C.</given-names>
            <surname>Panigutti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Hamon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Hupont</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. Fernandez</given-names>
            <surname>Llorca</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. Fano</given-names>
            <surname>Yela</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Junklewitz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Scalzo</surname>
          </string-name>
          , G. Mazzini,
          <string-name>
            <given-names>I.</given-names>
            <surname>Sanchez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. Soler</given-names>
            <surname>Garrido</surname>
          </string-name>
          , E. Gomez,
          <article-title>The role of explainable AI in the context of the AI Act</article-title>
          ,
          <source>in: Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency</source>
          , FAccT '23,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2023</year>
          , pp.
          <fpage>1139</fpage>
          -
          <lpage>1150</lpage>
          . URL: https://dl.acm.org/doi/10.1145/3593013.3594069. doi:
          <volume>10</volume>
          .1145/3593013.3594069.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>S.</given-names>
            <surname>Saha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Majumdar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mitra</surname>
          </string-name>
          ,
          <source>Explainability of Text Processing and Retrieval Methods: A Critical Survey</source>
          ,
          <year>2022</year>
          . URL: http://arxiv.org/abs/2212.07126. doi:
          <volume>10</volume>
          .48550/arXiv.2212. 07126, arXiv:
          <fpage>2212</fpage>
          .07126 [cs].
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>M.</given-names>
            <surname>Rizzo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Veneri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Albarelli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Lucchese</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Nobile</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Conati</surname>
          </string-name>
          ,
          <article-title>A Theoretical Framework for AI Models Explainability with Application in Biomedicine</article-title>
          ,
          <source>in: 2023 IEEE Conference on Computational Intelligence in Bioinformatics and Computational Biology (CIBCB)</source>
          , IEEE, Eindhoven, Netherlands,
          <year>2023</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>9</lpage>
          . URL: https://ieeexplore.ieee.org/document/ 10264877. doi:
          <volume>10</volume>
          .1109/CIBCB56990.
          <year>2023</year>
          .
          <volume>10264877</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>M.</given-names>
            <surname>Verma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Ganguly</surname>
          </string-name>
          , LIRME: Locally Interpretable Ranking Model Explanation,
          <source>in: Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          , SIGIR'19,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2019</year>
          , pp.
          <fpage>1281</fpage>
          -
          <lpage>1284</lpage>
          . URL: https://doi.org/10.1145/3331184.3331377. doi:
          <volume>10</volume>
          .1145/3331184.3331377.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>L.</given-names>
            <surname>Lyu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Anand</surname>
          </string-name>
          ,
          <article-title>Listwise Explanations for Ranking Models Using Multiple Explainers</article-title>
          , in: J.
          <string-name>
            <surname>Kamps</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Goeuriot</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Crestani</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Maistro</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Joho</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Davis</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Gurrin</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          <string-name>
            <surname>Kruschwitz</surname>
            ,
            <given-names>A</given-names>
          </string-name>
          . Caputo (Eds.),
          <source>Advances in Information Retrieval</source>
          , Springer Nature Switzerland, Cham,
          <year>2023</year>
          , pp.
          <fpage>653</fpage>
          -
          <lpage>668</lpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>031</fpage>
          -28244-7_
          <fpage>41</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>O.</given-names>
            <surname>Khattab</surname>
          </string-name>
          , M. Zaharia,
          <article-title>ColBERT: Eficient and Efective Passage Search via Contextualized Late Interaction over BERT</article-title>
          ,
          <source>in: Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          , SIGIR '20,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2020</year>
          , pp.
          <fpage>39</fpage>
          -
          <lpage>48</lpage>
          . URL: https://doi.org/10.1145/3397271.3401075. doi:
          <volume>10</volume>
          .1145/3397271.3401075.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>C.</given-names>
            <surname>Lucchese</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. M.</given-names>
            <surname>Nardini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Orlando</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Perego</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Veneri</surname>
          </string-name>
          , ILMART:
          <article-title>Interpretable Ranking with Constrained LambdaMART</article-title>
          ,
          <source>in: Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          , SIGIR '22,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2022</year>
          , pp.
          <fpage>2255</fpage>
          -
          <lpage>2259</lpage>
          . URL: https://dl.acm.org/doi/10.1145/3477495.3531840. doi:
          <volume>10</volume>
          .1145/3477495.3531840.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>M.</given-names>
            <surname>Graziani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Dutkiewicz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Calvaresi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. P.</given-names>
            <surname>Amorim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Yordanova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Vered</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Nair</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. H.</given-names>
            <surname>Abreu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Blanke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Pulignano</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. O.</given-names>
            <surname>Prior</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Lauwaert</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Reijers</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Depeursinge</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Andrearczyk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Müller</surname>
          </string-name>
          ,
          <article-title>A global taxonomy of interpretable AI: unifying the terminology for the technical</article-title>
          and
          <source>social sciences, Artificial Intelligence Review</source>
          <volume>56</volume>
          (
          <year>2023</year>
          )
          <fpage>3473</fpage>
          -
          <lpage>3504</lpage>
          . URL: https://doi.org/10.1007/s10462-022-10256-8. doi:
          <volume>10</volume>
          .1007/s10462-022-10256-8.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>W. J.</given-names>
            <surname>Murdoch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kumbier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Abbasi-Asl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Yu</surname>
          </string-name>
          , Definitions, methods, and
          <article-title>applications in interpretable machine learning</article-title>
          ,
          <source>Proceedings of the National Academy of Sciences</source>
          <volume>116</volume>
          (
          <year>2019</year>
          )
          <fpage>22071</fpage>
          -
          <lpage>22080</lpage>
          . URL: https://www.pnas.org/doi/abs/10.1073/pnas. 1900654116. doi:
          <volume>10</volume>
          .1073/pnas.1900654116, company:
          <source>National Academy of Sciences Distributor: National Academy of Sciences Institution: National Academy of Sciences Label: National Academy of Sciences Publisher: Proceedings of the National Academy of Sciences.</source>
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>I.</given-names>
            <surname>Goodfellow</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bengio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Courville</surname>
          </string-name>
          ,
          <article-title>Deep learning</article-title>
          , MIT press,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>J.</given-names>
            <surname>Vig</surname>
          </string-name>
          ,
          <article-title>A Multiscale Visualization of Attention in the Transformer Model</article-title>
          , in: M.
          <string-name>
            <surname>R. Costajussà</surname>
          </string-name>
          , E. Alfonseca (Eds.),
          <article-title>Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics: System Demonstrations, Association for Computational Linguistics</article-title>
          , Florence, Italy,
          <year>2019</year>
          , pp.
          <fpage>37</fpage>
          -
          <lpage>42</lpage>
          . URL: https://aclanthology.org/P19-3007. doi:
          <volume>10</volume>
          . 18653/v1/
          <fpage>P19</fpage>
          -3007.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>S.</given-names>
            <surname>Jain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. C.</given-names>
            <surname>Wallace</surname>
          </string-name>
          ,
          <article-title>Attention is not Explanation, in: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Association for Computational Linguistics</article-title>
          , Minneapolis, Minnesota,
          <year>2019</year>
          , pp.
          <fpage>3543</fpage>
          -
          <lpage>3556</lpage>
          . URL: https://aclanthology.org/ N19-1357. doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>N19</fpage>
          -1357.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>S.</given-names>
            <surname>Wiegrefe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Pinter</surname>
          </string-name>
          ,
          <article-title>Attention is not not Explanation</article-title>
          ,
          <source>in: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)</source>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Hong Kong, China,
          <year>2019</year>
          , pp.
          <fpage>11</fpage>
          -
          <lpage>20</lpage>
          . URL: https://aclanthology.org/D19-1002. doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>D19</fpage>
          -1002.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>S.</given-names>
            <surname>Serrano</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. A.</given-names>
            <surname>Smith</surname>
          </string-name>
          , Is Attention Interpretable?,
          <source>in: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics</source>
          , Association for Computational Linguistics, Florence, Italy,
          <year>2019</year>
          , pp.
          <fpage>2931</fpage>
          -
          <lpage>2951</lpage>
          . URL: https://aclanthology.org/P19-1282. doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>P19</fpage>
          -1282.
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>J.</given-names>
            <surname>Adebayo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gilmer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Muelly</surname>
          </string-name>
          , I. Goodfellow,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hardt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <article-title>Sanity checks for saliency maps</article-title>
          ,
          <source>in: Proceedings of the 32nd International Conference on Neural Information Processing Systems</source>
          , NIPS'18, Curran Associates Inc.,
          <string-name>
            <surname>Red</surname>
            <given-names>Hook</given-names>
          </string-name>
          ,
          <string-name>
            <surname>NY</surname>
          </string-name>
          , USA,
          <year>2018</year>
          , pp.
          <fpage>9525</fpage>
          -
          <lpage>9536</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>A.</given-names>
            <surname>Câmara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Hauf</surname>
          </string-name>
          ,
          <article-title>Diagnosing BERT with Retrieval Heuristics</article-title>
          , in: J. M. Jose, E. Yilmaz,
          <string-name>
            <given-names>J.</given-names>
            <surname>Magalhães</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Castells</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ferro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. J.</given-names>
            <surname>Silva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Martins</surname>
          </string-name>
          (Eds.),
          <source>Advances in Information Retrieval, Lecture Notes in Computer Science</source>
          , Springer International Publishing, Cham,
          <year>2020</year>
          , pp.
          <fpage>605</fpage>
          -
          <lpage>618</lpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>030</fpage>
          -45439-5_
          <fpage>40</fpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>