<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Justification in Case-Based Reasoning</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Wijnand van Woerkom</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Davide Grossi</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Henry Prakken</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Bart Verheij</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Amsterdam Center for Law and Economics, University of Amsterdam</institution>
          ,
          <country country="NL">The Netherlands</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Bernoulli Institute for Mathematics</institution>
          ,
          <addr-line>Computer Science and Artificial Intelligence</addr-line>
          ,
          <institution>University of Groningen</institution>
          ,
          <country country="NL">The Netherlands</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Department of Information and Computing Sciences, Utrecht University</institution>
          ,
          <country country="NL">The Netherlands</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Faculty of Law, University of Groningen</institution>
          ,
          <country country="NL">The Netherlands</country>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>Institute for Logic, Language and Computation, University of Amsterdam</institution>
          ,
          <country country="NL">The Netherlands</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The explanation and justification of decisions is an important subject in contemporary data-driven automated methods. Case-based argumentation has been proposed as the formal background for the explanation of data-driven automated decision making. In particular, a method was developed in recent work based on the theory of precedential constraint which reasons from a case base, given by the training data of the machine learning system, to produce a justification for the outcome of a focus case. An important role is played in this method by the notions of citability and compensation, and in the present work we develop these in more detail. Special attention is paid to the notion of compensation; we formally specify the notion and identify several of its desirable properties. These considerations reveal a refined formal perspective on the explanation method as an extension of the theory of precedential constraint with a formal notion of justification.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Precedential constraint</kwd>
        <kwd>Interpretability</kwd>
        <kwd>Law</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        In [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] a case-based reasoning method is proposed to explain data-driven automated decisions
for binary classification, based on the theory of precedential constraint introduced in [
        <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
        ].
This method is motivated by an analogy between the way in which a machine learning system
draws on training data to assign a label to a new data point and the way in which a court of
law draws on previously decided cases to make a decision about a new fact situation, because
in both of these situations the precedent that has been set must be adhered to as closely as
possible. The theory of precedential constraint, which has been developed to describe the type
of a fortiori reasoning used for legal decision making on the basis of case law, can therefore be
applied to analyze machine-learned decisions that are made on the basis of training data.
      </p>
      <p>
        More specifically, the method of [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] formally models the kind of dialogue in which lawyers
cite precedents to argue in favor of their preferred outcome of the new fact situation. These
citations, and the way in which they attack the opponent’s citation, are formalized using an
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Precedential Constraint</title>
      <p>
        The theory of precedential constraint was developed in [
        <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
        ] to describe the a fortiori reasoning
involved with case law. It is taken as the point of departure of the explanation method in [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]
and so we begin by recalling those aspects of it that are necessary for the rest of this work. The
contents of this section are largely similar to [6, Section 2].
      </p>
      <p>In order to describe the fact situation of a case we use what are called dimensions in the ai &amp;
law literature, which are formally just partially ordered sets.</p>
      <sec id="sec-2-1">
        <title>Definition 2.1.</title>
        <sec id="sec-2-1-1">
          <title>A dimension is a partially ordered set (, ⪯ ).</title>
          <p>We will frequently omit explicit reference to the dimension order ⪯ and instead refer to just
the set  when we speak of a dimension. A model of precedential constraint of a specific domain
assumes there is a set of these dimensions , relative to which the rest of the definitions are
specified.</p>
          <p>Definition 2.2. A fact situation  is a choice function on the set of dimensions , i.e. for each
dimension  ∈  an element  () ∈  of that dimension is chosen by  . A case  is a fact
situation  paired with an outcome  ∈ {0, 1}, written  =  : . A set CB of cases is called a
case base. If  =  :  we may write () instead of  ().</p>
          <p>In the context of a case , , , . . . we will refer to its fact situation by the corresponding
upper case letters , , , . . . without further explicit mention.</p>
          <p>The order ⪯ of a dimension  specifies the relative preference the elements of  have towards
either of two outcomes 0 and 1. More specifically, if  ≺  for ,  ∈  this means  prefers
outcome 1 relative to , and conversely  prefers outcome 0 relative to . Usually we want to
compare preference towards an arbitrary outcome , so to do this we define for any dimension
(, ⪯ ) the notation ⪯  := ⪯ if  = 1 and ⪯  := ⪰ if  = 0.</p>
          <p>Definition 2.3. Given fact situations  and  we say  is at least as good as  for an outcome
, denoted  ⪯  , if it is at least as good for  on every dimension :</p>
          <p>⪯   if and only if  () ⪯  () for all  ∈ .</p>
          <p>If moreover  =  :  is a previously decided case we say that  forces the decision of  for . A
case base CB forces the decision of  for  if it contains a case that does so.
Definition 2.4. Given two cases  =  :  and  =  :  such that  ⪯   we say that the
outcome of  for  was forced by the case , and write  ⪯ .</p>
          <p>To give some intuition for these definitions we consider a running example of risk of
recidivism, as in [6, Example 2.1].</p>
          <p>Example 2.1. Convicts are described along three dimensions: age (Age, ⪯ Age), the number
of prior ofenses (Priors, ⪯ Priors), and sex (Sex, ⪯ Sex). Age and number of priors have the
natural numbers as possible values, so Age := N and Priors := N. The values for sex are
Sex := {M, F}. The outcome for this domain is a judgement of whether the person is at high
(1) or low (0) risk of recidivism. The associated orders are as follows:</p>
          <p>(Age, ⪯ Age) := (N, ≥ ),
(Priors, ⪯ Priors) := (N, ≤ ),</p>
          <p>(Sex, ⪯ Sex) := ({M, F}, {(F, F), (M, M), (F, M)}).</p>
          <p>If a relation  is defined on all dimension we can, for fact situations  and , refer to the set
of dimensions on which  holds with [(, )] := { ∈  | ( (), ())}. For instance,
instantiating  := ̸⪯  we have [ ̸⪯  ] = { ∈  |  () ̸⪯  ()}; the dimensions on
which  is not at least as good for  as  . Besides fact situations we will also consider partial
fact situations, i.e. fact situations defined only on a particular subset of the dimensions. We can
do so conveniently using the well established notation for function restriction. Let  :  → 
and  ⊆ , we obtain a function  ↾  :  →  by restriction:  ↾  := {(, ) ∈  |  ∈ }.
For cases  and  with the same outcome  we write  (, ) :=  ↾ [ ̸⪯  ], the values of 
on which  is worse than  for , and (, ) :=  ↾ [ ⪯  ], the values of  on which  is
better than  for .</p>
          <p>Example 2.2. Suppose we have a case base of recidivism risk judgements, and two cases , 
with outcome 1 (i.e. judged high risk of recidivism) such that:
(Age) = 45,</p>
          <p>Now we can compute that  (, ) = {(Age, 50)} and (, ) = {(Priors, 5), (Sex, M)}.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. A Case-Based Reasoning Explanation Method</title>
      <p>
        In this section we detail the workings of the dimension-based model of explanation of [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], which
was inspired by the work of [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. A more detailed comparison between [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], and other related
works, can be found in [1, Section 8]. The method is built upon the theory of precedential
constraint of [
        <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
        ] and conceptually tries to mimic the arguments relating to precedent used by
lawyers with respect to case law. In such discussions, precedent cases are cited by both sides as
a means of arguing that the present (focus) case should be decided similarly as the precedent.
Both sides may attack the other’s citations, by pointing to important diferences between the
citation and the focus case; and they may defend themselves against such attacks, by pointing
to aspects of the focus case which compensates for these diferences. Each of the elements of
such a discussion – case citations, pointing to diferences, and compensating for diferences –
has its counterpart in the formal model of explanation.
      </p>
      <p>A key idea underlying the approach is that a tabular dataset for binary classification can be
interpreted as a case base CB in the sense of Definition 2.2. The method assumes access to the
training data used by the system, and interprets each of the features in the data as a dimension
in the sense of Definition 2.1. The corresponding dimension orders may be determined by
knowledge engineering, statistical methods, or a combination thereof. This gives us a body of
precedent CB upon which the machine learning system bases its decisions.</p>
      <p>Under this interpretation the machine learning system can be seen as deciding new fact
situations for sides. The goal is to explain a particular decision of a fact situation  for a side
, called the focus case  =  : . This explanation is provided in the form of a best citable
precedent  ∈ CB together with an explanation dialogue in which the choice for this  is justified.
This dialogue is formalized as a winning strategy in the grounded argument game of a particular
abstract argumentation framework.</p>
      <p>
        Before we can apply the theory of precedential constraint, we should specify the dimensions
as in Definition 2.1, and we begin in Section 3.1 by mentioning the method used for doing
so in [
        <xref ref-type="bibr" rid="ref1 ref6">1, 6</xref>
        ]. Any explanation dialogue should start with the citation of a best citable case. A
suggestion for the definition of this notion is given in [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] and we continue by recalling it in
Section 3.2, after which we explain and motivate the presence of the arguments occurring in
the argumentation framework in Sections 3.3 and 3.4. We are then ready to give the formal
definition of the framework in Section 3.5, explain what it means to have a winning strategy in
the argument game it induces, and as such what constitutes an explanation according to the
model.
3.1. Determining the Dimension Orders
In order to instantiate the explanation method for a particular dataset, we should specify the
dimension orders as in Definition 2.1. As just noted, this may be done on the basis of knowledge
engineering and/or statistical methods. In [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] a general method for determining the orders
corresponding to the dimensions was proposed, using a function  that associates each ordinal
feature  in the data with a coeficient expressing the degree to which the values in the range
of  prefer outcome 1. See [6, Section 4.2] for a more detailed explanation.
3.2. Citability of Cases
An important aspect of the explanations produced by the method of [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] is the selection of the
precedent case  with which it initiates its explanation of the outcome of the focus case  . We
will now describe how this selection procedure works; later in Section 4 we return to this topic
to suggest improvements. We begin with the notion of citability.
      </p>
      <sec id="sec-3-1">
        <title>Definition 3.1.</title>
        <p>A case  is citable for a case  if
(a) both cases have the same outcome ; and
(b) there is a dimension  such that () ⪯   ().</p>
        <p>Since this is a quite weak requirement there may in general be very many citable cases 
for any given  . For this reason the notion is strengthened by requiring that  should have a
minimal number of relevant diferences with  , according to some suitable notion of minimality.
To make this formal we should first define what a relevant diferences is. This is accomplished
by [1, Definition 11], which we repeat here.</p>
        <p>Definition 3.2. The set (,  ) of relevant diferences between  =  :  and  =  :  is
(,  ) :=  ↾ [ ̸⪯   ] = {(,  ()) |  ∈ ,  () ̸⪯   ()}.</p>
        <p>In other words, the relevant diferences are given by the values of the precedent  on the
dimensions on which  is not better than  for . Now a best citable precedent should minimize
this set of diferences, in the following sense.</p>
      </sec>
      <sec id="sec-3-2">
        <title>Definition 3.3.</title>
        <p>
          A case  is a best citable case for a case  if
(a)  is citable for  ; and
(b) there is no other  satisfying (a) for which (,  ) ⊂ (,  ).
3.3. Compensation of Relevant Diferences
An idea central to the explanation dialogues is that when a precedent  does not force a focus
case  , the values  (,  ) on which  is worse than  for their outcome can be compensated
for by the values (,  ) on which  is better than . This idea is often encountered in the
literature on case-based reasoning, see e.g. [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ], where certain compensations are described as
“showing that at a more abstract level, a parallel exists between the cases, arguing in efect that the
apparent distinction is merely a mismatch of details."
        </p>
        <p>In our context we assume the existence of a relation SC on partial fact situations , , where
SC(, ) says that  compensates for . This is used in practise as follows. Consider a precedent
 and a focus case  , both with outcome . If  forces the decision of  then  is at least as good
as  for  on all dimensions, so ∅ =  (,  ) or equivalently (,  ) =  . If this is not the case,
then ∅ ⊂  (,  ) or equivalently (,  ) ⊂  , and for  to justify the outcome of  we should
have that (,  ) compensates for  (,  ) as determined by whether SC((,  ),  (,  ))
holds.
3.4. Opposing Citations and Case Transformations
The last component of the dialogue is opposing citations, to which a response is possible through
the use of case transformations. The idea is that the proponent of the decision of  for its
outcome  can have their citation countered by the citation of a case  with outcome ¯, as a
means of saying that  should be a more appropriate precedent to draw on. This is analogous
to the argument between lawyers in a legal case.</p>
      </sec>
      <sec id="sec-3-3">
        <title>Definition 3.4.</title>
        <p>We define a semantics function J· K on the compensation arguments by:</p>
        <p>JCompensates(, )K := ( ∖  ↾ dom()) ∪  : .</p>
        <p>A case  can be transformed into  if  =  or there exists  ∈  such that JK = .</p>
        <p>
          The goal of the semantics function is to change  into a case  that forces the outcome of  .
It does so by replacing the values of the precedent case with those of the focus case, on those
dimensions on which the focus case is not at least as good as the precedent.
3.5. An Abstract Argumentation Framework for Explanation
We are now ready to describe the formal account of the explanation dialogues in [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ] through
the use of an abstract argumentation framework, a concept introduced in [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]. An abstract
argumentation framework AF = (Arg, Attack) is a directed graph, in which the nodes are
interpreted as arguments and the edges as an attack relation between them.
        </p>
        <p>
          An argumentation framework (Arg, Attack) is defined in [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ] that combines the types of
arguments defined in the preceding Sections 3.2, 3.3, and 3.4, relative to a focus case  =  : .
To do so we first define, for a particular precedent  =  :  that may be cited in defense of the
decision of  for , a subset  ⊆ Arg as follows:
 := ⋃︁ {︁{Worse() |  =  (,  ) ̸= ∅},
(1)
{Compensates(, ) | Worse() ∈ ,  ⊆ (,  ), SC(, )},
{Transformed() |  can be transformed into a case  with  ⪯  }}︁.
        </p>
        <p>Definition 3.5. Given a finite case base CB, a focus case  =  : , and a compensation
relation SC, an abstract argumentation framework for explanation with dimensions is a pair
AF = (Arg, Attack) where the arguments Arg are given by</p>
        <p>Arg := CB ∪ ⋃︁{ |  ∈ CB if  has the same outcome as  },
and for arguments ,  ∈ Arg we have Attack(,  ) if and only if either:
• ,  ∈ CB have diferent outcomes and [ ̸⪯  ] ̸⊂ [ ̸⪯  ];
•  ∈ CB and  is of the form Worse ();
•  is of the form Worse() and  is of the form Compensates(, ); or
•  ∈ CB has outcome ¯ and  is of the form Transformed().</p>
        <p>
          A dialogue now takes the form of a grounded argument game played on (Arg, Attack). For
the sake of brevity we only give an intuitive explanation of how this works, the reader is referred
to [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ] for a detailed treatment of the subject.
        </p>
        <p>An argument game on an AF (, ) is a two-player game, in which the players take turns
playing arguments from  which must attack the previously played argument according to
the attack relation . A player can win the game by moving an argument to which the other
player cannot reply, and a winning strategy for a player is a method of playing that ensures a
win regardless of how the opponent plays.</p>
        <p>
          We now have the formal machinery in place to define explanations as in [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ].
        </p>
        <p>Definition 3.6. An explanation of a focus case  is a winning strategy in the grounded argument
game starting with the citation of a best citable precedent  ∈ CB for  , played on the abstract
argumentation framework for explanation with dimensions (Arg, Attack).</p>
        <p>The winning strategies may be viewed as trees and have the following general shape:</p>
        <p>Worse()
Compensates(, )</p>
        <p>Transformed(1)

1
. . .</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. On the Citability of Cases</title>
      <p>Let us now consider some possible modifications of Definition 3.3 to better formalize the intuitive
notion of a most closely related case  of our focus case  .</p>
      <p>Firstly, since Definition 3.2 does not gather just the dimensions on which  is worse than
 but also the value of  at that dimension, a situation can arise where there is some case 
with [ ̸⪯   ] ⊂ [ ̸⪯   ] but  ↾ [ ̸⪯   ] ̸⊂  ↾ [ ̸⪯   ], just because there is some
dimension  ∈ [ ̸⪯   ] with () ̸=  (). It does not seem correct to dismiss  as a good
citation simply because it disagrees with  on a single dimension, especially when [ ̸⪯   ] is
only a very small subset of [ ̸⪯   ]. Let us look at an example to illustrate this point.
Example 4.1. We consider three cases , ,  with outcome 1 (meaning they were judged high
risk of recidivism) in the recidivism scenario of Example 2.1:
(Age) = 20,</p>
      <p>We have that (,  ) = {(Age, 20), (Priors, 3)} and (,  ) = {(Priors, 1)}. Therefore,
even though there are fewer dimensions on which  has relevant diferences with  – as
{Priors} ⊂ { Age, Priors} – this does not prevent  from being considered a best citable
precedent for  – as {(Priors, 1)} ̸⊂ { (Age, 20), (Priors, 3)}.</p>
      <p>This consideration suggests the definition should require minimality of [ ̸⪯   ] instead
of  ↾ [ ̸⪯   ]. However, this modification leaves room for a second type of scenario
where there is some precedent  which is intuitively much closer to the focus case relatively
to some other , without hindering  from being considered best citable. To see why we
consider a set of  + 1 dimensions {0, . . . , }. Now we may have that [ ̸⪯   ] = {0}
and [ ̸⪯   ] = {1, . . . , }. This means that the presence of  does not hinder ’s being
considered a best citable precedent for  , even though  is worse than  on  times as many
dimensions as it is worse on than . To remedy this, we could require minimality of the number
of dimensions rather than the set of dimensions itself, i.e. of |[ ̸⪯   ]|.</p>
      <p>In addition to looking just at diferences between the precedent and focus case it may be
beneficial to also consider the similarities since after all, the stare decisis doctrine states that
similar cases must be decided similarly. To achieve this we can require the best citable precedent
to subsequently maximize |[ =  ]|, so that it both minimizes diferences and maximizes
similarities. In all, this leads us to the following definition.</p>
      <p>Definition 4.1. A case  is a best citable case for a case  if it satisfies the conditions (a)  is
citable for  ; (b) there is no other  satisfying (a) with |[ ̸⪯   ]| &lt; |[ ̸⪯   ]|; (c) there is no
other  satisfying (a) and (b) with |[ =  ]| &gt; |[ =  ]|.</p>
      <p>
        Experimental results in [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] showed that there are in general many cases satisfying Definition
3.3 for any  . Measured on three datasets, the mean and standard deviation of the number of best
citable cases were respectively 82 ± 123.6, 76 ± 134, and 106 ± 116.5 [1, Table 5]. Recalculating
these statistics for the same datasets with Definition 4.1 instead results in respectively 5.6 ± 2.0,
2.1 ± 2.6, and 2.6 ± 2.5 average number of best citable cases; a substantial decrease. Still, the
definition remains somewhat ad-hoc, and more research is needed to assess its adequacy.
      </p>
    </sec>
    <sec id="sec-5">
      <title>5. Specifying the Compensation Relation</title>
      <p>
        In [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] no further explicit assumptions are made of the compensation relation SC. However in
order for this relation to function according to our intuitions it may be necessary to do so, and
we now consider a few such requirements. Let us first illustrate SC through a continuation of
Example 2.2.
      </p>
      <p>Example 5.1. We saw two example cases ,  where  was worse than  on the dimensions
Age and Sex, but better on Priors. Suppose that for a number of priors higher than 4, we no
longer care about values besides the number of priors. Then we may define</p>
      <p>SC(, ) if and only if (Priors) ≥ 4.</p>
      <p>In this case the worse values  (, ) would indeed be compensated for by the better values
(, ), since (Priors) = 5.</p>
      <p>A point to consider is whether the compensation relation should itself adhere to an a fortiori
principle. That is to say, if a set  is capable of compensating for a set , should a superset  ⊇ 
be capable of compensating for  as well?
Definition 5.1. A compensation relation SC is monotone if for any partial fact situations , , 
it holds that SC(, ) implies SC( ∪ , ).</p>
      <p>The same goes for values that are being compensated for; if a set  can compensate for a set
 then we might require of it to compensate any subset  ⊆  as well.</p>
      <p>Definition 5.2. A compensation relation SC is antitone if for any partial fact situations , , 
it holds that SC(, ) implies SC(,  ∩ ).</p>
      <p>
        In the factor based model of explanation in [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], i.e. the special case where the dimensions are
all two element sets with a linear order, it is possible to compensate for a set of worse values in
parts through the use of a pSubstitutes(, , )&amp;cCancels(′, ′, ) move [1, Definition 5]. We
can translate this to the dimensional setting as follows.
      </p>
      <p>Definition 5.3. A compensation relation SC is linear if for any partial fact situations , , , 
it holds that SC(, ) and SC(, ) imply SC( ∪ ,  ∪ ).</p>
      <p>A more fundamental question regarding the compensation relation is that of context
dependence; should the compensation of two sets be allowed to depend on the context in which it
takes place? This question and its consequences are the subject of Section 6.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Justification as an Extension of Forcing</title>
      <p>An interesting way to think of the compensation relation is as an extension of the notion of
forcing between cases. In essence a compensation says that while a precedent  might not force
the decision of some other case , the obstructing relevant diferences can be compensated, and
so the precedent  may still be said to justify the outcome of .
6.1. Context-Dependent Compensations
A downside of the formal specification of this compensation relation is that it is defined on partial
fact situations, rather than just fact situations. This makes it impossible for compensations to
take the values of the precedent into account when allowing compensations to be made.
Example 6.1. In Example 2.2 the diference in age between  and  is only 5, and we may want
to say that (, ) compensates for  (, ) in this case if we find this diference small enough
to be insignificant. To make this compensation possible formally we would need to postulate
SC({(Age, 50)}, {(Priors, 5), (Sex, M)} but this would inadvertently sanction compensations
where the age of the precedent case is, say, 20, in which case we may find the diference in age
large enough to be significant.</p>
      <p>Modifying SC so that it takes the precedents’ values into account yields a relation on full fact
situations. A natural requirement of any such relation is that it extends the forcing relation ⪯
of Definition 2.4. This is akin to saying that any set can compensate for the empty set. This
leads us to the following definition.</p>
      <p>Definition 6.1. A relation ⊑ on cases is called a justification relation if it extends the forcing
relation ⪯ , i.e. if ⪯ ⊆ ⊑ .</p>
      <p>Note that any compensation relation SC gives rise to a justification relation
⊑SC:
 ⊑SC  if and only if  ⪯  or SC((, ),  (, )).
The converse does not hold, precisely because a justification relation takes into account the
context of the compensation. To see this, consider the naïve approach of obtaining a compensation
relation SC⊑ from a justification relation ⊑:</p>
      <p>SC⊑(, ) if and only if  ⊑  for some ,  with  =  (, ),  = (, ).
(3)
The problem is that this definition is not necessarily well defined , meaning that the truth value of
SC⊑(, ) may depend on the particular representatives  and  that are used for its evaluation.
This leads us to define the notion of a context-independent ⊑, requiring exactly that the relation
SC⊑ above is well defined.</p>
      <p>
        Definition 6.2. A justification relation ⊑ is context-independent if for any four cases , , , 
with  (, ) =  (, ) and (, ) = (, ) it holds that  ⊑  if  ⊑ .
6.2. Winning Strategies and Justification
The terminology of Definition 6.1 is inspired by [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], where an argument is said to be justified if
and only if the proponent has a winning strategy in the grounded argument game about the
argument. We will now formally justify this comparison by showing that for any compensation
relation SC the proponent of an initial citation  has a winning strategy in the game on the
argumentation framework if and only if  ⊑SC  (of Eq. (2)).
      </p>
      <p>
        Let us fix a precedent case  and a focus case  , and introduce some shorthand terminology
to ease our work. We will say a case  has a winning strategy if the proponent has a winning
strategy in the grounded argument game on the explanation AF (Arg, Attack) of Definition 3.5,
starting with a citation of . Following [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] we distinguish between nontrivial winning strategies
for , in which  can be attacked by a Worse() move, and trivial winning strategies for ,
in which there is no Worse() attack possible. In other words, a winning strategy for  is
nontrivial if Worse() ∈  and trivial if Worse() ̸∈ , with  as defined in Eq. (1).
Proposition 6.1. There is a trivial winning strategy for  if and only if  ⪯  .
Proof. Note that Worse() ̸∈  if  (,  ) = ∅ if  ⪯  . Hence left to right is immediate.
For right to left we note in addition that any citation made by the opponent can be attacked
with a Transformed() move, and so since there is no reply possible to a Transformed move
the proponent has a (trivial) winning strategy for .
      </p>
      <p>Proposition 6.2. There is a nontrivial winning strategy for  if and only if  (,  ) ̸= ∅ and
SC((,  ),  (,  )).</p>
      <p>Proof. Suppose the proponent has a winning strategy. Since Worse() ̸∈  attacks the initial
citation of  there should be a Compensates(, ) response to the Worse() move available
to the proponent, with  = (, ). This implies that SC((,  ),  (,  )).</p>
      <p>For the other direction we begin by noting that because  (, ) ̸= ∅ there is
Worse() ∈ , and so the assumption SC((,  ),  (,  )) guarantees that there is
 = Compensates(, ) ∈ . Now, there are two types of moves available to the opponent
to which we need a reply.</p>
      <p>1. The first is Worse() ∈ . As mentioned we have a reply  available, and since a
compensation move cannot be replied to the game is won by the proponent.
2. The second is the citation of a case  ∈ CB with outcome ¯ for which it holds that
[ ̸⪯  ] ̸⊂ [ ̸⪯  ]. By Definition 3.4 we have that  can be transformed into ′ = JK,
and so we can reply to the citation with Transformed() ∈ . There are no more
moves available to the opponent and so the proponent wins the game.</p>
      <sec id="sec-6-1">
        <title>Corollary 6.2.1. There is a winning strategy for  if and only if  ⊑SC  .</title>
        <p>Proof. Applying Eq. (2) and then Propositions 6.1 and 6.2 we get
 ⊑SC  if  ⪯  or SC((, ),  (, ))
if  has a trivial winning strategy or  has a nontrivial winning strategy
if  has a winning strategy.</p>
        <p>Under this view of the winning strategies, and employing a fully general definition of
compensation through a justification relation ⊑, we can now rephrase Definition 3.6 of explanations
in the following way.</p>
        <sec id="sec-6-1-1">
          <title>Definition 6.3.</title>
        </sec>
      </sec>
      <sec id="sec-6-2">
        <title>An explanation of a case  is a best citable precedent  ∈ CB with  ⊑  .</title>
        <p>
          The theory of precedential constraint describes how the outcome of a fact situation can
be forced by precedent. However the collection of precedents may not be suficient to force
the outcome of all possible new fact situations. If such an undecided fact situation presents
itself there may still be a precedent which, on the basis of additional reasoning, can be argued
to justify an outcome for the fact situation. This is the view suggested by Corollary 6.2.1; a
justification relation goes beyond the forcing relation by sanctioning citations of precedents
that do not strictly force the outcome of the focus case.
6.3. A Relational Description of the Explanation Model
Having shown that a justification relation in some sense corresponds to the winning strategies
underlying the explanations of [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ], we can give a succinct description of the explanation method
just through the use of relations on cases. Let us think of citability as a relation ⊴, then those
 ∈ CB related to the focus case through the intersection ⊑ ∩ ⊴ with  are said to explain the
focus case  , i.e. those  with  ⊑  and  ⊴  .
        </p>
        <p>
          The model in [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ] is a top-level model as it does not give explicit definitions of these notions,
apart from suggesting a definition for the citability relation ⊴ as in Definition 3.3, and a method
for determining ⪯ on the basis of Pearson correlation coeficients. In its running example and
the experiments in [1, Section 6] all compensations are allowed, so that ⊑ ∩ ⊴ = ⊴. Through
the relational view we summarize these inputs as follows: 1. The forcing relation ⪯ , determined
by specifying the dimensions and their orders. 2. The justification relation ⊑, determined by
specifying the compensations. 3. The citability relation ⊴, determined by the definition of a
best citable precedent. This view considerable simplifies the presentation of the model as it
does not rely on the concepts of argumentation frameworks and winning strategies.
        </p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>7. Discussion and Conclusion</title>
      <p>
        We have described the explanation model of [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] in Section 3, which provides explanations as
winning strategies on the grounded argument game of an abstract argumentation theory. In
Section 6 we showed that this model admits an equivalent rephrasing in terms of relations, in
which explanations are provided as cases related to the focus case through justification and
citation relations. Most notably this shows that the explanation model can in some sense be
seen as adding a notion of justification to the theory of precedential constraint as a relation ⊑
extending the forcing relation ⪯ .
      </p>
      <p>We conclude by noting an important consideration for future work on the topic. In order to
apply this notion of justification to the explanation of machine-learned decisions, it is imperative
that the input parameters – that being the forcing, justification, and citation relations – are
constructed in such a way that they are faithful to the rationale of the black-box under
consideration, because otherwise such a justification runs a high risk of becoming a rationalization if
it does not reflect the real reasons behind the decision.</p>
    </sec>
    <sec id="sec-8">
      <title>Acknowledgments</title>
      <p>This research was (partially) funded by the Hybrid Intelligence Center, a 10-year programme
funded by the Dutch Ministry of Education, Culture and Science through the Netherlands
Organisation for Scientific Research, grant number 024.004.022. We also thank the referees for
very useful commentary and suggestions.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>H.</given-names>
            <surname>Prakken</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ratsma</surname>
          </string-name>
          ,
          <article-title>A top-level model of case-based argumentation for explanation: Formalisation and experiments</article-title>
          ,
          <source>Argument &amp; Computation</source>
          <volume>13</volume>
          (
          <year>2022</year>
          )
          <fpage>159</fpage>
          -
          <lpage>194</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>J. F.</given-names>
            <surname>Horty</surname>
          </string-name>
          ,
          <article-title>Rules and reasons in the theory of precedent</article-title>
          ,
          <source>Legal Theory</source>
          <volume>17</volume>
          (
          <year>2011</year>
          )
          <fpage>1</fpage>
          -
          <lpage>33</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>J.</given-names>
            <surname>Horty</surname>
          </string-name>
          ,
          <article-title>Reasoning with dimensions and magnitudes</article-title>
          ,
          <source>Artificial Intelligence and Law</source>
          <volume>27</volume>
          (
          <year>2019</year>
          )
          <fpage>309</fpage>
          -
          <lpage>345</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>P.</given-names>
            <surname>Dung</surname>
          </string-name>
          ,
          <article-title>On the acceptability of arguments and its fundamental role in nonmonotonic reasoning, logic programming and n-person games</article-title>
          ,
          <source>Artificial Intelligence</source>
          <volume>77</volume>
          (
          <year>1995</year>
          )
          <fpage>321</fpage>
          -
          <lpage>357</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>V.</given-names>
            <surname>Aleven</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. D.</given-names>
            <surname>Ashley</surname>
          </string-name>
          ,
          <article-title>Evaluating a learning environment for case-based argumentation skills</article-title>
          ,
          <source>in: Proceedings of the 6th international conference on Artificial intelligence and law</source>
          ,
          <year>1997</year>
          , pp.
          <fpage>170</fpage>
          -
          <lpage>179</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>W. van Woerkom</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Grossi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Prakken</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Verheij</surname>
          </string-name>
          ,
          <article-title>Landmarks in case-based reasoning: From theory to data</article-title>
          ,
          <source>in: Proceedings of the First International Conference on Hybrid Human-Machine Intelligence, Frontiers of AI</source>
          , IOS Press,
          <year>2022</year>
          , p.
          <fpage>tbd</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>K.</given-names>
            <surname>Čyras</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Birch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Toni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Dulay</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Turvey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Greenberg</surname>
          </string-name>
          , T. Hapuarachchi,
          <article-title>Explanations by arbitrated argumentative dispute</article-title>
          ,
          <source>Expert Systems with Applications</source>
          <volume>127</volume>
          (
          <year>2019</year>
          )
          <fpage>141</fpage>
          -
          <lpage>156</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>V.</given-names>
            <surname>Aleven</surname>
          </string-name>
          ,
          <article-title>Using background knowledge in case-based legal reasoning: A computational model and an intelligent learning environment</article-title>
          ,
          <source>Artificial Intelligence</source>
          <volume>150</volume>
          (
          <year>2003</year>
          )
          <fpage>183</fpage>
          -
          <lpage>237</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>