<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>On Diagnostic Arguments in Abstract Argumentation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Jordan Robinson</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Katie Atkinson</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Simon Maskell</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Chris Reed</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Dundee</institution>
          ,
          <addr-line>Scotland</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>University of Liverpool</institution>
          ,
          <country country="UK">United Kingdom</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2024</year>
      </pub-date>
      <fpage>27</fpage>
      <lpage>40</lpage>
      <abstract>
        <p>We are interested in employing argumentation for intelligence analysis due to its obvious potential for real-world impact. The Analysis of Competing Hypotheses, a well-known technique for multiple hypothesis evaluation from the intelligence community, includes sensitivity analysis as a task which helps analysts identify diagnostic information. We draw upon this notion of sensitivity analysis in this paper to set out a novel algorithm, called the Diagnostic Argument Identifier, that is able to identify diagnostic arguments. We employ a labelling-based approach to compute acceptance probabilities between partitions of arguments, which are then used to calculate the mutual information between the labels of each partition before and after the sequential removal of each argument from a framework. We present the results from one experiment on an abstract framework to assess whether our method can identify diagnostic arguments, and thus aid intelligence analysts. We argue that our algorithmic approach systematises and, therefore, reduces the subjectivity of sensitivity analysis; thus, yielding benefits to intelligence analysts - or any other expert working within a decision or deliberation setting - who need to objectively reevaluate the dependence of their set of conclusions on observed data present within an analysis.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Abstract argumentation</kwd>
        <kwd>probabilistic argumentation</kwd>
        <kwd>information theory</kwd>
        <kwd>intelligence analysis</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Our work is conducted within the setting of intelligence analysis where one of the most
wellknown techniques within the intelligence community is the Analysis of Competing Hypotheses
(ACH) [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. The ACH provides a systematic approach to hypothesis evaluation. In short, the
procedure starts with: hypothesis generation; listing of evidence; instantiation of a matrix with
hypotheses (as column headers) and information, evidence and assumptions (as row headers);
analysis of the consistency or otherwise of each row entry with each hypothesis; refinement of
the matrix, removing or combining superfluous or overlapping hypotheses, respectively; drawing
tentatively-held conclusions; sensitivity analysis to identify diagnostic row entries; and finally,
reporting the probability of hypotheses and diagnostic row entries to stakeholders.
      </p>
      <p>Sensitivity analysis forces intelligence analysts to think diagnostically in that they must
establish whether the likelihood of their conclusions changes after the removal of a row entry.
A diagnostic data point is one where its removal from the analysis changes the conclusions
drawn. Although the ACH does its best to formalise hypothesis evaluation, sensitivity analysis is
a challenging and subjective task because it is difficult to remove an observed data point and act
as though it never existed when reevaluating.</p>
      <p>
        To this end, we make two contributions. The first is an evaluation-based approach to
probabilistic argumentation which uses the set of labellings discovered by a semantics to calculate
joint and marginal argument acceptance probabilities of partitions of arguments and their labels.
Second, we introduce the Diagnostic Argument Identifier (DAI), a novel algorithm, which applies
the equations from the first contribution to quantify diagnosticity scores of arguments within an
argumentation framework (AF), measuring the change in evaluation after the sequential removal
of each argument from a framework. Our algorithm emulates the task of sensitivity analysis and
should alleviate the reliance on human effort, through use of an algorithmic approach. In [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ],
we presented an application which executes the DAI and visualises the results output from the
algorithm. The goal of this paper is to introduce the formal details underpinning the application.
      </p>
      <p>The remainder of this paper is structured as follows. We begin by providing a brief overview of
abstract argumentation in Section 2. In Section 3, we present our two main contributions: the joint
and marginal argument acceptance probability equations and the DAI, which is accompanied by
the relevant pseudo-code in Section 3.4. We present and discuss a result from a single experiment
on an abstract example in Section 4. In Section 5, we relate the DAI to others’ work, and Section
6 concludes with some avenues for future research.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Preliminaries</title>
      <p>
        We consider (finite) Dung AFs (A , R) containing a set of arguments A and attack relations
R ✓ A ⇥ A [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. For an AF G with arguments {a, b} ✓ A , we say that a attacks b if and only
if (a, b) 2 R. The set of arguments that attack the argument a 2 A is denoted by Att(a) =
{b | (b, a) 2 R}. An admissible set of arguments has to satisfy conflict-freeness and acceptability.
A set of arguments S ✓ A is conflict-free if and only if @(a, b) 2 R where a, b 2 S. An argument
a 2 S ✓ A is acceptable with respect to S if and only if 8 b 2 A such that (b, a) 2 R, then
9 c 2 S such that (c, b) 2 R.
      </p>
      <p>
        We employ a labelling-based approach for this work because the set of labellings resulting
from semantic evaluation enabled the computation of probabilities. A labelling L of a set
of arguments S ✓ A within an AF G is a total function L (S) : S ! LAB that assigns all the
arguments a 2 S to a label l 2 LAB, where LAB = {IN, OUT, UND} in the case of complete
semantics. We only consider complete labellings in this work as they are the foundation upon
which all other semantics can be defined [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. Let G = (A , R) be an AF and L (A ) : A ! LAB
be a labelling function. A labelling is a complete labelling if and only if 8 ai 2 A , it holds that:
1. L (ai) = IN if and only if L (a j) = OUT, 8 a j such that a j 2 Att(ai);
2. L (ai) = OUT if and only if 9 a j such that a j 2 Att(ai) and L (a j) = IN.
      </p>
      <p>As a consequence of a labelling being a total function, arguments that are neither labelled IN nor
OUT are labelled UND.</p>
      <p>Arguments labelled IN, OUT, or UND for one or all labellings are referred to as credulously or
sceptically IN, OUT, or UND, respectively.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Diagnostic Argument Identifier</title>
      <p>We now introduce our main contributions: the labelling-based argument acceptance probability
equations for partitions of argument labels, derived from a set of probability spaces (Sec. 3.1),
and then the DAI, which is an algorithm that is capable of identifying the most critical arguments
within a Dung AF. We explain how to calculate the mutual information (MI) between partitions
of labelling vectors (Sec. 3.2) and show how to conduct sensitivity analysis on an AF (Sec. 3.3).
The section closes with our second contribution, the pseudo-code for the DAI (Sec. 3.4).</p>
      <sec id="sec-3-1">
        <title>3.1. Probability Spaces</title>
        <p>We let G = (A , R) be an AF which contains a set of N arguments A . We order the set of
arguments using a function f (A ) : A ! A, where A = (a1, ..., aN ) is an ordered vector of
arguments. We assume we have a function g(A) : A ! LM which assigns all arguments ai 2 A
to a set of labelling vectors, such that</p>
        <p>M</p>
        <p>LM = {Li}i=1
where M is the number of labelling vectors, and Li is the i-th labelling vector containing N
argument labels, such that Li = (l1, ..., lN ), where l j 2 Li is the label of the argument a j 2 A and
l j 2 LAB (Sec. 2).</p>
        <p>We now partition the set of arguments into two sets, named Af and Ay .</p>
        <p>Definition 3.1. For an AF G = (A , R), let the partitions of A be Af ✓ A and Ay ✓ A , where
Af [ Ay = A and Af \ Ay = 0/, and the dichotomous sets Af c and Ay c are complements such
that Af c = A \ Ay and Ay c = A \ Af .</p>
        <p>Both the sets Af and Ay are mapped to argument vectors through the function f , such that
f (Af ) : Af !
f (Ay ) : Ay !</p>
        <p>Af = {(a1, ..., a|Af |) | 8 ai 2 A</p>
        <p>where ai 62 Ay } and
Ay = {(a1, ..., a|Ay |) | 8 a j 2 A
where a j 62 Af },
respectively, where i 6= j.</p>
        <p>The partitions Af and Ay are mapped to a corresponding set of labelling vectors through the
same function g, such that g(Af ) : Af ! Lf and g(Ay ) : Ay ! Ly , respectively, such that
Mf
Lf = {Lf ,i}i=1 and</p>
        <p>My</p>
        <p>Ly = {Ly ,i}i=1
where Mf  M, My  M, Lf ✓ LM, Ly ✓
for the partitions Af and Ay , respectively.</p>
        <sec id="sec-3-1-1">
          <title>LM, and Lf ,i and Ly ,i are the i-th labelling vectors</title>
          <p>Example 3.1. Consider a Dung AF G with arguments A = {p, q, r, s, t} and relations R =
{(q, p), (r, q), (s, q), (r, s), (s, r), (r, t), (t, r), (s, t), (t, s)} (Fig. 1a). Evaluating the AF G under
complete semantics produces four distinct labellings (Tab. 1). Following Def. 3.1, we partition
the set of arguments A into dichotomous sets, where Af = {p, q} and Ay = {r, s, t}. Using the
(1)
(2)
(3)
(4)
(5)
p
t
p
t
function f , we map the sets of arguments A , Af and Ay to the argument vectors A = (p, q, r, s,t),
Af = (p, q) and Ay = (r, s,t), which enables the creation of sets of labelling vectors LM, Lf
and Ly , as shown in Eqs. 6, 7 and 8, respectively.</p>
          <p>We let (W f , Ff , Pf ) and (W y , Fy , Py ) be probability spaces, where W f and W y are sample
spaces, Ff and Fy are event spaces, and Pf and Pf are functions such that Pf : Ff ! (0, 1] and
Py : Fy ! (0, 1], respectively. We consider two random variables Xf and Xy that are real-valued
measurable functions Xf : W f ! R and Xy : W y ! R that map results from from the sample
spaces W f and W y to numerical values; thus, modelling a random experiment which, in our case,
is the resulting set of labellings output from semantic evaluation of an AF.</p>
          <p>Definition 3.2. (Random Vector). Let (W , F , P) be a probability space where the random
vector X : W ! R is a measurable function. The random vector X contains random variables
X = (Xf , Xy ) defined on two probability spaces Xf : W f ! R and Xy : W y ! R.</p>
          <p>Before semantic evaluation and given the constraints of argumentation, the number of possible
elements (or labelling vectors) in the sample space W is the number of unique combinations of
argument labels, such that</p>
          <p>|W | = |LAB|N
where |W | is the number of potential elements in the sample space W , N is the number of
arguments, and LAB is defined in Sec. 2. After semantic evaluation of an AF, the set of labelling
vectors in the sample space is reduced to a set of M labelling vectors, such that W = LM.</p>
          <p>The probability spaces (W f , Ff , Pf ) and (W y , Fy , Py ) are measurable spaces, where Ff ⌦ Fy
is the smallest s -field of potential subsets of W f ⇥ W y ◆ W , containing all sets of the form
Lf ,i ⇥ Ly , j where Lf ,i 2 Ff and Ly , j 2 Fy , such that Ff ⌦ Fy is the product s -field. Thus,
the event space F is a s -algebra containing the powerset of all elements in the product space
Ff ⌦ Fy , including the empty set and the set of all events. However, after semantic evaluation,
the set of combinations of labelling vectors for each partition is, again, reduced to sets of realised
events such that Lf ⌦ Ly ⇢ Ff ⌦ Fy is the smallest subset of events from which we can
calculate non-zero probabilities, where Lf and Ly are shown in Eqs. 4 and 5, respectively.</p>
          <p>The observation of the i-th outcome is denoted x(i) 2 W which is the i-th labelling vector</p>
          <p>LM, such that x(i) = (xf(i), xy(i)) = (Lf ,i, Ly ,i) 2 W . Similarly, we can observe outcomes
x(i) 2
from different labellings, where xf(i), xy( j) 2 W are the i-th and j-th labelling vector for the partitions
Af and Ay , respectively, where i 6= j. With a slight abuse of notation, the probability of Af and
Ay ’s labelling vectors being contained in the i-th and j-th outcome in the sample space is denoted
as P(xf(i)) or P(xy( j)) which refers to P(Xf = xf(i)) or P(Xy = xy( j)), respectively.</p>
          <p>The joint and marginal probabilities of argument labels in each segment are the only
probabilities concern our calculation of the MI for each combination of Af and Ay .
(9)
(10)</p>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>3.1.1. Joint Probability</title>
        <p>First, the joint probability of the labels of arguments in each partition measured across the i-th
and j-th outcome in the sample space W is computed using Eq. 10.</p>
        <p>P(xf(i), xy( j)) = 1 M</p>
        <p>Â Ix(i)=xf(k);xy( j)=xy(k)</p>
        <p>M k=1 f
where IA is unity if A is true and zero otherwise.</p>
        <p>Due to the distinct nature of labellings, there will only be one labelling vector in the sample
space W that contains the same arrangement of argument labels as the labels for arguments in
each partition. Thus, there will be M pairs of labellings vectors which produce non-zero joint
probabilities across the product space of realised argument labels in each segment. Consider
a list M containing |M | distinct i- j pairs, where |M | = Mf ⇥ My , from the observed product
space Lf ⌦ Ly , such that we compute |M | joint probabilities. Each joint probability across the
product space is the reciprocal of the number of labellings, if and only if both observed outcomes
feature in the same labelling vector across the sample space W , and for any other i- j pair the
probability is zero, as shown in Eq. 11.</p>
        <p>P(xf(i), xy( j)) =
( M1 , iff 9 m 2 { 1, . . . , M} such that i = i(m) and j = j(m)
0, otherwise.</p>
        <p>(11)</p>
        <p>It follows that the list M contains M i- j pairs of labelling vectors with a joint probability
greater than zero, corresponding to the number of times that the sample space W contained those
distinct segments of argument labels, or outcomes x(m) = (xf(m), xy(m)) 2 W .</p>
        <p>Example 3.2. Continuing our running example, we employ the sets of labelling vectors Lf
and Ly , presented in Eqs. 7 and 8, to compute the joint probability of labels for arguments in
each partition. The product space Lf ⌦ Ly contains |M | = Mf ⇥ My combinations of potential
labelling vectors based on the unique vectors in each partition, where |M | = 12 in this example.
Using both Eqs. 10 and 11, it is easy to see that the first, second and last i- j pair from the product
space Lf ⌦ Ly , shown in Eq. 12, is equal to 14 , which is M1 in this example. The joint probability
of the third i- j pair in Eq. 12 is equal to zero in both Eqs. 10 and 11 because that event was not
observed as a labelling vector in the sample space W (see Tab. 1).
(12)
(13)
(14)</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.1.2. Marginal Probability</title>
        <p>The second probability that we wish to measure is the marginal probability of each unique
labelling vector in the sets Lf and Ly , which turns out to be a normalised count across a vector
subspace, counting how many times Xf = xf(i) and Xy = xy(i) occurred in the sample space W , as
shown in Eqs. 13 and 14, respectively.</p>
        <p>P(xf(i)) =
Example 3.3. Consider again the running example from Fig. 1a and the partition Af with its set
of distinct labelling vectors Lf , as presented in Eq. 7. Using Eq. 13, there are three events within
the space Lf which we can calculate marginal probabilities for. The marginal probability of the
first Lf ,1 2 W , second Lf ,2 2 W and third event Lf ,3 2 W are equal to 12 , 41 and 14 , respectively.</p>
      </sec>
      <sec id="sec-3-4">
        <title>3.2. Mutual Information</title>
        <p>
          We wish to quantify the amount of information in the initial AF and the extent to which removing
an argument affects the distribution of labels between partitions of other arguments. We measure
this change through the computation of the MI between partitions of argument labels before and
after the removal of an argument. The MI is a symmetric function that quantifies the conditional
dependence between two random variables and is able to determine the amount of information
communicated, on average, about one random variable through observation of another [
          <xref ref-type="bibr" rid="ref5 ref6">5, 6</xref>
          ]. The
MI between pairs of argument labels across the realised product space Lf ⌦
Ly are defined Eqs. 4 and 5, respectively, is shown in Eq. 15.
        </p>
        <sec id="sec-3-4-1">
          <title>Ly , where Lf and</title>
          <p>The units of MI depend on the base of the logarithm used in the calculation. For the purposes
of this paper, information will be measured in bits which is the logarithm to the base of two.</p>
          <p>We simplify the MI calculation presented in Eq. 15 by conducting one summation over |M |
pairs of outcomes xf(i) and xy( j) which feature in the realised product space Lf ⌦ Ly , where
|M | = Mf ⇥ My , as shown in Eq. 16.</p>
          <p>I(Xf ; Xy ) = Â Â
xf(i)2 Lf xy( j)2 Ly
I(Xf ; Xy ) =</p>
          <p>Â
(xf(i),xy( j))2 Lf ⌦ Ly
0
0</p>
          <p>P(xf(i), xy( j)) 1
P(xf(i))P(xy( j)) A</p>
          <p>P(xf(i), xy( j)) 1
P(xf(i))P(xy( j)) A</p>
          <p>We note that many of the joint probabilities in this sum will be zero for i- j pairs of outcomes in
the product space that do not feature in the sample space. So, the only non-zero contributions to
the MI summation will be i- j pairs that are in both the observed sample space W and the product
space Lf ⌦ Ly . Thus, we restrict the summation to only include outcomes xf(i) and xy( j) that have
a joint probability greater than zero, as in Eq. 17.</p>
          <p>I(Xf ; Xy ) =</p>
          <p>Â
(xf(i),xy( j))2 Lf ⌦ Ly ,P(xf(i),xy( j))&gt;0
0</p>
          <p>P(xf(i), xy( j)) 1
P(xf(i))P(xy( j)) A</p>
          <p>In light of this constraint and using Eq. 11, it is easy to see that the only non-zero addends
to the MI will be from contributions where the joint probability of pairs of outcomes from the
product space feature in the same labelling from the sample space. For the sake of computational
efficiency, we substitute Eq. 11 into Eq. 17 so that the MI calculation is, therefore, reduced to a
single sum over the set of labellings for our problem setting, such that</p>
          <p>M 1
I(Xf ; Xy ) = x(mÂ)2 W M log @
0</p>
          <p>1
M P(xf(m))P(xy(m)) A
1
where x(m) = (xf(m), xy(m)) and W ⇢</p>
          <p>Lf ⌦</p>
        </sec>
        <sec id="sec-3-4-2">
          <title>Ly is the space containing non-zero joint probabilities.</title>
          <p>The MI is a symmetric function which means there will be a total of 22A distinct combinations
of Af and Ay which produce unique MI calculations. We, therefore, compute the MI between
divisions of arguments for up to half the powerset to completely explore the distribution of
information communicated across the sets of labelling vectors within the initial AF.
Example 3.4. Turning back to our running example and remembering that Af = (p, q) and Ay =
(r, s, t). The MI between each segment’s set of labelling vectors was found to be I(Xf ; Xy ) =
1.5 bits (1 d.p.), under complete semantics. We see that observing the labels of, say, the arguments
in Af tells us 1.5 bits of information about the labels of arguments in Ay , and vice versa.
(15)
(16)
(17)
(18)</p>
        </sec>
      </sec>
      <sec id="sec-3-5">
        <title>3.3. Sensitivity Analysis</title>
        <p>is the set of sensitive relations and R
removed, where Ra ✓</p>
        <p>R.</p>
        <p>Now that we have explained how to determine the MI between segments of labelling vectors
within the initial AF, we explain how to conduct sensitivity analysis. To start this task, we
sequentially remove each argument a 2 A and the relations containing that argument from the
initial AF – creating a sensitive AF G a – which we evaluate using the same semantics chosen
earlier, and compute the MI between divisions of labelling vectors in the sensitive AF.
Definition 3.3. For an AF G = (A , R) undergoing sensitivity analysis, we refer to G a =
(A a, Ra) as a sensitive AF which does not include the argument a1, the argument of
interest, where A a = A \ a such that A a
⇢</p>
        <p>A and a
26</p>
        <p>A a. For a relation r 2
a
2 r features in that relation, then it is removed from the set of relations such that Ra = R \ R</p>
        <sec id="sec-3-5-1">
          <title>R, if the argument</title>
          <p>= { r | 8 r 2</p>
        </sec>
        <sec id="sec-3-5-2">
          <title>R where a</title>
          <p>2 r} are the relations to be
Notation 3.1. When we say Af a or Ay a, we are referring to the two partitions of A a within the
sensitive AF G a that obey Def. 3.1, however they use the set of sensitive arguments A a, as
outlined in Def. 3.3, instead of the set of all arguments A in the initial AF.
a set of labelling vectors through the function g(Aa) : Aa !
LMa , such that
Again, we order the set of sensitive arguments A a using a function f (A a) : A a</p>
        </sec>
        <sec id="sec-3-5-3">
          <title>Aa, such</title>
          <p>that A
a = (a1, ..., aN 1) is a vector of arguments. We map the vector of sensitive arguments Aa to
!
a Ma</p>
          <p>LMa = {Li }i=1
where Ma is the number of labelling vectors output from semantic evaluation of the sensitive
AF G a, and Lia is the i-th labelling vector containing N
1 argument labels, such that Lia =
(l1a, ..., lNa 1), where l aj 2 Lia is the label of the argument a j 2 Aa and l aj 2 LAB (Sec. 2).</p>
          <p>The sets Af a and Ay a are mapped to the argument vectors Afa and Afa through
f (Af a) : Af a !
f (Ay a) : Ay a !</p>
          <p>A
A
fa = {(a1, ..., a|Af a|) | 8 ai 2 A a where ai 62 Ay a} and
ya = {(a1, ..., a|Ay a|) | 8 a j 2 A a where a j 62 Af a},
respectively, where i 6= j.</p>
          <p>The sets of distinct labelling vectors corresponding to the partitions Af and Ay are found using
g(Afa ) : Afa !
g(Aya ) : Aya !</p>
          <p>Lf a = {Lfa ,i}i=f 1 and
Ly a = {Lya ,i}i=1</p>
          <p>Ma
Mya
where Mfa </p>
          <p>Ma, Mya </p>
          <p>Ma, Lf a ✓
for the partitions Afa and Aya , respectively.</p>
          <p>LMa , Ly a ✓</p>
          <p>LMa , and Lfa ,i and Lya ,i are the i-th labelling vectors</p>
          <p>Using the same semantics as earlier, we evaluate the sensitive AF to observe two sets of
labelling vectors for the arguments in the vectors A
fa and Aya ; thus, enabling us to compute
argument acceptance probabilities for partitions of arguments within the sensitive AF.
1It is important for the reader to note that this paper only considers the removal of one argument from an initial AF
while conducting sensitivity analysis. However, we note this approach could be extended to remove more than one
argument from an initial AF to understand how this affects the results output from sensitivity analysis.
(19)
(20)
(21)
(22)
(23)
Notation 3.2. Let G a be a sensitive AF with the argument a removed (Def. 3.3) and Afa and Aya
be two sensitive argument vectors, mapped from the sets Af a and Ay a, which follow Def. 3.1.
The random vector X˜ = (X˜ f , X˜ y ) is a measurable function from a probability space (W a, F a, Pa)
where X˜ maps elements from the sensitive sample space W a = LMa to events Lf a ⌦ Ly a ⇢ F a.
Example 3.5. Back to our running example. We present an MI calculation for the sensitive AF
G t . To start, we employ Def. 3.3 to create a sensitive AF G t with arguments A t = {p, q, r, s} and
relations Rt = {(q, p), (r, q), (s, q), (r, s), (s, r)} (Fig. 1b). We let Af t = {p, q} and Ay t = {r, s}
because that is the same as the original partitions (i.e., Af and Ay in Ex. 3.1) where neither Ay t
nor Ay t contain the argument t. We map the dichotomous sets of arguments Af t and Ay t to the
argument vectors Atf = (p, q) and Aty = (r, s), respectively, through the function f . We, again,
use complete semantics to evaluate G t to observe the set labellings for each partition (Tab. 2),
allowing the realisation and mapping of each argument in Atf and Aty to their respective sets of
labelling vectors Lf t and Ly t , through the function g. The MI I(X˜ f ; X˜ y ) between the labelling
vectors for this partition of sensitive arguments is equal to 0.9 bits (1 d.p.).</p>
          <p>For a sensitive AF G a, the total number of combinations of Af a and Ay a that produce unique
MI calculations is equal to 2A2 a , which is equal to half the number of unique MI results (i.e., 22A )
that we calculate in the initial AF G .</p>
          <p>Once we have computed the set of distinct MI results using half the possible combinations of
Af and Ay in the initial AF and Af a and Ay a in the sensitive one, we can quantify a diagnosticity
score, which describes how much information was lost or gained between either Af and Ay ’s set
of labelling vectors after the removal of the sensitive argument, as stated in Def. 3.4.
Definition 3.4. The diagnosticity score is defined as the change in MI before (calculated using
the labelling vectors Lf and Ly for all arguments in Af and Ay from G ) and after (calculated
using Lf a and Ly a for all arguments in Af a and Ay a from G a) the removal of the argument of
interest. Eq. 24 computes the diagnosticity score when the argument of interest a was removed
from either Af or Ay .</p>
          <p>D (Af , Ay ; Af a _ Ay a) = I(Xf ; Xy )</p>
          <p>I(X˜ f ; X˜ y )
(24)</p>
          <p>An interesting point to note about Eq. 24 is the sign. A positive diagnosticity score infers
that there was more information communicated, on average, between partitions in the initial AF
G , whereas a score below zero indicates that there was more information transferred between
partitions of argument labels within the sensitive AF G a.</p>
          <p>Example 3.6. Coming back to our running example for the last time, we consider the impact
that removing the argument t had on the acceptability of arguments within each partition in the
initial and sensitive AFs. We employed the set of labelling vectors of Af and Ay to calculate the
MI between Xf and Xy in the initial AF (Ex. 3.4), and used the set of labelling vectors of Atf
and Aty to compute the MI between X˜ f and X˜ y in the sensitive AF G t (Ex. 3.5). We now show
the use Def. 3.4 by presenting the diagnosticity score D (Af , Ay ; Ay t ) for the aforementioned
partitions, which turns out to be 0.6 bits (1 d.p.).</p>
        </sec>
      </sec>
      <sec id="sec-3-6">
        <title>3.4. Pseudo-code for the Diagnostic Argument Identifier</title>
        <p>We present the pseudo-code for the DAI in Algorithm 1. The algorithm takes as input an AF
G and a semantics S capable of producing more than one labelling, and returns a diagnosticity
vector D[A ][ 22A ], which contains 22A diagnosticity scores for every argument in the graph.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experimental Evaluation</title>
      <p>We have introduced a novel algorithm that uses an evaluation-based approach to quantify how
much the removal of an argument changes the semantic evaluation of an AF (Algo. 1). To evaluate
the effectiveness of our method, we now present the results of an experiment conducted using the
AF in our running example (Fig. 1). One of the reasons we chose to use the AF presented in Fig.
1 was because we assumed it possessed a similar topology to graphs found within an intelligence
setting (i.e., many symmetric attacks as a result of conflicting information).</p>
      <p>Results and Discussion. We computed 22A (i.e., 16) diagnosticity scores, corresponding to the
total number of segments of labelling vectors, for all arguments in the running example (Fig. 2).</p>
      <p>Looking to Fig. 3, we present a violin plot which groups together the individual diagnosticity
scores in Fig. 2 to show the distribution of change in MI for partitions of labelling vectors after
the sequential removal of each argument in the AF G within the running example. We also
included the median and mean diagnosticity scores, as well as the average absolute values, to
better understand and compare the distribution and sign (i.e., positive or negative) of scores for
all the removed arguments.</p>
      <p>I(Xf ; Xy ) += M1 log M P(xf(m1))P(xy(m))
m-th labelling vector of the initial AF (Eq. 18).</p>
      <p>MIs
{ Af , Ay , I(Xf ; Xy )}
✓</p>
      <p>◆
for a 2 A do</p>
      <p>G a = (A a, Ra)
L (A a)
Aa = f (A a) : A a ! Aa
W a = g(A) : A ! LMa
Ma = |LMa |
for Af , Ay , I(Af ; Ay , ) 2 MIs do
if a 2 Af then</p>
      <p>Af a = Af \ a; Ay a = Ay
else if a 2 Ay then</p>
      <p>Ay a = Ay \ a; Af a = Af</p>
      <p>. Append the result to the array of initial MI calculations.
. Create Af a and Ay a by removing a from Af (Defs. 3.1 &amp; 3.3).</p>
      <p>. Create Af a and Ay a by removing a from Ay (Defs. 3.1 &amp; 3.3).</p>
      <sec id="sec-4-1">
        <title>Algorithm 1 The Diagnostic Argument Identifier</title>
        <p>Input: An AF G = (A , R) and a semantics S</p>
        <p>Output: D[A ][ 22A ]
1: function DAI(G , S)
2: L (A )
3: A = f (A ) : A ! A
4: W = g(A) : A ! LM
5: M = |LM |
6: MIs[ 22A ]</p>
        <p>D[A ][ 22A ]</p>
        <p>22A do
for Af 2</p>
        <p>Ay = A \ Af
Af = f (Af ) : Af ! Af ;
Lf = g(Af ) : Af ! Lf ;
I(Xf ; Xy ) = 0
for (xf(m), xy(m)) 2 W do</p>
        <p>P(xf(m)); P(xy(m))
7:</p>
        <p>The argument t was arguably the most diagnostic as it had the largest median, mean and
average absolute diagnosticity scores (Fig. 3). This result is intuitive because t has symmetric
attacks between the arguments r and s such that removing t reduces the number of complete
labellings from four (Tab. 1) to three (Tab. 2).</p>
        <p>The argument q had the second largest set of diagnosticity scores (Fig. 3). Even though q only
attacked one argument (Fig. 1) and its removal does not reduce the number of labellings (Tab.
1), the conditional dependence between t and q meant that q produced greater median and mean
diagnosticity scores. An interesting point to note is that the removal q from the initial AF would
change the label of the argument p so that it was sceptically labelled IN within the sensitive AF
G q. Thus, for the sensitive AF G q, the MI scores would equal zero for partitions where p was
the only argument in a segment (i.e., Af q = {p} or Ay q = {p}) because the MI is always zero for
partitions that only contain arguments which are sceptically labelled.</p>
        <p>While the arguments r and s are not the most diagnostic, their removal from the initial AF
results in the same distribution of change in MI (Fig. 3), which can be attributed to the symmetry
between them in the initial graph. The median change in MI was below zero (Fig. 3) which
indicates that, on average, more information was communicated between segments of labelling
vectors in the sensitive AFs G r and G s after the removal of r and s, respectively.</p>
        <p>The preliminary result presented in this paper is a first attempt within the literature to combine
abstract argumentation and a technique from the information-theoretic literature for sensitivity
analysis. We were able to quantify the sensitivity, dependence and robustness of an AF’s set of
conclusions based upon the arguments it was comprised of.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Related Work</title>
      <p>
        The prior literature covers work on probabilistic argumentation, namely the epistemic [
        <xref ref-type="bibr" rid="ref7 ref8">7, 8</xref>
        ]
and constellation [
        <xref ref-type="bibr" rid="ref10 ref9">9, 10</xref>
        ] approaches. Our method of probability computation differs from both
approaches because we do not rely on probability functions to evaluate AFs, as is the case with
the epistemic approach. Nor do we iterate through all the permutations of a graph’s topology,
counting the number of times where a user-chosen set of arguments features in an induced AF’s
set of extensions, as per the constellation approach. We note one proposal that employed the
frequency of individual argument labels to compute marginal probabilities which were then used
in combination with a Markov network to semantically evaluate an AF [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. Our formalised
probability calculations (see Eqs. 10, 11, 13 and 14) are the first attempt within the literature to
employ partitions of arguments and their labels to compute joint and marginal probabilities.
      </p>
      <p>
        To the best of our knowledge, there has only been two proposals before this one which use
argumentation for sensitivity analysis. The first employs argumentation and Markov random
fields to quantify the sensitivity of items of information [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. The second proposal evaluates the
sensitivity of initial weights assigned to arguments within the context of inverse argumentation,
where they consider whether changes in an argument’s weight affects the acceptability degree of
other arguments, computed using gradual semantics [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. The algorithm presented in this paper
combines Dung’s original abstract AF and semantics with a set of novel probability calculations
to compute the change in MI before and after the removal of an argument of interest.
      </p>
      <p>
        Aside from the two above works on sensitivity analysis, we consider the argument strength
literature to be the most closely related to the DAI. Strength has been represented by arbitrarily
assigning a weight to an argument [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] or using ranking-based semantics [
        <xref ref-type="bibr" rid="ref15 ref16 ref17 ref18 ref19">15, 16, 17, 18, 19</xref>
        ].
However, our approach differs from all of the above because we neither choose an argument’s
weight nor do we not alter the well-known semantics to produce a preorder on the set of arguments.
      </p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusions and Future Work</title>
      <p>The novel algorithm presented in this paper is one of the first attempts within the literature to use
of argumentation for sensitivity analysis. We use the labellings output from the evaluation of an
AF to compute joint and marginal argument acceptance probabilities which are employed in MI
calculations before and after the removal of an argument. We argue that the diagnosticity scores
output by the DAI provide a holistic quantification of the sensitivity, dependence and robustness
of an AF’s evaluation. We contend that such a tool would provide benefit to intelligence analysts
by algorithmically identifying sensitive arguments found within an analysis.</p>
      <p>Future work could involve intelligence analysts, comparing the outputs from the DAI with the
outcome from sensitivity analysis conducted by a set of analysts. And, of course, extending the
DAI to include more flavoursome AFs is an obvious extension. Another avenue for future work
could be to investigate different domains in which the DAI could be applied. For instance, the
DAI could aid in decision and deliberation problems, allowing users to utilise the rational logic of
argumentation and probability theory to identify and focus on crucial arguments with their task,
or by providing reasoners with an indication of which arguments are the most important to attack
when analysing a debate. Adding or removing multiple arguments is an interesting idea to pursue
as it might better explain the diagnosticity of sets of arguments. Finally, one could argue that the
DAI, in its current state, is computationally expensive. Cheaper approaches, such as counting the
number of attacks from and to an argument or whether an argument is in a dense area of an AF,
should be explored to identify whether they are capable of producing the same results as the DAI.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgements</title>
      <p>This work was funded by both the Defence Science and Technology Laboratory and the United
Kingdom’s Engineering and Physical Sciences Research Council (Grant number: EP/S023445/1).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>R.</given-names>
            <surname>Heuer</surname>
          </string-name>
          , Psychology of Intelligence Analysis,
          <source>Center for the Study of Intelligence</source>
          ,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>J.</given-names>
            <surname>Robinson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Atkinson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Maskell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Reed</surname>
          </string-name>
          ,
          <article-title>Identifying diagnostic arguments in abstract argumentation</article-title>
          ,
          <source>in: Proc. of COMMA</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>P. M.</given-names>
            <surname>Dung</surname>
          </string-name>
          ,
          <article-title>On the acceptability of arguments and its fundamental role in nonmonotonic reasoning, logic programming and n-person games</article-title>
          ,
          <source>Artificial Intelligence</source>
          <volume>77</volume>
          (
          <year>1995</year>
          )
          <fpage>321</fpage>
          -
          <lpage>357</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>M. W. A.</given-names>
            <surname>Caminada</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. M.</given-names>
            <surname>Gabbay</surname>
          </string-name>
          ,
          <article-title>A logical account of formal argumentation</article-title>
          ,
          <source>Studia Logica</source>
          <volume>93</volume>
          (
          <year>2009</year>
          )
          <fpage>109</fpage>
          -
          <lpage>145</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>C. E.</given-names>
            <surname>Shannon</surname>
          </string-name>
          ,
          <source>A mathematical theory of communication</source>
          ,
          <source>The Bell System Technical Journal</source>
          <volume>27</volume>
          (
          <year>1948</year>
          )
          <fpage>379</fpage>
          -
          <lpage>423</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>T.</given-names>
            <surname>Cover</surname>
          </string-name>
          , J. Thomas, Elements of Information Theory, John Wiley &amp; Sons,
          <year>2005</year>
          , pp.
          <fpage>13</fpage>
          -
          <lpage>55</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>M.</given-names>
            <surname>Thimm</surname>
          </string-name>
          ,
          <article-title>A probabilistic semantics for abstract argumentation</article-title>
          ,
          <source>in: Proc. of ECAI</source>
          , volume
          <volume>242</volume>
          , IOS Press,
          <year>2012</year>
          , pp.
          <fpage>750</fpage>
          -
          <lpage>755</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>M.</given-names>
            <surname>Thimm</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hunter</surname>
          </string-name>
          ,
          <article-title>Probabilistic argumentation with incomplete information</article-title>
          ,
          <source>in: Frontiers in Artificial Intelligence and Applications</source>
          , volume
          <volume>263</volume>
          ,
          <year>2014</year>
          , pp.
          <fpage>1033</fpage>
          -
          <lpage>1034</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>P.</given-names>
            <surname>Dung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Thang</surname>
          </string-name>
          ,
          <article-title>Towards (probabilistic) argumentation for jury-based dispute resolution</article-title>
          ,
          <source>in: Proc. of COMMA</source>
          ,
          <year>2010</year>
          , pp.
          <fpage>171</fpage>
          -
          <lpage>182</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>H.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Oren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Norman</surname>
          </string-name>
          ,
          <article-title>Probabilistic argumentation frameworks</article-title>
          ,
          <source>in: Theories and Applications of Formal Argumentation</source>
          , volume
          <volume>7132</volume>
          ,
          <year>2011</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>16</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>N.</given-names>
            <surname>Potyka</surname>
          </string-name>
          ,
          <article-title>Abstract argumentation with markov networks</article-title>
          ,
          <source>in: Proc. of ECAI</source>
          , volume
          <volume>325</volume>
          , IOS Press,
          <year>2020</year>
          , pp.
          <fpage>865</fpage>
          -
          <lpage>872</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Oren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Sycara</surname>
          </string-name>
          ,
          <article-title>Markov argumentation random fields</article-title>
          ,
          <source>in: Proc. of AAAI Conference on Artificial Intelligence</source>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>N.</given-names>
            <surname>Oren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Yun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Vesic</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. S.</given-names>
            <surname>Baptista</surname>
          </string-name>
          ,
          <article-title>Inverse problems for gradual semantics</article-title>
          ,
          <source>in: International Joint Conference on Artificial Intelligence</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>2719</fpage>
          -
          <lpage>2725</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>L.</given-names>
            <surname>Amgoud</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Cayrol</surname>
          </string-name>
          ,
          <article-title>Integrating preference orderings into argument-based reasoning</article-title>
          ,
          <source>in: Qualitative and Quantitative Practical Reasoning</source>
          , volume
          <volume>1244</volume>
          , Springer,
          <year>1997</year>
          , pp.
          <fpage>159</fpage>
          -
          <lpage>170</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>C.</given-names>
            <surname>Cayrol</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.-C.</surname>
          </string-name>
          Lagasquie-Schiex, Graduality in argumentation,
          <source>J. Artif. Intell. Res. (JAIR) 23</source>
          (
          <year>2005</year>
          )
          <fpage>245</fpage>
          -
          <lpage>297</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>L.</given-names>
            <surname>Amgoud</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Ben-Naim</surname>
          </string-name>
          ,
          <article-title>Ranking-based semantics for argumentation frameworks</article-title>
          ,
          <source>in: Scalable Uncertainty Management</source>
          , volume
          <volume>8078</volume>
          , Springer,
          <year>2013</year>
          , pp.
          <fpage>134</fpage>
          -
          <lpage>147</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>F.</given-names>
            <surname>Pu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Luo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , G. Luo,
          <article-title>Argument ranking with categoriser function</article-title>
          ,
          <source>in: Knowledge Science, Engineering and Management</source>
          , volume
          <volume>8793</volume>
          , Springer,
          <year>2014</year>
          , pp.
          <fpage>290</fpage>
          -
          <lpage>301</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>L.</given-names>
            <surname>Amgoud</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Ben-Naim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Doder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Vesic</surname>
          </string-name>
          ,
          <article-title>Ranking arguments with compensation-based semantics</article-title>
          ,
          <source>in: Proc. of Knowledge Representation and Reasoning</source>
          , AAAI Press,
          <year>2016</year>
          , p.
          <fpage>12</fpage>
          -
          <lpage>21</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>D.</given-names>
            <surname>Grossi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Modgil</surname>
          </string-name>
          ,
          <article-title>On the graded acceptability of arguments in abstract and instantiated argumentation</article-title>
          ,
          <source>Artificial Intelligence</source>
          <volume>275</volume>
          (
          <year>2019</year>
          )
          <fpage>138</fpage>
          -
          <lpage>173</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>