<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>How Much Do I Argue Like You? Towards a Metric on Weighted Argumentation Graphs</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Markus BRENNEIS</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Maike BEHRENDT</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Stefan HARMELING</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Martin MAUVE</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science, University of D u ̈sseldorf</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <fpage>2</fpage>
      <lpage>13</lpage>
      <abstract>
        <p>When exchanging arguments with other people, it is interesting to know who of the others has the most similar opinion to oneself. In this paper, we suggest using weighted argumentation graphs that can model the relative importance of arguments and certainty of statements. We present a pseudometric to calculate the distance between two weighted argumentation graphs, which is useful for applications like recommender systems, consensus building, and finding representatives. We propose a list of desiderata which should be fulfilled by a metric for those applications and prove that our pseudometric fulfills these desiderata.</p>
      </abstract>
      <kwd-group>
        <kwd />
        <kwd>argumentation graphs</kwd>
        <kwd>online argumentation</kwd>
        <kwd>metric</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        In real-world discussions, people exchange arguments on a dedicated issue, such as
improving the course of study [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], the distribution of funds [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], or which party to vote for
at the next general election. In all those cases, participants discuss positions like “there
should be a universal basic income” or “special math courses should be introduced”,
state their pro and contra arguments and attack other people’s arguments.
      </p>
      <p>Each individual participant of an argumentation has a personal view on the
arguments and their relative importance: Users can decide for themselves which arguments
they consider more convincing, thus which arguments they agree to and how much they
agree with a statement. They may consider some positions more important than others.</p>
      <p>Based on those individual views, there are useful applications for measuring the
similarity or distance between two users: Clustering can be used to find representatives
for a group of people with similar argumentation behavior or for finding a consensus.
Another application is opinion polling, where one wants to find out why two persons or
organizations come to different conclusions. What is more, collaborative filtering, which
needs some definition of distance between users, can be used for pre-filtering arguments
in applications like Kialo2.
Brenneis et al. / How Much Do I Argue Like You? Towards a Metric on Weighted Argumentation Graphs 3</p>
      <p>In this paper, we propose solutions to the two main challenges to achieving the goal
of comparing the argumentations of two users: We define weighted argumentation graphs
which are a suitable representation of argumentation covering the mentioned aspects,
including importance of arguments and agreement with statements. Secondly, we suggest
a pseudometric for calculating the distance between two weighted argumentation graphs,
which considers the specific structure of argumentation graphs (e.g., opinions deeper in
a graph are less important). We contribute a list of useful desiderata for a metric which
compares argumentations, and prove that our pseudometric fulfills those properties.</p>
      <p>In the following chapter, we present our definition of weighted argumentation
graphs. The third chapter introduces our pseudometric and desiderta for a useful metric.
Finally, we discuss some limitations of our pseudometric and take a look at related work.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Definition of Weighted Argumentation Graphs</title>
      <p>
        To be able to determine the similarity of real-world argumentations, there has to be a
suitable representation of them. This representation should be able to capture all aspects
mentioned in the introduction, and it should be as simple as possible. Therefore, the
following definition is based on the IBIS model [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], which has been successfully tested
with users without background in argumentation theory using our D-BAS system [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ].
      </p>
      <p>
        For the application purposes described in the introduction, the model should be able
to represent the known opinions and arguments of a person as close as possible. Thus, we
use statements, not arguments, as atomic elements in our definition, which then can be
composed to arguments which can support or attack another statement. Note, though, that
this definition can be translated to classical abstract argumentation frameworks based on
Dung’s definition [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], e.g. using DABASCO [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ].
      </p>
      <p>Let S be the (finite) set of all statements, with the special statement I 2 S. The set
A S n fIg S is a set of arguments. For an argument a = (s1; s2) 2 A, s1 is called
premise, s2 conclusion of a. Let s 2 S, then a!s := f(t; u) 2 A j u = sg is the set of
arguments with conclusion s.</p>
      <p>
        Note that I is excluded to be a premise since it is the issue in the IBIS model. We
refer to I as “personal well-being”, which allows us to interpret an edge like (b; I) in
Figure 1 as “My personal well-being improves, because more wind power plants will be
built.” The premises of arguments with conclusion I are called positions. Positions are
actionable items like “A wall between Austria and Germany should be built,” and play
an important role in real-world argumentation, e.g. decision-making problems [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
Definition 1 (Argumentation Graph). An argumentation graph is a directed, weakly
connected graph G = (S; A), A S n fIg S, where the statements are nodes and the
arguments edges, and there is exactly one I 2 S which has no outgoing edges.
      </p>
      <p>Note that this model does not include different relations for attack and support, as
known from bipolar argumentation frameworks. Whether an argument is supportive are
not, is up to a person’s interpretation of the natural language representation of the
arguments. The purpose of the model is solely to capture the hierarchy of statement, which
we later need for our metric; bipolarity would add unnecessary complexity in this paper.</p>
      <p>Every person can have a personal view with personal attitudes on a common
argumentation graph G, as depicted in G0 in Figure 1. To get an intuition for our next
definia1
a11</p>
      <p>G
a
a2</p>
      <p>I
b
b1</p>
      <p>G0 on G
tion, let us have a look at what we can deduce about Alice’s attitudes from her graph G0:
Alice strongly accepts position b (rating :4) and is slightly against a (rating :2). She
accepts the statement a2 more than statement a1 (rating :5 &gt; rating :3). The
counterargument (a2; a) “No wall should be built, because a wall is expensive.” is far more important
for her than the argument (a1; a) (relative importance :8 &gt; :2).</p>
      <p>Furthermore, it makes sense to sort the positions: She considers building more wind
power plants (b) more important than building a wall (a, relative importance 32 &gt; 13 ).
So when comparing her attitudes with someone else’s attitudes, she would consider a
contrary opinion on b more severe than a different opinion on a. For ordinary statements,
which are not positions, having an importance does not make sense: One cannot say
that “The wall stops illegal immigration of cows” is twice as important as “Cows can
come via Switzerland” (important regarding what?); one can only say that the arguments
regarding building a wall which are built by those statements are of differing importance.</p>
      <p>We will use real numbers to represent those weights and ratings.</p>
      <p>Definition 2 (Weighted Argumentation Graph). Let G = (S; A) be an argumentation
graph. A weighted argumentation graph G0 on G is a quadruple (S; A; r; w) with
functions r and w. r : S ! [ 0:5; 0:5] assigns an agreement score (rating) to every statement,
where negative values mean disagreement, 0 no opinion/don’t care, and positive values
agreement. w : A ! [0; 1] assigns an importance weight to each argument. The value
indicates the importance of that argument relative to other arguments with the same
conclusion. The value 0 means that the argument is not used (i.e. has no relevance), and 1
means that the argument is the only relevant argument for the conclusion. The following
conditions must hold:</p>
      <p>Formula 1 means that the sum of weights of arguments with the same conclusion is
1 if there is an argument with positive weight (cf. (a1; a) and (a2; a) in Figure 1); the sum
is 0 iff no argument for a common conclusion has a weight (cf. w(b1; b) = 0 in Figure 1).
This assures that w represents relative, not absolute importance. To simplify notation, we
write w( ; ) instead of w(( ; )). If the underlying argumentation graph G happens to be
a directed tree, we call G0 a weighted argumentation tree.</p>
      <p>w and r can be represented as matrix or vector, respectively, where undefined values
are set to the default value 0. For the example in Figure 1, one gets:</p>
      <p>Brenneis et al. / How Much Do I Argue Like You? Towards a Metric on Weighted Argumentation Graphs 5
0 0 0 0 0 0 0 01
B0:33 0 0 0 0 0 0C</p>
      <p>BB0:67 0 0 0 0 0 0CC
w = BB 0 0:2 0 0 0 0 0CC ;</p>
      <p>BB 0 0:8 0 0 0 0 0CC
B@ 0 0 0 0 0 0 0CA
0 0 0 0 0 0 0
0 0 1
B 0:2C</p>
      <p>BB 0:4 CC
r = BB 0:3 CC</p>
      <p>BB 0:5 CC
B@ 0 CA
0
(3)</p>
      <p>The entry in row i column j of w is the weight of the argument with premise i and
conclusion j. The first column and first row must refer to I as premise or conclusion,
respectively. Because of Formula 1, the column sum is always 0 or 1.</p>
      <p>If we draw or talk about a weighted argumentation graph, “non-existing” edges a
are edges with w(a) = 0 (argument with no importance), and “non-existing” nodes s are
nodes with r(s) = 0 (neutral statement). An example is G0 shown in Figure 1.</p>
      <p>
        The importance of a position p is represented as weight of the “argument” (p; I)
leading to the “personal well-being” I. An application could obtain weights and ratings
from a user, for example, by asking them to mark statements which are considered more
important, or sorting arguments by relevance, which we do in our deliberate system [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Proposal of a Pseudometric for Weighted Argumentation Graphs</title>
      <p>We now propose a pseudometric for calculating a distance between two weighted
argumentation graphs, and prove several properties we expect of a function which compares
two argumentations. The goal of the metric is to indicated how close the opinions and
used arguments of two persons are, considering graph structure and individual
assessments of importance; we do not want to compare argumentations on abstract levels like
consistency, number of arguments used, or if other person’s arguments are countered.
3.1. The Pseudometric
We define a distance measure of two weighted argumentation graphs G1 = (S; A; r1; w1),
G2 = (S; A; r2; w2) on G = (S; A) as:
dG(G1; G2) = (1</p>
      <p>¥
a) å aikwi1[:;1] r1
i=1
wi2[:;1] r2k1
(4)
where a 2 (0; 1) determines the influence of opinions deeper in the graph: A lower a
emphasizes opinions on statements r(s), a higher value the similarity of the
argumentation underneath a statement s. w[:;1] denotes the first column of the weight matrix w,
ai is the i-th power of the scalar a, wi the i-th power of the square matrix w, and
the Hadamard (entrywise) product. The i-th summand calculates the contribution of the
paths with length i ending at I. We drop the index G if the underlying argumentation
graph is clear from the context.</p>
      <p>The intuition behind this distance measure becomes clearer when rephrasing it for
the special case of argumentation trees (which have no cycles or re-used statements, thus
unique paths from each statement to I). In case of argumentation trees T1 and T2 on T ,
the distance dG is equivalent to:
dT (T1; T2) = (1
a) å adepth(s) r1(s) Õ w1(a) r2(s) Õ w2(a)
s2S a2rI!s a2rI!s
(5)
rs1!s2 is the sequence of all arguments (edges) on the path from s1 to s2 2 S (where
s2 is deeper in the tree). depth(s) = jrI!sj is the length of the path from I to s, i.e. the
number of arguments; depth(I) = 0.</p>
      <p>The terms in the absolute value measure the similarity of the opinions of a statement
s as difference of their ratings, scaled with the product of the “importances” of the
arguments leading to s. Hence, statements which are deeper in the argumentation tree get
a smaller weight, and the overall relevance of an argumentation branch is limited to its
importance.</p>
      <p>To see how the calculation works and that the results match intuition, let us calculate
the distance between the graphs G0 (Figure 1), T2, and T3 (Figure 2) for a = 0:5. We can
expect that T2 is closer to T3 than to G0, because the opinions on the statements a and b
match and only the weights are different. The results confirm the expectation:
d(T2; G0) =(1
a) a1 0:5 0:6
( 0:2)
+ a1 0:5 0:4
0:4</p>
      <p>
        However, dG is not a metric, because dG(G1; G2) can be 0 even if G1 is not equal to
G2: Consider G1 where all statements are agreed to and every argument weight is 0, and
G2 where all statements have a rating of 0. Because weights and ratings are multiplied,
the distance is 0 though the weighted argumentation graphs are different.
3Remember that r1(root(T )) = r2(root(T )) because root(T ) = I, and r(I) := 0.
3.2. Desiderata for a Metric for Weighted Argumentation Trees
Our pseudometric is only one of many possible metrics for the applications described.
We present a list of intuitive desiderata which, as we think, should be fulfilled by any
metric comparing two person’s argumentations, and thus should be considered when
constructing alternative metric proposals. It is, however, hard to capture intuitive
properties in graphs which can contain circular references and re-used statements. Therefore,
we focus on argumentation trees in this section. Field experiments like [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] have also
shown that users seldom create cycles or re-use statements in different branches in real
discussions.
      </p>
      <p>After each desideratum, we prove that our pseudometric (Formula 5) fulfills it for
weighted argumentation trees. We consider those properties important in many
realworld application domains of a metric, albeit not everywhere, as pointed out in Section 5.</p>
      <p>For each desideratum, we indicate why we think it is intuitive. Most desiderata are
followed by a visual example making the choice of variable names clearer. Note that
each tree in the examples is considered to be part of a bigger weighted argumentation
tree, i.e. not all existing nodes and edges are drawn, and irrelevant statement ratings and
argument weights are left out.
d(Tk2 ; Tk3 ).</p>
      <p>Desideratum 1 (Proportionally bigger overlap is better). Consider trees T1; T2; T3, where
T2 is like T1, but uses one additional argument for a statement s, and T3 is like T2, but
uses one additional argument for s. Although T2 and T1 differ in only one argument, and
T3 and T2 differ in only one argument, we expect d(T1; T2) &gt; d(T2; T3) because T2 and T3
have a greater overlap regarding the used arguments.</p>
      <p>More formally: For every statement s in a tree T which only has leaves s1; : : : ; sn
(n &gt; 2) as premises for s with r(s1) = = r(sn) 6= 0 and w(s1; s) = = w(sn; s) and
8a 2 rI!s : w(a) 6= 0, consider the trees Tk, n k &gt; 0, which only contain s1; : : : ; sk.
Then, given k1 &lt; k2 &lt; k3, we want to have kk12 &lt; kk23 =) d(Tk1 ; Tk2 ) &gt; d(Tk2 ; Tk3 ), i.e. if the
relative overlap of the number of arguments used is greater, the distance is smaller.
Likewise, we demand kk12 &gt; kk32 =) d(Tk1 ; Tk2 ) &lt; d(Tk2 ; Tk3 ), and kk12 = kk23 =) d(Tk1 ; Tk2 ) =</p>
      <p>We require 8a 2 rI!s : w(a) 6= 0 (i.e. no argument (edge) with weight 0 along the
path from I to s), because if a user gives an argument a weight of 0, they say the premise
is not related to the conclusion, thus not related to the topic of the discussion. This means
that the user actually does not care about the opinions underneath that argument, which
may be treated as if no opinion has been given.</p>
      <p>In the following example, we have k1 = 1, k2 = 2, and k3 = 3, thus 21 &lt; 23 =)
dT (T1; T2) &gt; dT (T2; T3):</p>
      <p>T1</p>
      <p>s
1
k1
r s1</p>
      <p>T2</p>
      <p>s
k12 k12
r s1 r s2</p>
      <p>T3</p>
      <p>s
k13 k13 k13
r s1 r s2 r s3
Proof. Only the argument weights for (si; s) (namely k1 , k1 , k1 , respectively, as used
1 2 3
in (8) = (9)) are different and contribute to the sum, all other summands are zero. The
common weight products and common values for r are summarized as wr for readability
and are factored out. For (10) &gt; (11), remember that kk1 &lt; kk2 .
2 3
(8)
(9)
(10)
(11)
dT (Tk1 ; Tk2 ) = wr(1
adepth(s0)jwk1 (s0; s)
wk2 (s0; s)j</p>
      <p>The other cases are proven by replacing “&gt;” with “&lt;” or “=”, respectively.
Desideratum 2 (Contrary opinion is worse than no opinion). Consider trees T1, T2, T3,
where all trees are identical, but T1 has no opinion on a statement s, T2 a positive opinion
on s and T3 a negative opinion on s. As we definitely know that T2 and T3 disagree on s,
we want to have d(T2; T3) &gt; d(T1; T2).</p>
      <p>Formally: For any statement s in a tree T with 8a 2 rI!s : w(a) 6= 0, let T + be like
T but with r+(s) = q &gt; 0, T like T with r (s) = p &lt; 0, and T 0 like T with r0(s) = 0.
Then d(T +; T ) &gt; d(T +; T 0).</p>
      <p>T0 0 s</p>
      <p>T+ q s</p>
      <p>T
p s
Proof. The only positive summand is the summand for s (which has rating 0, q or p for
T 0, T +, and T , respectively). The argument weights and the a term are common to all
summands, can be factored out, and are summarized as wa .</p>
      <p>dT (T +; T ) = wa jq
pj = wa (q + p) &gt; wa jq
0j = dT (T +; T 0)
(12)
Desideratum 3 (Deviation in deeper parts has less influence than deviation in higher
parts). Consider the trees T1; T2; T3, where T1 has an argument (sA; s) with no
children and 8a 2 rI!s : w(a) 6= 0, T2 is constructed from T1 by adding a new statement
sB and argument (sB; s) with w2(sB; s) = w2(sA; s) = w1(sA;s) , and T3 is constructed
2
from T1 by adding a new statement sA1 and argument (sA1; sA) with w3(sA1; sA) = 1. If
r(sA) = r(sB) = r(sA1) 6= 0, then we want to have d(T1; T2) &gt; d(T1; T3), because
arguments deeper in the tree should have a smaller influence since we consider them less
important for the overall opinion.</p>
      <p>We require r(sA) = r(sB) = r(sA1) 6= 0 because adding a statement with “don’t care”
opinion should not actually change the distance. We want equality because this
desideratum should cover only differences in the depth of the statements, not their rating.</p>
      <p>T1</p>
      <p>s
w
r sA</p>
      <p>T2</p>
      <p>s
w2 w2
r sA r sB</p>
      <p>T3</p>
      <p>s
w
r sA</p>
      <p>Brenneis et al. / How Much Do I Argue Like You? Towards a Metric on Weighted Argumentation Graphs 9
Proof. The only differences are the summands including s, sA and sA1. Let r := r(sA) =
r(sB) = r(sA1) 6= 0 and w := w1(sA; s). As before, we summarize common values for
weights and a as wa .</p>
      <p>dT (T1; T2) = wa
rw
&gt; wa ajrwj = wa (a jr 0 w
r 1 wj) = dT (T1; T3)
(13)
(14)
Desideratum 4 (Influence of deeper parts depends on weights in higher parts). Consider
trees T1; T2; T3, where all trees are identical, have statements sA and sB with conclusion s,
and w(sA; s) &gt; w(sB; s), but T1 has an additional argument for sA and T3 has an additional
argument for sB. Although the difference in both cases is only one argument, we expect
d(T1; T2) &gt; d(T3; T2) because (sA; s) has a larger weight.</p>
      <p>Formally: Let T2 be a weighted argumentation tree with arguments (sA; s); (sB; s) and
w(sA; s) &gt; w(sB; s), no premises for sA and sB and 8a 2 rI!s : w(a) 6= 0. T1 is constructed
from T2 by adding (sA1; sA) and T3 from T2 by adding (sB1; sB), each with a weight of 1
and r1(sA1) = r3(sB1) 6= 0. Then dT (T1; T2) &gt; dT (T3; T2).</p>
      <p>T1
w(sA;s)
sA
1
r sA1
s
w(sB;s)
sB</p>
      <p>T2
w(sA;s)
sA
s
w(sB;s)
sB</p>
      <p>T3
w(sA;s)
sA
s</p>
      <p>1
r sB1
w(sB;s)
sB
Proof. Let r := r1(sA1) = r3(sB1) 6= 0. Only the summand which includes sA1 or sB1,
respectively, contributes a value greater than 0.</p>
      <p>dT (T1; T2) = wa jw(sA; s) r
0j &gt; wa j0
w(sB; s) rj = dT (T2; T3)
(15)
Desideratum 5 (Weights of arguments have influence even if they are the only
difference). Consider trees T1; T2, where all trees are identical and have the arguments (sA; s)
and (sB; s), but the weights are different: w1(sA; s) 6= w2(sA; s) and w1(sB; s) 6= w2(sB; s).
We want to have d(T1; T2) &gt; 0 if there exists a statement s0 below sA (or sA itself) with
r1(s0) = r2(s0) 6= 0 and 8a 2 rI!s0 : w(a) 6= 0.</p>
      <p>We demand r1(s0) = r2(s0) 6= 0 because it makes sense if weights leading only to
statements which are rated as “don’t care” are ignored.</p>
      <p>In this example, we have sA = s0:</p>
      <p>T1
w1(sA;s)
r sA
s
w1(sB;s)
sB
Proof. It is enough to show that there is at least one summand greater than 0.
(16)
(17)
(18)
(19)
(20)
(21)
(22)</p>
      <p>Desideratum 6 (Symmetry regarding negation of opinion). Let T1, T2 be any weighted
argumentation trees, and T3, T4, respectively, the same trees, but the opinion for each
statement is negated, i.e. r3(s) = r1(s) and r4(s) = r2(s) for all s 2 S. We expect that
a metric is symmetric regarding negation, i.e. d(T1; T2) = d(T3; T4).</p>
      <p>Proof. This holds because jr1(s) r2(s)j = j( r3(s)) ( r4(s))j = jr3(s) r4(s)j.
Desideratum 7 (Trade-off between argument weights and agreement). Consider trees
T1; T2; T3 which are nearly identical and have leaf statements sA and sB with
common conclusion s and 8a 2 rI!s : w(a) 6= 0. We have r1(sA) = r2(sA) = 0:5, r3(sA) =
0:5, r1(sB) = r2(sB) = r3(sB) = 0, and w1(sA; s) &gt; w2(sA; s). Furthermore, w2(sB; s) =
w1(sB; s) + w1(sA; s) w2(sA; s), w3(sB; s) = w1(sB; s) + w1(sA; s) w3(sA; s), i.e. sB is
neutral and “collects” remaining weight such that the sum is 1.</p>
      <p>If w1(sA; s) w2(sA; s) &gt; w2(sA; s) + w3(sA; s), although T1 and T2 have the same
opinion on sA, we want to have d(T1; T2) &gt; d(T2; T3), because both T2 and T3 do not
care much about their (different) opinions on sA. Likewise, if w1(sA; s) w2(sA; s) &lt;
w2(sA; s) + w3(sA; s), we expect d(T1; T2) &lt; d(T2; T3) because the weights w1 and w2 are
closer to each other and give a greater weight for opposing opinions on sA.</p>
      <p>The following example trees depict the first case with concrete weight values.
Because the different opinions on statement sA are underneath an argument edge with small
weight, we want to have d(T1; T2) &gt; d(T2; T3).</p>
      <p>T1
0 s
0:9 0:1
0:5 sA 0 sB</p>
      <p>T2
0 s
0:1 0:9
0:5 sA 0 sB</p>
      <p>T3
0 s
0:2 0:8
0:5 sA 0 sB
Proof. We proof the first case. For (20) = (21), remember that w1(sA; s) &gt; w2(sA; s).
d(T1; T2) = wa j0:5 w1(sA; s) 0:5 w2(sA; s)j
= wa 0:5 (w1(sA; s)</p>
      <p>w2(sA; s))
&gt; wa 0:5 (w2(sA; s) + w3(sA; s))
= wa j0:5 w2(sA; s) ( 0:5) w3(sA; s)j = d(T2; T3)
Brenneis et al. / How Much Do I Argue Like You? Towards a Metric on Weighted Argumentation Graphs 11</p>
      <p>The other case follows by replacing “&gt;” with “&lt;”.</p>
      <p>Desideratum 8 (Trade-off between statement ratings and agreement). Consider trees
T1; T2; T3 which are nearly identical and have a statement s and 8a 2 rI!s : w(a) 6= 0. We
have r1(s) &gt; r2(s) &gt; 0 &gt; r3(s) such that jr1(s) r2(s)j &gt; jr2(s) r3(s)j. Although T1 and
T2 have the same positive opinion on s, we want to have d(T1; T2) &gt; d(T2; T3), because
both T2 and T3 have a weak opinion on s. Likewise, if jr1(s) r2(s)j &lt; jr2(s) r3(s)j, we
expect d(T1; T2) &lt; d(T2; T3) because the ratings r1(s) and r2(s) are closer to each other
than r2(s) and r3(s).</p>
      <p>The first case, d(T1; T2) &gt; d(T2; T3), is shown in the following example:
T1 0:4 s</p>
      <p>T2 0:1 s</p>
      <p>T3 0:1 s
Proof. Only the summand for s contributes to the distance, all other summands are 0.
Remember that all weights on the path to s are positive.</p>
      <p>d(T1; T2) = wa jr1(s) r2(s)j &gt; wa jr2(s) r3(s)j = d(T2; T3)
(24)</p>
      <p>The other case follows by replacing “&gt;” with “&lt;”.</p>
      <p>Desideratum 9 (Weights limit the influence of a path). Consider graphs T1; T2 which
are nearly identical, have an argument (sA; s) with w = w1(sA; s) = w2(sA; s) and only
the ratings and weights below (and including) sA may differ in any way. No matter how
those values are chosen, we want to have d(T1; T2) w, i.e. the maximum influence of
the differences below sA is limited by the weight of sA to its conclusion.
Proof. It is enough to consider paths which include a, since the summands for all other
parts are 0. Let Sa be the set of all statements which have an argument leading to a,
including a itself. We abbreviate adepth(s0)(1 a) as a(s0).</p>
      <p>dT (T1; T2) = å a(s0) r1(s0)
s02Sa</p>
      <p>Õ
sA0 2rS!I
As the factor after w is in [0; 1], we get dT (T1; T2)</p>
    </sec>
    <sec id="sec-4">
      <title>4. Limitations</title>
      <p>Although the proposed pseudometric fulfills several intuitive desiderata, there are some
limitations which we will discuss in the following.</p>
      <p>If a weight of an argument is 0, the proposed pseudometric ignores all weights which
are underneath this argument. When comparing how similar people argue for or against
T1
0</p>
      <p>I
the top-level positions, this is okay, but if also the way how arguments which are not
supported are attacked should influence the distance, the metric has to be extended.</p>
      <p>Moreover, ordering can be changed by adding an unrelated opinion. Consider trees
T1; T2; T3 with d(T1; T3) &lt; d(T2; T3). At first sight, it might seem unexpected that this
order can be changed to d(T1; T30) &gt; d(T2; T30) by adding a new position c to T3 and
keeping the relative weights of the other positions. Depending on the application context,
e.g. a voting advice application (VAA), this might be unwanted. An example is depicted
in Figure 2. This is due to the normalization of the argument weights.</p>
      <p>Although even end-user friendly systems like D-BAS support for undercuts, i.e.
arguments that have an argument as conclusion, undercuts are currently not explicitly
modeled in our model. This is no big problem, because in many applications, for instance, in
a VAA, arguments can be preselected such that no undercut attack is necessary (since the
arguments make sense), or rephrased such that the premise is attacked. For example,
consider the argument “We should build more nuclear power plants because cats are cute”
and the attack “Cats and nuclear power plants are unrelated”. Though this is technically
an undercut, a user interface may present this as an attack on “Cats are cute”.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Related Work</title>
      <p>
        Calculating the distance between argumentation graphs to compare how similar the
attitudes of two agents are has already been used in other systems. The Carneades opinion
formation and polling tool presented in [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] is able to compare one’s argumentation with
the argumentation of other entities like organizations. This comparison is simply done
by counting the number of statements where the agreement/disagreement is the same.
This approach is much simpler than our proposal, but uses neither weights nor ratings
and violates i.a., Desideratum 3.
      </p>
      <p>
        The mobile application described in [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] also bases on IBIS and extends it with an
agreement value for each argument in the argumentation tree. This information is used,
for opinion prediction using collaborative filtering. In contrast to our work, the idea of
relative argument importance in combination with statement rating is not present.
      </p>
      <p>
        Another application of calculating the similarity of weighted trees is match-making
of agents, which are represented by weighted trees. In [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], a recursive similarity measure
for this application is proposed. Its parameter N serves a similar purpose to our parameter
a . They also give examples which are similar to our desiderata, e.g. Example 4 is like
our Desideratum 2. Nodes, however, do not have a weight, and some desiderata are not
fulfilled; for instance, Desideratum 5 is explicitly not demanded in their Example 2.
      </p>
      <p>
        There are already other definitions of weighted argumentation graphs based upon
Dung’s definition of argumentation frameworks, but many lack the differentiation
between argument relevance and statement ratings (e.g. [
        <xref ref-type="bibr" rid="ref10 ref2 ref6 ref8">2,8,10,6</xref>
        ]), which we think is
important since argument weights limit how much a branch of an argumentation is
relevant, whereas statement ratings are only relevant for the single statement. Furthermore,
in most cases, there is a global assignment of values in the graph and no user-specific
views, which is why a strength of 0 for attacks would be meaningless in the model
presented in [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. Note that most related work in this field is concerned about evaluating
consistency or calculating extensions, whereas our main goal is comparing the attitudes
of agents, not caring about whether an agents’ attitude is logically consistent or not.
      </p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion and Future Work</title>
      <p>In this paper we proposed a pseudometric to calculate the distance between two
argumentation graphs representing the attitudes of different persons, and several desiderata
which should be considered when proposing other metrics for the same purpose, and
are fulfilled by our pseudometric. Possible next steps include developing other sensible
metrics and comparing them regarding theoretical properties and practicality.</p>
      <p>In future work, we want to check if the desiderata are not only intuitive for experts in
argumentation theory, but are following the intuition of untrained humans. We also want
to test the metric in a VAA to compare the argumentation of voters with those of parties.
Thereby we see whether the results of the metric are accepted in an application context.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Althuniyan</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sirrianni</surname>
            ,
            <given-names>J.W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rahman</surname>
            ,
            <given-names>M.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>X.F.</given-names>
          </string-name>
          :
          <article-title>Design of mobile service of intelligent largescale cyber argumentation for analysis and prediction of collective opinions</article-title>
          .
          <source>In: International Conference on AI and Mobile Services</source>
          . pp.
          <fpage>135</fpage>
          -
          <lpage>149</lpage>
          . Springer (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Amgoud</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ben-Naim</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Doder</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vesic</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Acceptability semantics for weighted argumentation frameworks</article-title>
          .
          <source>In: IJCAI</source>
          . vol.
          <year>2017</year>
          , pp.
          <fpage>56</fpage>
          -
          <lpage>62</lpage>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Bhavsar</surname>
            ,
            <given-names>V.C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Boley</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>A weighted-tree similarity algorithm for multi-agent systems in e-business environments</article-title>
          .
          <source>Computational Intelligence</source>
          <volume>20</volume>
          (
          <issue>4</issue>
          ),
          <fpage>584</fpage>
          -
          <lpage>602</lpage>
          (
          <year>2004</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Brenneis</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mauve</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <article-title>: deliberate - online argumentation with collaborative filtering</article-title>
          .
          <source>In: COMMA</source>
          (
          <year>2020</year>
          ), to appear
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Dung</surname>
            ,
            <given-names>P.M.</given-names>
          </string-name>
          :
          <article-title>On the acceptability of arguments and its fundamental role in nonmonotonic reasoning, logic programming and n-person games</article-title>
          .
          <source>Artificial Intelligence</source>
          <volume>77</volume>
          (
          <issue>2</issue>
          ),
          <fpage>321</fpage>
          -
          <lpage>357</lpage>
          (
          <year>1995</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Dunne</surname>
            ,
            <given-names>P.E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hunter</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>McBurney</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Parsons</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wooldridge</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Weighted argument systems: Basic definitions, algorithms, and complexity results</article-title>
          .
          <source>Artificial Intelligence</source>
          <volume>175</volume>
          (
          <issue>2</issue>
          ),
          <fpage>457</fpage>
          -
          <lpage>486</lpage>
          (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Ebbinghaus</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Decision making with argumentation graphs</article-title>
          . arXiv preprint arXiv:
          <year>1908</year>
          .
          <volume>03357</volume>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Gordon</surname>
            ,
            <given-names>T.F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Walton</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Formalizing balancing arguments</article-title>
          .
          <source>In: COMMA</source>
          . pp.
          <fpage>327</fpage>
          -
          <lpage>338</lpage>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Gordon</surname>
            ,
            <given-names>T.F.</given-names>
          </string-name>
          :
          <article-title>Structured consultation with argument graphs. From Knowledge Representation to Argumentation in AI. A Festschrift in Honour of Trevor Bench-Capon on the Occasion of his 60th Birthday pp</article-title>
          .
          <fpage>115</fpage>
          -
          <lpage>133</lpage>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Hunter</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>A probabilistic approach to modelling uncertain logical arguments</article-title>
          .
          <source>International Journal of Approximate Reasoning</source>
          <volume>54</volume>
          (
          <issue>1</issue>
          ),
          <fpage>47</fpage>
          -
          <lpage>81</lpage>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Krauthoff</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Meter</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mauve</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Dialog-Based Online Argumentation: Findings from a Field Experiment</article-title>
          .
          <source>In: Proceedings of the 1st Workshop on Advances in Argumentation in Artificial Intelligence</source>
          . pp.
          <fpage>85</fpage>
          -
          <lpage>99</lpage>
          (
          <year>November 2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Kunz</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rittel</surname>
            ,
            <given-names>H.W.J.</given-names>
          </string-name>
          :
          <article-title>Issues as elements of information systems</article-title>
          , vol.
          <volume>131</volume>
          .
          <string-name>
            <surname>Citeseer</surname>
          </string-name>
          (
          <year>1970</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Martınez</surname>
            ,
            <given-names>D.C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Garcıa</surname>
            ,
            <given-names>A.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Simari</surname>
            ,
            <given-names>G.R.:</given-names>
          </string-name>
          <article-title>An abstract argumentation framework with varied-strength attacks</article-title>
          .
          <source>In: Proceedings of the Eleventh International Conference on Principles of Knowledge Representation and Reasoning (KR'08)</source>
          . pp.
          <fpage>135</fpage>
          -
          <lpage>144</lpage>
          (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Neugebauer</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          : Dabasco:
          <article-title>Generating af, adf, and aspic+ instances from real-world discussions</article-title>
          .
          <source>In: COMMA</source>
          . pp.
          <fpage>469</fpage>
          -
          <lpage>470</lpage>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>