<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Using the Dempster-Shafer theory of evidence to resolve ABox inconsistencies</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Andriy Nikolov</string-name>
          <email>a.nikolov@open.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Victoria Uren</string-name>
          <email>v.s.uren@open.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Enrico Motta</string-name>
          <email>e.motta@open.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Anne de Roeck</string-name>
          <email>a.deroeck@open.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Knowledge Media Institute, The Open University</institution>
          ,
          <addr-line>Milton Keynes</addr-line>
          ,
          <country country="UK">UK</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Automated ontology population using information extraction algorithms can produce inconsistent knowledge bases. Confidence values assigned by the extraction algorithms may serve as evidence helping to repair produced inconsistencies. The Dempster-Shafer theory of evidence is a formalism, which allows appropriate interpretation of extractors' confidence values. The paper presents an algorithm for translating the subontologies containing conflicts into belief propagation networks and repairing conflicts based on the Dempster-Shafer plausibility.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        One of the approaches for ontology population considers using automatic
information extraction algorithms to annotate natural language data already
available on the Web [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ]. Unsupervised information extraction algorithms do not
produce 100% correct output, which may lead to inconsistency of the whole
knowledge base produced in this way. Also information extracted from different
sources can be genuinely contradictory. So when performing fusion of knowledge
extracted from different sources it is important to resolve such inconsistencies
automatically or provide the user with a ranking of conflicting options estimating
how likely each statement is to be wrong. Extraction algorithms can often
estimate the reliability of their output by attaching confidence values to produced
statements [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Uncertain reasoning using these confidence values can help to
evaluate the plausibility of statements and rank the conflicting options. Most
of the ongoing research in the field of applying uncertain reasoning to the
Semantic Web focuses on fuzzy logic and probabilistic approaches. Fuzzy logic was
designed to deal with representation of vagueness and imprecision. This
interpretation is not relevant for the problem occurring during population of crisp
OWL knowledge bases, where we need to assess the likelihood for a statement
to be true or false. The probabilistic approach is more appropriate for dealing
with such problems. However, as stated in [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], axioms of probability theory are
implied by seven properties of belief measures. One of them is completeness,
which states that “a degree of belief can be assigned to any well-defined
proposition”. However, this property cannot be ensured when dealing with confidence
degrees assigned by extractors, because they do not always carry information
about the probability of a statement being false. The Dempster-Shafer theory of
evidence [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] presents a formalism that helps to overcome this problem. It allows
belief measurements to be assigned to sets of propositions, thus specifying
explicitly degrees of ignorance. In this paper, we describe an algorithm for resolving
conflicts using the Dempster-Shafer belief propagation approach.
2
      </p>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <p>
        There are several studies dealing with inconsistency handling in OWL
ontologies, among others [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] and [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. The general algorithm for the task of repairing
inconsistent ontologies consists of two steps:
– Ontology diagnosis: finding sets of axioms, which contribute to inconsistency;
– Repairing inconsistencies: changing/removing the axioms most likely to be
erroneous.
      </p>
      <p>
        Choosing the axioms for change and removal is a non-trivial task. Existing
algorithms working with crisp ontologies (e.g., [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]) utilize such criteria as syntactic
relevance (how often each entity is referenced in the ontology), impact (the
influence of removal of the axiom on the ontology should be minimized) and
provenance (reliability of the source of the axiom). The last criterion is especially
interesting for the automatic ontology population scenario since extraction
algorithms do not extract information with 100% accuracy. A study described in
[
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] specifies an algorithm which utilizes the confidence value assigned by the
extraction algorithm. The strategy employed there was to order the axioms
according to their confidence and add them incrementally, starting from the most
certain one. If adding the axiom led to inconsistency of the ontology then a
minimal subconsistent ontology was determined and the axiom with the lowest
confidence was removed from it. A disadvantage of such a technique is that it
does not take into account the impact of an axiom: e.g., when an axiom
violates several restrictions, it does not increase its chances to be removed. Also
it does not consider the influence of redundancy: if the same statement was
extracted from several sources, this should increase its reliability. Using uncertain
reasoning would provide a more sound approach to rank potentially erroneous
statements and resolve inconsistencies.
      </p>
      <p>
        In the Semantic Web domain the studies on uncertain reasoning are mostly
focused on two formalisms: probability theory and fuzzy logic. Existing
implementations of fuzzy description logic [
        <xref ref-type="bibr" rid="ref10 ref9">9, 10</xref>
        ] are based on the notion of fuzzy
set representing a vague concept. The uncertainty value in this context denotes
a membership function μF (x) which specifies the degree to which an object x
belongs to a fuzzy class F . Probabilistic adaptations of OWL-DL include Bayes
OWL[
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] and PR-OWL [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. However, both of these formalisms do not fully
reflect the properties of the problems we are dealing with in the fusion scenario.
In [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] a framework for choosing an appropriate uncertainty handling formalism
was presented. The framework is based on the following seven properties of belief
measurements:
1. Clarity : Propositions should be well-defined.
2. Scalar continuity : A single real number is both necessary and sufficient for
representing a degree of belief.
3. Completeness: A degree of belief can be assigned to any well-defined
proposition.
4. Context dependency : The belief assigned to a proposition can depend on the
belief in other propositions.
5. Hypothetical conditioning : There exists some function that allows the belief
in a conjunction of propositions to be calculated from the belief in one
proposition and the belief in the other proposition given that the first proposition
is true.
6. Complementarity : The belief in the negation of a proposition is a
monotonically decreasing function of the belief in the proposition itself.
7. Consistency : There will be equal belief in propositions that have the same
truth value.
      </p>
      <p>It was proven that accepting all seven properties logically necessitates the axioms
of probability theory. Alternative formalisms allow weakening of some
properties. Fuzzy logic deals with the case when the clarity property does not hold,
i.e., when concepts and relations are vague. Such an interpretation differs from
the one we are dealing with in the fusion scenario, where the ontology TBox
contains crisp concepts and properties. Confidence value attached to a type
assertion ClassA(Individual1) denotes a degree of belief that the statement is true
in the real world rather than the degree of inclusion of the entity Individual1
into a fuzzy concept ClassA. This makes fuzzy interpretation inappropriate for
our case.</p>
      <p>
        Probabilistic interpretation of the extraction algorithm’s confidence may lead to
a potential problem. If we interpret the confidence value c attached to a
statement returned by an extraction algorithm as a Bayesian probability value p, we,
at the same time, introduce a belief that the statement is false with a
probability 1 − p. However, the confidence of an extraction algorithm reflects only the
belief that the document supports the statement and does not itself reflect the
probability of a statement being false in the real world. Also while statistical
extraction algorithms ([
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]) are able to assign a degree of probability to each
extracted statement, rule-based algorithms ([
        <xref ref-type="bibr" rid="ref14 ref15">14, 15</xref>
        ]) can only assign the same
confidence value to all statements extracted by the same rule based on the rule’s
performance on some evaluation set. Any extraction produced by a rule with a
low confidence value in this case will serve as a negative evidence rather than
simply lack of evidence. This issue becomes more important if the reliability of
sources is included into analysis: it is hard to assign the conditional probability
of a statement being false given that the document supports it. It means that
the completeness property does not always hold.
      </p>
      <p>
        The Dempster-Shafer theory of evidence [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] allows weakening of the
completeness property. Belief can be assigned to sets of alternative options rather than
only to atomic elements. In the case of binary logic, it means that the degree
of ignorance can be explicitly represented by assigning a non-zero belief to the
set {true;false}. On the other hand, it still allows the Bayesian interpretation of
confidence to be used, when it is appropriate (in this case the belief assigned to
the set {true;false} is set to 0). This paper presents an algorithm for resolving
inconsistencies by translating the inconsistency-preserving subset of ontology into
the Dempster-Shafer belief network and choosing the axioms to remove based
on their plausibility. We are not aware of other studies adapting the
DempsterShafer approach to the Semantic Web domain.
      </p>
      <p>
        Alternative approaches to uncertainty representation, which were not applied so
far to ontological modelling, include probability intervals [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] and higher-order
probability [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]. However, the first of these approaches uses min and max
operators for aggregation, which makes it hard to represent cumulative evidence, and
the second focuses on resolving different kinds of problems (namely expressing
probability estimations of other probability estimations). There are also other
approaches to belief fusion in the Semantic Web (e.g., [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] and [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ]). These
studies deal with social issues of representing trust and provenance in a distributed
knowledge base and focus on the problem of establishing the certainty of
statements asserted by other people. These approaches, however, do not focus on
resolving the inconsistencies and just deal with direct conflicts (i.e., statement
A is true vs statement A is false). They do not take into account ontological
inference and mutual influence of statements in the knowledge base. In this way,
they can be considered complementary to ours.
3
      </p>
    </sec>
    <sec id="sec-3">
      <title>The Dempster-Shafer Belief Theory</title>
      <p>Dempster-Shafer theory of evidence differs from the Bayesian probability
theory as it allows assigning beliefs not only to atomic elements but to sets of
elements as well. The base of the Dempster’s belief distribution is the frame of
discernment (Ω) - an exhaustive set of mutually exclusive alternatives. A belief
distribution function (also called mass function or belief potential) m(A)
represents the influence of a piece of evidence on subsets of Ω and has the following
constraints:
– m( ) = 0 and
– PA⊆Ω m(A) = 1
m(A) defines the amount of belief assigned to the subset A. When m(A) &gt; 0, A
is referred to as a focal element. If each focal element A contains only a single
element, the mass function is reduced to be a probability distribution. Mass also
can be assigned to the whole set of Ω. This represents the uncertainty of the
piece of evidence about which of the elements in Ω is true. In our case each mass
function is defined on a set of variables D = {x1, ..., xn} called the domain of m.
Each variable is boolean and represents an assertion in the knowledge base. For
a single variable we can get degree of support Sup(x) = m({true}) and degree
of plausibility P l(x) = m({true}) + m({true; f alse}). Plausibility specifies how
likely it is that the statement is false. Based on plausibility it is possible to select
from a set of statements the one to be removed.</p>
    </sec>
    <sec id="sec-4">
      <title>Building Belief Networks</title>
      <sec id="sec-4-1">
        <title>Our algorithm consists of four steps:</title>
      </sec>
      <sec id="sec-4-2">
        <title>1. Inconsistency detection.</title>
        <p>At this stage we select the subontology containing all axioms contributing
to an inconsistency.
2. Constructing a belief network.</p>
        <p>At this stage the subontology found at the previous step is translated into
a belief network.
3. Assigning mass distributions.</p>
        <p>At this stage we assign mass distribution functions to nodes.
4. Belief propagation.</p>
        <p>At this stage we propagate uncertainties through the network and update
the confidence degrees of ABox statements.
4.1</p>
        <sec id="sec-4-2-1">
          <title>Illustrating Example</title>
          <p>In order to illustrate our algorithm, we use an example from the banking
domain. Supposedly, we have an ontology describing credit card applications, which
defines two disjoint classes of applicants: reliable and risky. In order to be
reliable, an applicant has to have UK citizenship and evidence that (s)he was never
bankrupt in the past. For example, the TBox contains the following axioms:
T1: RiskyApplicant v CreditCardApplicant
T2: ReliableApplicant v CreditCardApplicant
T3: RiskyApplicant v ¬ReliableApplicant
T4: ReliableApplicant ≡ ∃wasBankrupt.F alse u ∃hasCitizenship.U K
T5: &gt; v≤ 1wasBankrupt (wasBankrupt is functional)
The ABox contains the following axioms (with attached confidence values):
A1: RiskyApplicant(Ind1) : 0.7
A2: wasBankrupt(Ind1, F alse) : 0.6
A3: hasCitizenship(Ind1, U K) : 0.4
A4: wasBankrupt(Ind1, T rue) : 0.5
As given, the ontology is inconsistent: the individual Ind1 is forced to belong
to both classes RiskyApplicant and ReliableApplicant, which are disjoint, and
the functional property wasBankrupt has two different values. If we choose to
remove the axioms with the lowest confidence values, it will require removing A3
and A4. However, inconsistency can also be repaired by removing a single
statement A2. The fact that A2 leads to the violation of two ontological constraints
should increase the likelihood it is wrong.
4.2</p>
        </sec>
        <sec id="sec-4-2-2">
          <title>Inconsistency Detection</title>
          <p>
            The task of the inconsistency detection step is to retrieve all minimal inconsistent
subontologies (MISO) of the ontology and combine them. As defined in [
            <xref ref-type="bibr" rid="ref6">6</xref>
            ], an
ontology O0 is a minimal inconsistent subontology of an ontology O, if O0 ⊆ O
and O0 is inconsistent and for all O00 such that O00 ⊂ O0 ⊆ O, O00 is consistent.
OWL reasoner Pellet [
            <xref ref-type="bibr" rid="ref7">7</xref>
            ] is able to return the MISO for the first encountered
inconsistency in the ontology. To calculate all MISO O01, ..., O0n in the ontology
we employ Reiter’s hitting set tree algorithm [
            <xref ref-type="bibr" rid="ref20">20</xref>
            ]. After all conflict sets were
identified, the next step involves constructing belief networks from each set. If
for two subontologies O0i ∩ O0j 6= then these two subontologies are replaced
with O0 = O0i ∪ O0j .
          </p>
          <p>For our illustrating example, the conflict detection algorithm is able to identify
two conflict sets in this ontology: the first, consisting of {T3, T4, A1, A2, A3}
(individual Ind1 belongs to two disjoint classes), and the second {T5, A2, A4}
(individual Ind1 has two instantiations of a functional property). The statement
A2 belongs to both sets and therefore the sets are merged.
4.3</p>
        </sec>
        <sec id="sec-4-2-3">
          <title>Constructing Belief Networks</title>
          <p>
            The networks for propagation of Dempster-Shafer belief functions (also called
valuation networks) were described in [
            <xref ref-type="bibr" rid="ref21">21</xref>
            ]. By definition the valuation network
is an undirected graph represented as a 5-tuple: {Ψ, {ΩX }X∈Ψ , {τ1, ..., τn}, ↓, ⊗},
where Ψ is a set of variables, {ΩX }X∈Ψ is a collection of state spaces, {τ1, ..., τn}
is a collection of valuations (belief potentials of nodes), ↓ is a marginalization
operator and ⊗ is a combination operator. In our case Ψ consists of ABox
assertions, every {ΩX }X∈Ψ = {0; 1} and {τ1, ..., τn} are created using rules described
below. The operators are used for propagation of beliefs and are described in the
following subsections. The network contains two kinds of nodes: variable nodes
corresponding to explicit or inferred ABox assertions and valuation nodes
representing TBox axioms. Variable nodes contain only one variable, while valuation
nodes contain several variables.
          </p>
          <p>Translation of an inconsistent subontology into a belief propagation network is
performed using a set of rules (Table 1). Each rule translates a specific OWL-DL
construct into a set of network nodes and links between them. Rules 1 and 2
directly translate each ABox statement into a variable node. Other rules process
TBox axioms and create two kinds of nodes: one valuation node to represent the
TBox axiom and one or more variable nodes to represent inferred statements.
Such rules only fire if the network already contains variable nodes for ABox
axioms, which are necessary to make the inference. For example, a rule processing
the class equivalence axiom (Rule 4) is interpreted as the following: “If there is
a node N1 representing the type assertion I ∈ X and an owl : equivalentClass
axiom X ≡ Y , then:
– Create a node N2 representing the assertion I ∈ Y ;
– Create a node N3 representing the axiom X v Y ;
– Create links between N1 and N3 and between N3 and N2.”</p>
          <p>If a rule requires creating a node, which already exists in the network, then
the existing node is used.</p>
          <p>Applying the rules described above to our illustrating example (rules 1, 2, 4, 5,
6, 9, 20) will result in the following network (Fig. 1).</p>
          <p>N Pre-conditions
1 I ∈ X
2 R(I1, I2)
3 N1 : I ∈ X, X v Y
4 N1 : I ∈ X, X ≡ Y
5 N1 : I ∈ X, X v ¬Y
6 N1 : I ∈ X, X u Y
7 N1 : I ∈ X, X t Y
8 N1 : I ∈ X, ¬X</p>
          <p>Nodes to create Links to create
N1 : I ∈ X
N2 : R(I1, I2)
N2 : I ∈ Y , N3 : X v Y (N1,N3),(N3,N2)
N2 : I ∈ Y , N3 : X v Y (N1,N3),(N3,N2)
N2 : I ∈ Y , N3 : X v ¬Y (N1,N3),(N3,N2)
N2 : I ∈ X u Y , N3 : X u Y , (N1,N3),(N4,N3),
N4 : I ∈ Y (N3,N2)
N2 : I ∈ X t Y , N3 : X t Y , (N1,N3),(N4,N3),
N4 : I ∈ Y (N3,N2)</p>
          <p>N2 : I ∈ ¬X, N3 : ¬X (N1,N3),(N3,N2)
9 &gt;N2v:≤R(1IR, ,o2N)1 : R(I, o1), N3 : &gt; v≤ 1R
10 &gt;N2v:≤R(1IR3,−I,1N)1 : R(I2, I1), N3 : &gt; v≤ 1R−
11 R ≡ R−, N1 : R(I1, I2) N2 : R ≡ R−, N3 : R(I2, I1) (N1,N2),(N2,N3)
12 R ≡ Q, N1 : R(I1, I2) N2 : R ≡ Q,N3 : Q(I1, I2) (N1,N2),(N2,N3)
13 R v Q,N1 : R(I1, I2) N2 : R v Q, N3 : Q(I1, I2) (N1,N2),(N2,N3)
14 R ≡ Q−,N1 : R(I1, I2) N2 : R ≡ Q−, N3 : Q(I2, I1) (N1,N2),(N2,N3)
15 TNr2a:nRs((RI2),,IN3)1 : R(I1, I2), N3 : T rans(R), N4 : R(I1, I3) (N3,N4)
(N1,N3),(N2,N3),
21 ∃NR2.:&gt;I1v∈XX, N1 : R(I1, I2), N3 : ∃R.&gt; v X
After the nodes were combined into the network, the next step is to assign the
mass distribution functions to the nodes. There are two kinds of variable nodes:
(i) nodes representing statements supported by the evidence and (ii) nodes
representing inferred statements. Initial mass distribution for the nodes of the first
type is assigned based on their extracted confidence. If a statement was
extracted with a confidence degree c, it is assigned the following mass distribution:
m(T rue) = c, m(T rue; F alse) = 1 − c. It is possible that the same statement is
extracted from several sources. In this case, multiple pieces of evidence have to
be combined using Dempster’s rule of combination.</p>
          <p>Nodes created artificially during network construction are only used for
propagation of beliefs from their neighbours and do not contain their own mass
assignment. Valuation nodes specify the TBox axioms and are used to propagate
beliefs through the network. For the crisp OWL ontologies only mass
assignments of 0 and 1 are possible. The principle for assigning masses is to assign
the mass of 1 to the set of all combinations of variable sets allowed by the
corresponding axiom. Table 2 shows the mass assignment functions for OWL-DL
T-Box axioms 1.</p>
          <p>In our example, we assign distributions based on the extractor’s confidence
values to the variable nodes representing extracted statements: A1:(m(1)=0.7,
m({0;1})=0.3), A2: (m(1)=0.6, m({0;1})=0.4), A3: (m(1)=0.4, m({0;1})=0.6),
1 For nodes allowing multiple operands (e.g., intersection or cardinality) only the case
of two operands is given. If the node allows more than two children, then number
of variables and the distribution function is adjusted to represent the restriction
correctly
A4: (m(1)=0.5, m({0;1})=0.5). The valuation nodes obtain their distributions
according to the rules specified in the Table 2: T3 (rule 3), T4 (rules 2, 4, 18)
and T5 (rule 7).</p>
        </sec>
        <sec id="sec-4-2-4">
          <title>4.5 Belief Propagation</title>
          <p>
            The axioms for belief propagation were formulated in [
            <xref ref-type="bibr" rid="ref22">22</xref>
            ]. The basic operators
for belief potentials are marginalization ↓ and combination ⊗. Marginalization
takes a mass distribution function m on domain D and produces a new mass
distribution on domain C ⊆ D.
          </p>
          <p>m↓C(X) = X m(Y )</p>
          <p>Y↓C=X
For instance, if we have the function m defined on domain {x,y} as m({0;0}) =
0.2, m({0;1}) = 0.35, m({1;0}) = 0.3, m({1;1}) = 0.15 and we want to find
a marginalization on domain {y}, we will get m(0) = 0.2 + 0.3 = 0.5 and
m(1) = 0.35 + 0.15 = 0.5. The combination operator is represented by the
Dempster’s rule of combination:
m1 ⊗ m2(X) =</p>
          <p>PX1∩X2=X m1(X1)m2(X2)
1 − PX1∩X2= m1(X1)m2(X2)
Belief propagation is performed by passing messages between nodes according
to the following rules:
1. Each node sends a message to its inward neighbour (towards the root of the
tree). If μA→B is a message from A to B, N (A) is a set of neigbours of A
and the potential of A is mA, then the message is specified as a combination
of messages from all neighbours except B and the potential of A:
μA→B = (⊗{μX→A|X ∈ (N (A) − {B}) ⊗ mA})↓A∩B
2. After a node A has received a message from all its neighbors, it combines all
messages with its own potential and reports the result as its marginal.
As the message-passing algorithm assumes that the graph is a tree, it is
necessary to eliminate loops. All valuation nodes constituting the loop are replaced
by a single node with the mass distribution equal to the combination of mass
distributions of its constituents. The marginals obtained after propagation for
the nodes corresponding to initial ABox assertions will reflect updated mass
distributions. After the propagation we can remove the statement with the lowest
plausibility from each of the MISO found at the diagnosis stage.
Calculating the beliefs for our example gives the following Dempster-Shafer
plausibility values for ABox statements: Pl(A1)=0.94, Pl(A2)=0.58, Pl(A3)=0.8,
Pl(A4)=0.65. In order to make the ontology consistent it is sufficient to
remove from both conflict sets an axiom with the lowest plausibility value (A2).
In this example, we can see how the results using Dempster-Shafer belief
propagation differ from the Bayesian interpretation. Bayesian probabilities, in this
case, are calculated in the same way as Dempster-Shafer support values. If we
use confidence values as probabilities and propagate them using the same
valuation network we will obtain the results: P(A1)=0.66, P(A2)=0.35, P(A3)=0.32
and P(A4)=0.33. In this scenario, we would remove A3 and A4 because of the
negative belief bias. Also we can see that all three statements A2, A3 and A4
will be considered wrong in such a scenario (resulting probability is less than
0.5). The Dempster-Shafer approach provides more flexibility by making it
possible to reason about both support (“harsh” queries) and plausibility (“lenient”
queries).
5</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Conclusion and Future Work</title>
      <p>
        In this paper, we described how the Dempster-Shafer theory of evidence can be
used for dealing with ABox-level inconsistencies produced by inaccurate
information extraction. It would be interesting to investigate if the capabilities of
the Dempster-Shafer uncertainty representation (e.g., explicit representation of
ignorance) can be utilized for knowledge modelling at the TBox level. In [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ] it
was shown that the Dempster-Shafer approach may lead to problems when it
is used to represent uncertainty of inferencing rules (i.e., TBox-level) and not
only of pieces of evidence (ABox assertions). These problems occur if the
ontology contains contradictory pieces of knowledge, and are caused by the fact that
the Dempster-Shafer approach does not distinguish pieces of evidence
regarding specific individuals from generic rules applicable to all individuals. It will
be interesting to investigate if these problems can be avoided when modelling
description logic axioms.
      </p>
      <p>The algorithm described in the paper focuses on only one aspect of provenance
information: confidence values assigned by extraction algorithms. However, such
an approach has its limitations: for instance, we know that rule-based extractors
tend to repeat their errors when applied to several documents. Such reoccurring
errors lead to erroneous inconsistency resolution if interpreted as independent
pieces of evidence. In order to improve the quality of the fusion procedure, it
would be useful to take into account other kinds of provenance information,
in particular, the reliability of the extraction algorithm itself, the reliability of
sources from which statements were extracted, and the timestamp reflecting
when each statement was produced. This we consider our primary direction for
the future work.
6</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgements</title>
      <p>This work was funded by the X-Media project (www.x-media-project.org)
sponsored by the European Commission as part of the Information Society
Technologies (IST) programme under EC grant number IST-FP6-026978.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Welty</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Murdock</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          :
          <article-title>Towards knowledge acquisition from information extraction</article-title>
          .
          <source>In: 5th International Semantic Web Conference</source>
          , Athens, GA, USA, November 5-
          <issue>9</issue>
          ,
          <year>2006</year>
          . (
          <year>2006</year>
          )
          <fpage>709</fpage>
          -
          <lpage>722</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Popov</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>KIM - a semantic platform for information extraction and retrieval</article-title>
          .
          <source>Natural language engineering 10</source>
          (
          <year>2004</year>
          )
          <fpage>375</fpage>
          -
          <lpage>392</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Iria</surname>
          </string-name>
          , J.:
          <article-title>Relation extraction for mining the Semantic Web</article-title>
          .
          <source>In: Dagstuhl Seminar on Machine Learning for the Semantic Web</source>
          , Dagstuhl, Germany (
          <year>2005</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Horvitz</surname>
            ,
            <given-names>E.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Heckerman</surname>
            ,
            <given-names>D.E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Langlotz</surname>
            ,
            <given-names>C.P.:</given-names>
          </string-name>
          <article-title>A framework for comparing alternative formalisms for plausible reasoning</article-title>
          .
          <source>In: AAAI-86</source>
          . (
          <year>1986</year>
          )
          <fpage>210</fpage>
          -
          <lpage>214</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Shafer</surname>
          </string-name>
          , G.:
          <article-title>A mathematical theory of evidence</article-title>
          . Princeton University Press (
          <year>1976</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Haase</surname>
          </string-name>
          , P.,
          <string-name>
            <surname>van Harmelen</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stuckenschmidt</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sure</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          :
          <article-title>A framework for handling inconsistency in changing ontologies</article-title>
          .
          <source>In: Proceedings of the International Semantic Web Conference (ISWC2005)</source>
          . (
          <year>2005</year>
          )
          <fpage>353</fpage>
          -
          <lpage>367</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Kalyanpur</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Parsia</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sirin</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grau</surname>
            ,
            <given-names>B.C.</given-names>
          </string-name>
          :
          <article-title>Repairing unsatisfiable concepts in OWL ontologies</article-title>
          .
          <source>In: ESWC2006</source>
          . (
          <year>2006</year>
          )
          <fpage>170</fpage>
          -
          <lpage>184</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Haase</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Volker</surname>
          </string-name>
          , J.:
          <article-title>Ontology learning and reasoning - dealing with uncertainty and inconsistency</article-title>
          .
          <source>In: International Workshop on Uncertainty Reasoning for the Semantic Web (URSW)</source>
          .
          <article-title>(</article-title>
          <year>2005</year>
          )
          <fpage>45</fpage>
          -
          <lpage>55</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Stoilos</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stamou</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tzouvaras</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pan</surname>
            ,
            <given-names>J.Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Horrocks</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <string-name>
            <surname>Fuzzy</surname>
            <given-names>OWL</given-names>
          </string-name>
          :
          <article-title>Uncertainty and the semantic web</article-title>
          . In: International Workshop of OWL: Experiences and Directions,
          <string-name>
            <surname>ISWC</surname>
          </string-name>
          <year>2005</year>
          .
          <article-title>(</article-title>
          <year>2005</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Straccia</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          :
          <article-title>Towards a fuzzy description logic for the Semantic Web (preliminary report)</article-title>
          .
          <source>In: 2nd European Semantic Web Conference (ESWC-05). Number 3532 in Lecture Notes in Computer Science</source>
          , Crete, Springer Verlag (
          <year>2005</year>
          )
          <fpage>167</fpage>
          -
          <lpage>181</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Ding</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Peng</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          :
          <article-title>A probabilistic extension to ontology language OWL</article-title>
          .
          <source>In: 37th Hawaii International Conference On System Sciences (HICSS-37)</source>
          . (
          <year>2004</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12. da Costa,
          <string-name>
            <given-names>P.C.G.</given-names>
            ,
            <surname>Laskey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.B.</given-names>
            ,
            <surname>Laskey</surname>
          </string-name>
          ,
          <string-name>
            <surname>K.J.:</surname>
          </string-name>
          <article-title>PR-OWL: A Bayesian ontology language for the semantic web</article-title>
          .
          <source>In: Workshop on Uncertainty Reasoning for the Semantic Web</source>
          ,
          <string-name>
            <surname>ISWC</surname>
          </string-name>
          <year>2005</year>
          .
          <article-title>(</article-title>
          <year>2005</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bontcheva</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cunningham</surname>
          </string-name>
          , H.:
          <article-title>SVM based learning system for information extraction</article-title>
          .
          <source>In: Deterministic and Statistical Methods in Machine Learning</source>
          . (
          <year>2005</year>
          )
          <fpage>319</fpage>
          -
          <lpage>339</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Ciravegna</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wilks</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          :
          <article-title>Designing Adaptive Information Extraction for the Semantic Web in Amilcare</article-title>
          . In:
          <article-title>Annotation for the Semantic Web</article-title>
          . (
          <year>2003</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Zhu</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Uren</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Motta</surname>
          </string-name>
          , E.:
          <article-title>ESpotter: Adaptive named entity recognition for web browsing</article-title>
          .
          <source>In: Professional Knowledge Management Conference (WM2005)</source>
          . (
          <year>2005</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>de Campos</surname>
            ,
            <given-names>L.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Huete</surname>
            ,
            <given-names>J.F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moral</surname>
            ,
            <given-names>S.:</given-names>
          </string-name>
          <article-title>Uncertainty management using probability intervals</article-title>
          .
          <source>In: Advances in Intelligent Computing IPMU '94</source>
          . (
          <year>1994</year>
          )
          <fpage>190</fpage>
          -
          <lpage>199</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Gaifman</surname>
          </string-name>
          , H.:
          <article-title>A theory of higher order probabilities</article-title>
          .
          <source>In: TARK '86: Proceedings of the 1986 conference on Theoretical aspects of reasoning about knowledge</source>
          , San Francisco, CA, USA, Morgan Kaufmann Publishers Inc. (
          <year>1986</year>
          )
          <fpage>275</fpage>
          -
          <lpage>292</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Richardson</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Agrawal</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Domingos</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Trust management for the Semantic Web</article-title>
          .
          <source>In: 2nd International Semantic Web Conference, ISWC</source>
          <year>2003</year>
          ,
          <string-name>
            <given-names>Sanibel</given-names>
            <surname>Island</surname>
          </string-name>
          ,
          <string-name>
            <surname>Florida</surname>
          </string-name>
          (
          <year>2003</year>
          )
          <fpage>351</fpage>
          -
          <lpage>368</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Nickles</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Social acquisition of ontologies from communication processes</article-title>
          .
          <source>Journal of Applied Ontology</source>
          (
          <year>2007</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Reiter</surname>
          </string-name>
          , R.:
          <article-title>A theory of diagnosis from first principles</article-title>
          .
          <source>Artificial Intelligence</source>
          <volume>32</volume>
          (
          <issue>1</issue>
          ) (
          <year>1987</year>
          )
          <fpage>57</fpage>
          -
          <lpage>95</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>Shenoy</surname>
            ,
            <given-names>P.P.</given-names>
          </string-name>
          :
          <article-title>Valuation-based systems: a framework for managing uncertainty in expert systems</article-title>
          . In:
          <article-title>Fuzzy logic for the management of uncertainty</article-title>
          . John Wiley &amp; Sons, Inc., New York, NY, USA (
          <year>1992</year>
          )
          <fpage>83</fpage>
          -
          <lpage>104</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22.
          <string-name>
            <surname>Shenoy</surname>
            ,
            <given-names>P.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shafer</surname>
          </string-name>
          , G.:
          <article-title>Axioms for probability and belief-function propagation</article-title>
          . In:
          <article-title>Readings in uncertain reasoning</article-title>
          . Morgan Kaufmann Publishers Inc., San Francisco, CA, USA (
          <year>1990</year>
          )
          <fpage>575</fpage>
          -
          <lpage>610</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <surname>Pearl</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          :
          <article-title>Bayesian and belief-function formalisms for evidential reasoning: a conceptual analysis</article-title>
          .
          <source>In: Readings in uncertain reasoning</source>
          . Morgan Kaufmann Publishers Inc., San Francisco, CA, USA (
          <year>1990</year>
          )
          <fpage>575</fpage>
          -
          <lpage>610</lpage>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>