<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A Study of Kernel Contraction in E L</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Amr Dawood</string-name>
          <email>dawoodg@cs.sfu.ca</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>James Delgrande</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Zhiwei Liao</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Simon Fraser University Burnaby</institution>
          ,
          <addr-line>B.C.</addr-line>
          ,
          <country country="CA">Canada</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>An intelligent agent must be able to change its beliefs in a rational manner and with a plausible outcome. While the central belief change theories have well-developed formal characterisations, there has been limited work on applying these theories in a practical setting. In this paper we examine belief base contraction, specifically kernel contraction, in the framework of the description logic EL. This is potentially interesting first, because EL is used in large-scale knowledge bases and second, because the EL constructs for structuring knowledge allow the specification of heuristics to choose an intuitive, commonsense result for the contraction. We introduce algorithms that perform belief contraction based on hitting sets, and examine different ways of finding a suitable hitting set. Heuristics based on localization and specificity are introduced. As well, our main contraction operators are shown to conform to Hansson's postulates for smooth base contraction.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>An agent interacting with some domain will be required to
modify its beliefs, perhaps as a result of receiving additional
information about the domain, or perhaps in correcting an
inaccurate belief. The dominant approach to belief change
with respect to a static domain1 is the AGM approach. Two
primary belief change operators are considered, belief
revision in which the agent must consistently (when possible)
incorporate a new belief into its belief corpus, and belief
contraction, in which an agent ceased to believes a given
formula.</p>
      <p>In this paper we focus on belief contraction: It is seen
as the more basic operation; it is applicable even when the
underlying system lacks a notion of inconsistency; and as
covered in the Background section, revision may be defined
in terms of it.</p>
      <p>Our goal is to address belief contraction in a Description
Logic called E L. A Description Logic (DL) is one of a
family of formalisms that are used to represent knowledge.
Knowledge is structured, in that concepts are used to
represent classes of individuals and roles are used to represent
relationships between concepts. Different DLs have
different expressive powers and different reasoners with different
1The alternative, in which the domain may have changed, is
belief update, which we do not consider here.
complexities. E L is one of the simplest DLs, yet is
expressive enough to have found significant application.
Consequently it is of interest to consider change within an E L
knowledge base. A further point of interest in addressing
change within a DL is that, because of the constructs for
structuring knowledge, heuristics can be introduced to
select the most “reasonable”, or perhaps commonsense,
contraction from among a set of candidate contractions.</p>
      <p>
        A key feature of E L is that an E L knowledge base (KB)
can be translated in polynomial time into a normal form of
“small” independent assertions. Then contracting a formula
from a KB can be most easily accomplished by, first
determining all minimal sets of assertions (called kernels) that
imply and then removing an element of each kernel from
the KB. This means of contraction is called kernel
contraction, and was first studied by
        <xref ref-type="bibr" rid="ref15">Hansson (1994)</xref>
        .
      </p>
      <p>The problem of determining kernels has been previously
addressed (as described in the next section), so in this paper
we focus on the second part, determining which formulas
to remove from each kernel. We look at some basic
implementations as well as more advanced ones. Our focus
is on investigating the algorithmic aspects of these different
approaches. Algorithmic time complexity is taken into
consideration as well as the overall suitability of the solutions
found. “Suitability” can be measured in different ways,
including optimality, but also in terms of what makes more
sense as a “reasonable” solution.</p>
      <p>We begin by examining simplistic algorithms for
removing formulas from a kernel, and move on to examining things
from the point of view of the hitting set problem, where
the hitting set problem is simply the problem of finding the
smallest set of elements that intersect with a given set of sets
of elements. However, approaches that only consider the
size of the solution do not always make a meaningful choice
on which beliefs to remove. A contributions of this study is
the subsequent investigation of algorithms that perform
kernel contraction based on the structure of a E L knowledge
base. These approaches exploit the structure of E L
knowledge bases to find most reasonable solutions. A first
approach is to select beliefs that are somehow related to each
other, rather than selecting the beliefs that are not. This idea
motivates the heuristic that we call localization. A second
heuristic we introduce is specificity, which is based on the
idea that removing beliefs about more specific matters is
more reasonable than removing beliefs about more general
ones. So we can then use these heuristics as a tie-breaker
step when we get more than one solution.</p>
      <p>
        The next section covers background material, including
the prominent approaches to belief change, and a brief
introduction to description logic, focusing on E L. The following
section covers basic approaches to belief contraction in E L
while the next section discusses commonsense heuristics,
specifically dealing with localization and specificity.
Further details and full descriptions of algorithms may be found
in
        <xref ref-type="bibr" rid="ref11">(Dawood 2017)</xref>
        .
      </p>
    </sec>
    <sec id="sec-2">
      <title>Background</title>
      <p>
        In this section, we provide some background on the main
building blocks of our work. We discuss belief change
first with respect to the AGM framework
        <xref ref-type="bibr" rid="ref2">(Alchourro´n,
Ga¨rdenfors, and Makinson 1985)</xref>
        and second according to
        <xref ref-type="bibr" rid="ref16">Hansson (1999)</xref>
        . Following this, Description Logics (DLs)
are briefly introduced, focusing on the DL E L.
      </p>
      <sec id="sec-2-1">
        <title>Belief Change</title>
        <p>
          AGM Belief Change The AGM framework
          <xref ref-type="bibr" rid="ref2">(Alchourro´n,
Ga¨rdenfors, and Makinson 1985)</xref>
          , named after its founders,
is the most prominent framework for belief change. Belief
change is described on an abstract level, independent of how
beliefs are represented and manipulated. Belief states are
modeled by sets of sentences, called belief sets, closed
under the logical consequence operator of a logic that includes
classical propositional logic in a language L. Thus a belief
set K satisfies the constraint:
        </p>
        <sec id="sec-2-1-1">
          <title>If K logically entails then</title>
          <p>2 K:</p>
          <p>Three types of belief change are identified: expansion,
revision, and contraction. For expansion, a new belief is
simply added to a belief set and the deductive closure taken.
That is, for belief set K and formula , the expansion of K
by is defined by</p>
          <p>K +
:
= Cn(K [ f g)
where Cn( ) is the deductive closure of its argument.
Expansion can be reasonably applied when new information is
consistent with a belief set; it is also useful in specifying the
postulates for revision and contraction, and in their relation.</p>
          <p>Revision represents the situation in which new
information may be inconsistent with the reasoner’s beliefs, and
needs to be incorporated in a consistent manner where
possible. Consequently, if a new formula is inconsistent with
respect to an agent’s beliefs, some beliefs have to be dropped
before the formula can be consistently incorporated.</p>
          <p>
            Contraction represents the situation in which the reasoner
loses information. Informally, the contraction of a belief set
by a formula is another belief set in which that formula is not
believed. We denote the belief set resulting from contracting
the belief set K with respect to sentence by K . Since
contraction is the subject of this study, we discuss the
rationality criteria (or postulates) of such change. In the AGM
framework, the following eight postulates are introduced as
a basis for ensuring the rationality of contraction functions
(more details can be found in
            <xref ref-type="bibr" rid="ref18">(Peppas 2008)</xref>
            and
            <xref ref-type="bibr" rid="ref13">(Ga¨rdenfors
1988)</xref>
            ):
(K-1) K
(K-2) K
(K-3) If
is a belief set.
          </p>
          <p>The fifth postulate, the so-called recovery postulate, is
controversial; it asserts that following a contraction, all beliefs
are recoverable by an expansion by the formula for
contraction.</p>
          <p>
            In
            <xref ref-type="bibr" rid="ref13 ref2">(Alchourro´n, Ga¨rdenfors, and Makinson 1985;
Ga¨rdenfors 1988)</xref>
            , various constructions, including partial
meet and epistemic entrenchment are developed, and shown
to exactly capture the set of functions that conform to the
contraction postulates; we omit details, as they are not
directly pertinent here.
          </p>
          <p>Both revision and contraction are formulated to follow a
principle of informational economy. In the case of
contraction, this informally means not giving up a formula if there
is no need to do so. Compared to revision, contraction is
usually seen as a more basic change function, since a belief
set can only decrease. However, revision can be regarded
as a combination of contraction and expansion, via the Levi
Identity:</p>
          <p>
            K
= (K
: ) + :
Belief Bases The AGM approach presents belief change
at a high level, independent of how a belief is represented,
and where the agent’s beliefs are described by a
deductivelyclosed belief set.
            <xref ref-type="bibr" rid="ref16">Hansson (1999)</xref>
            presents an approach to
belief change where the agent’s beliefs are represented by an
arbitrary set of formulas, called a belief base. The intuition
is that a belief base may be “closer” to how an agent’s beliefs
may be represented in a finite computer.
          </p>
          <p>Two constructions for a belief base contraction B are
described:
1. Remainders: Compute all maximal subsets of B that do
not entail . The contraction is defined as the intersection
of some “select” set of these subsets.
2. Kernels: Compute all minimal subsets of B that entail .</p>
          <p>Remove (in some fashion) an element of each such set to
obtain a belief base that does not entail .</p>
          <p>Here, we consider kernel contraction on belief bases. The
function in a kernel contraction that removes an element of
each subset is called an incision function.</p>
          <p>For example, consider the following set where we wish to
remove the element b:
fa; a ! b; b; cg</p>
          <p>It is not enough to simply remove b from the set; if we do so,
we get: fa; a ! b; cg which still implies b. In this example,
there are two kernels, fbg and fa; a ! bg. An incision
function must remove fbg and one of a, a ! b. The two
possible resulting belief bases after contraction are:
fa; cg and fa ! b; cg</p>
          <p>In the following, B will refer to an arbitrary belief base.
We refer to the set of -kernels of B by B ? .</p>
          <p>
            Definition 1 (Kernel Set
            <xref ref-type="bibr" rid="ref16">(Hansson 1999)</xref>
            ) For a belief base
B and a formula , B ? is a set such that X 2 B ? if
and only if:
          </p>
          <p>is a kernel set that contains all -kernels of B.</p>
          <p>Kernels are the smallest subsets of the knowledge base that
imply a specific belief. Therefore, in order to give up that
belief, every one of those kernels needs to be incised. If
the kernels are not minimal, they might contain a belief that
does not contribute to the implication of the belief to be
contracted. For example, in fp; qg, q is a belief that does not
contribute to make the set imply p. In this case, removing q
only would not help in giving up p.</p>
          <p>
            An incision function is defined in as follows:
Definition 2 (Incision function
            <xref ref-type="bibr" rid="ref16">(Hansson 1999)</xref>
            ) An
incision function is a function such that for all beliefs :
Contraction is carried out by removing the beliefs that are
selected by the incision function from the original
knowledge base. Contraction using the incision function 2 can be
defined as:
Definition 3
            <xref ref-type="bibr" rid="ref16">(Hansson 1999)</xref>
            (Contraction)
          </p>
          <p>B
= B r
(B ?
)</p>
          <p>Hansson provides the following postulates that are shown
to exactly capture belief base contraction based on incision
functions:
(B-1) If 6`
(B-2) B
then B</p>
          <p>B
(B-3) If 2 B and
such that B0 6`</p>
          <p>62 B
but B0 [
(B-4) If for every B0</p>
          <p>B = B .</p>
          <p>6`</p>
          <p>`
B we have B0 `</p>
          <p>(Success)
(Inclusion)
As well, Hansson provides the following postulate which,
along with the preceding, characterises smooth kernel
contraction:
(B-5) B \ Cn(B
)</p>
          <p>B
2Technically the contraction function should be
parameterised by . For simplicity and because no confusion arises, we
omit it.
then there is a set B0 B
(Core retainment)
iff B0 ` , then
(Uniformity)</p>
        </sec>
        <sec id="sec-2-1-2">
          <title>See (Hansson 1999) for details.</title>
          <p>Although the AGM framework is very popular, we find
the postulates introduced by Hansson more suitable to be
applied to kernel contraction. First, in our approach we work
with a normalized form of an E L knowledge base, which is
to say, a belief base of formulas. Second, as we show at the
end of this section, contraction of E L belief bases do not
satisfy the recovery postulate of the AGM framework.</p>
        </sec>
      </sec>
      <sec id="sec-2-2">
        <title>Description Logic</title>
        <p>A Description Logic (DL) is one of a class of logics based on
describing sets of individuals using concepts and describing
the relationships between them using roles. Two types of
knowledge are represented: Intensional knowledge is
general knowledge about a problem or a domain, while
extensional knowledge is knowledge about a specific problem
instance. For this purpose, a knowledge base3 is composed of
two corresponding components: A TBox containing the
intensional knowledge and an ABox containing the extensional
knowledge.</p>
        <p>The ABox gives assertions, much like in a relational
database, such as Female(Sally) and FatherOf(Adam, Sally).
Since our focus is on contraction in TBoxes, we don’t
consider the ABox further. A TBox is composed of: a set of
assertions, made up of equivalences such as Man = Human
u Male (the individuals satisfying Man are exactly those
satisfying Human and Male) and subsumptions such as Man v
Human (every individual that is a Man is a Human). The
latter are referred to as General Concept Inclusions (GCIs).
Every equivalence can be trivially broken down into two
GCIs.</p>
        <p>
          There are many members of the DL family; they vary
widely in expressivity power and the complexity of their
reasoning algorithms. In the following section we discuss
a well-known member of the DL family, E L, which we will
use the representation language for the rest of this study.
The E L Description Logic E L is a description logic that
has recently attracted much attention. It is a “light-weight”
DL, in that it is limited expressively, although it is used in
well-known ontologies such as SNOMED CT
          <xref ref-type="bibr" rid="ref19">(Spackman
2000)</xref>
          , a large-scale commercial medical ontology with well
over 300000 concepts
          <xref ref-type="bibr" rid="ref7">(Baader 2011)</xref>
          . E L contains the
following concept constructors:
        </p>
        <p>Conjunction (u):
If C1 and C2 are concepts, then C1 u C2 is a concept
standing for the conjunction of C1; C2.</p>
        <p>Existential restriction (9):
If C is a concept and R is a role, then 9R:C is the concept
of those individuals that are R-related to an item
belonging to concept C.</p>
        <p>The top concept &gt;, true of all individuals.</p>
        <p>Assertions express either equivalences between concepts
or a subsumption relation. The following are example
assertions in E L.</p>
        <p>3By knowledge base we refer to a belief base – a set of sentences
that represent the beliefs of an agent.</p>
        <p>E L is simple and easy to use; as well it has a
polynomialtime subsumption algorithm, where the subsumption
algorithm determines whether a specific subsumption relation
holds or not in a given knowledge base.</p>
        <p>The subsumption algorithm is made up of four steps:</p>
        <sec id="sec-2-2-1">
          <title>1. Normalize the TBox.</title>
        </sec>
        <sec id="sec-2-2-2">
          <title>2. Translate the normalized TBox into a graph.</title>
        </sec>
        <sec id="sec-2-2-3">
          <title>3. Complete the graph using completion rules.</title>
          <p>4. Read off the subsumption relationships from the
normalized graph.</p>
          <p>
            We are most interested in the first step, normalization. It
is shown
            <xref ref-type="bibr" rid="ref4 ref5">(Baader et al. 2007)</xref>
            that any E L TBox can be
translated in polynomial time into a set of GCIs of the following
form:
          </p>
          <p>A v B
A1 u A2 v B
A v 9r:B
9r:A v B
A, A1, A2, and B are concept names, and r is a role name. A
normalized TBox can be thought of as a knowledge base that
has been translated into a canonical form, where each GCI
represents an individual item of information. For our
contraction functions, we will work with a normalized TBox.
Such TBoxes can be thought of as corresponding to a
(nonclosed) equivalent of a belief set.</p>
        </sec>
      </sec>
      <sec id="sec-2-3">
        <title>Related work</title>
        <p>
          There has been some related work regarding contraction in
light-weight description logics. In
          <xref ref-type="bibr" rid="ref1 ref10 ref21 ref9">(Booth, Meyer, and
Varzinczak 2009a)</xref>
          a contraction operator is introduced that
contracts E L belief sets (rather than belief bases) using
remainders. A representation result is given relating the first six
contraction postulates, adapted to E L, to the construction of
infra-remainder sets
          <xref ref-type="bibr" rid="ref1 ref10 ref21 ref9">(Booth, Meyer, and Varzinczak 2009b)</xref>
          .
        </p>
        <p>
          <xref ref-type="bibr" rid="ref21">Zhuang and Pagnucco (2009)</xref>
          also address belief
contraction in E L. They point out that the recovery postulate
doesn’t hold in E L. Informally the problem, as in other
inferentially-weak approaches (see e.g.
          <xref ref-type="bibr" rid="ref12">(Delgrande 2008)</xref>
          regarding Horn clause theories), is that E L doesn’t support
negation or disjunction. For a problematic example,
consider where the TBox contains just the closure of A v B
and one wishes to contract by A u C v B; in the
contraction, A v B must be removed, but it is not “recovered” if
one expands the result by A u C v B. The authors
propose a suitable alternative to the recovery postulate based on
the notion of logical difference, and develop algorithms that
implement this approach.
        </p>
        <p>
          A model-based approach to contraction of DL-Lite belief
bases is introduced in
          <xref ref-type="bibr" rid="ref22">(Zhuang et al. 2016)</xref>
          . The approach
depends on defining an alternative semantics, called type
semantics, that skirts the problem that a knowledge base with
a finite signature may have an infinite number of models.
from a belief base
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Kernel Contraction</title>
      <p>For the kernel contraction of a formula
B there are three steps:
1. Determine the kernels K of B with respect to .
2. Determine a set of incised formulas I from K.
3. Set B B n I.</p>
      <p>
        The third step is trivial. As well, the first step has been
addressed, in the pinpointing algorithm introduced in
        <xref ref-type="bibr" rid="ref4 ref5">(Baader,
Pen˜aloza, and Suntisrivaraporn 2007)</xref>
        .
      </p>
      <p>
        The pinpointing algorithm computes all kernels using a
modified version of the E L subsumption algorithm. It finds
a monotone Boolean formula called a pinpointing formula;
the propositions of the pinpointing formula are GCIs of the
TBox, and a propositional valuation represents the kernels.
The approach may take exponential time (w.r.t the size of
the TBox) to find all kernels, because there may be
exponentially many kernels. However,
        <xref ref-type="bibr" rid="ref4 ref5">(Baader, Pen˜aloza, and
Suntisrivaraporn 2007)</xref>
        also gives a polynomial-time
algorithm that computes a single kernel.
      </p>
      <p>
        Example 1 (
        <xref ref-type="bibr" rid="ref4 ref5">(Baader, Pen˜aloza, and Suntisrivaraporn
2007)</xref>
        ) Let the TBox, T , be:
ax1 : A v 9r:A,
ax3 : 9r:Y v B,
We have that T j= A v B. Let
ax2 : A v Y ,
ax4 : Y v B.
      </p>
      <p>= A v B. We obtain:</p>
      <p>T ? (A v B) = ffax2; ax4g; fax1; ax2; ax3gg
Since we can use the pinpointing algorithm to get the set
of all kernels, we just need to consider the issue of removing
an element from each kernel to perform a contraction. This
is the role of the incision function. We next look at some
basic incision functions to set the stage for later
considerations.</p>
      <sec id="sec-3-1">
        <title>Incision Functions</title>
        <p>We are given an E L TBox, T , and a formula for contraction
. We assume without loss of generality that T has been
normalized; this has the important consequence that for 2
T that T n f g 6j= .</p>
        <p>The main contraction algorithm is shown in Algorithm 1;
as discussed, the only issue is specifying the incision
function CUT.</p>
        <sec id="sec-3-1-1">
          <title>Algorithm 1 Contraction algorithm</title>
          <p>1: procedure CONTRACT(T , A)
2: kernelset = PINPOINT(T , A)
3: giveUpSet = CUT(kernelset)
4: T = T / giveUpSet</p>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>5: end procedure</title>
        <p>
          We next consider simple ways of implementing CUT. Due
to space considerations, here and elsewhere we omit the
pseudocode; see
          <xref ref-type="bibr" rid="ref11 ref17">(Dawood 2017; Liao 2014)</xref>
          . We first
consider non-optimal solutions, followed by optimal solutions.
Basic Incision Functions The first incision function is
overly simplistic, but serves to provide a base case. In this
case, one simply removes a formula from each kernel. The
time complexity is O(m), where m is the number of
kernels (size of the kernelset), assuming that a formula
selection takes constant time. However, this is clearly a poor
algorithm. For example, given a kernel set ffa; cg; fb; cgg,
one might remove a from the first set followed by c from the
second set. Then the result of giveU pSet = fa; cg is f g
b ,
which unnecessarily removes a. Removing a is unnecessary
because removing c alone is enough for contraction. More
generally, removing more beliefs than needed is undesired,
as it violates the informational economy principle.
        </p>
        <p>The second incision function improves on the previous by
selecting a formula from every kernel, as before. However
each time a formula is selected, every kernel that contains
that formula is excluded from consideration in the following
iterations. This algorithm has a worst-case time
complexity of O(m2): There are basically two loops, first scanning
through the kernels, and second, once a formula has been
selected, scanning the remaining kernels to see whether they
contain that formula.4</p>
        <p>
          The third function implements a greedy approach: At
each step, the algorithm finds the formula that appears in
the most kernels and removes it; these kernels are not
considered in the following steps. An sketch of the
implementation in
          <xref ref-type="bibr" rid="ref11">(Dawood 2017)</xref>
          is given as follows. First, for every
formula in a kernel, the number of kernels in which that
formula occurs is determined. Then the formula that has the
greatest number of occurrences is added to the set of incised
formulas; the kernels containing that formula are removed
from consideration; the count of formula occurrences in
kernels is updated and the process continued. If the number of
formulas in kernels is n, the number of kernels is m, and the
size of the largest kernel is k, then the time complexity is
O(m2 k n). This however could be significantly enhanced
by using more clever data structures.
        </p>
        <p>
          Hitting Set Approach According to the principal of
information economy
          <xref ref-type="bibr" rid="ref13">(Ga¨rdenfors 1988)</xref>
          , in every change to
an agent’s beliefs, loss of information should be minimized.
To satisfy the requirement of minimum change we require
an algorithm that removes the fewest GCIs while hitting all
the kernels. This however is exactly the minimal hitting set
problem
          <xref ref-type="bibr" rid="ref14">(Garey and Johnson 1979)</xref>
          , which has already
received extensive study.
        </p>
        <p>
          Consequently, another approach to cutting kernels in
order to perform contraction is the minimal hitting set
algorithm. The minimal hitting set problem is defined as follows
          <xref ref-type="bibr" rid="ref1 ref21">(Abreu and van Gemund 2009)</xref>
          :
Definition 4 Given a set S = fs1; s2; :::; sng of n
nonempty sets, a minimal hitting set d is a set such that:
Thus, d is a minimal hitting set if and only if it contains at
least one element from every set, and no proper subset of d
4This assumes that looking up a formula in a kernel takes
constant time, for example if kernels are stored as hash tables.
is a hitting set. In the context of kernel contraction in E L,
we can define the minimal hitting set contraction as:
Definition 5 (Minimal hitting set contraction) Given a
kernel set K = fk1; k2; :::; kng of n kernels, a minimal hitting
set giveUpSet is a set such that:
8ki2K [ki \ giveU pSet 6= ;]
^
@giveUpSet0 giveUpSet[8ki2K (ki \ giveU pSet0 6= ;)]
Since kernels are subsets of the TBox, and our goal to find
a giveU pSet that hits all kernels, we can use approaches to
the minimal hitting set problem to solve it. The computation
of all such hitting sets is known to be NP-hard
          <xref ref-type="bibr" rid="ref14">(Garey and
Johnson 1979)</xref>
          ; there are however known very good
approximation algorithms, such as the one introduced in
          <xref ref-type="bibr" rid="ref1 ref21">(Abreu and
van Gemund 2009)</xref>
          , for the minimal hitting set problem, that
are feasible in terms of running time.
        </p>
        <p>The following result shows that we obtain a well-behaved
contraction function:
Theorem 1 Let T be a normalized TBox, a (normalized)
set of GCIs, and with kernel set K and set giveU pSet
determined as in Definition 5. Then for defined by: T =
T n giveU pSet, we have that T satisfies the postulates
for smooth kernel contraction (B-1)-(B-5).</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Heuristics for Contraction</title>
      <p>In the last section, we discussed contraction algorithms
where the goal was to remove the least number of beliefs.
Here we consider heuristics based on the E L language and
subsumption hierarchy. The idea is to bring in commonsense
principles that may be expected to yield intuitive or
“reasonable” results for a contraction. With localization we choose
a hitting set that in some fashion affects a minimal region of
the TBox; with specificity we prefer hitting sets that refer to
more specific concepts.</p>
      <sec id="sec-4-1">
        <title>Localization</title>
        <p>The idea here is that change should be as local as possible in
a knowledge base. Intuitively, this will involve a lesser
overall change to a KB; as well, pragmatically, it would seem that
if a belief is incorrect, then it may be expected to arise from
a specific part of the KB (much like in diagnosis where
pragmatically one would prefer a diagnosis with fewer faults).</p>
        <p>Consider following E L example:</p>
        <p>(1) A v B; (2) B v D; (3) A v C; (4) C v D.
These imply that (5) A v D. We see that there are two
(5)-kernels: f1, 2g and f3, 4g</p>
        <p>So, to contract the knowledge base by (5), we need to
remove a belief from each kernel. However, removing (1) and
(4) seems unreasonable because it means removing
information concerning all four concepts. Removing (1) and (3), on
the other hand, removes knowledge concerning only three
concepts: A, B, and C. Not only that, but pragmatically,
(1) and (3) together imply that perhaps our knowledge of
A is wrong; conceivably our knowledge of B and C is not
necessarily false. Similar considerations apply to removing
(2) and (4), which together would suggest that the source of
incorrectness lies with D.</p>
        <p>
          Localization is implemented in
          <xref ref-type="bibr" rid="ref17">(Liao 2014)</xref>
          using a
graphbased approach. The algorithm operates by defining a graph
based on a set of GCIs, where an edge connects GCIs
sharing a concept or role. Then notions involving graph
connectivity and graph clustering can be used to determine
localized change.
        </p>
        <p>Definition 6 (Connected GCIs) Given a set of GCIs K,
define an undirected graph G = (V; E) by: V is the set of
GCIs in K and for g1, g2 2 K, (g1; g2) 2 E just if g1 and g2
share a concept or role name.</p>
        <p>So, for a normalized TBox, given that the set of kernels
has been determined and a set of minimal hitting sets H =
fh1; : : : ; hng has also been found, localization can be used
to select the preferred hitting set(s), based on the structure
of each hitting set:
1. For each hi 2 H, determine the underlying graph
(Definition 6), and determine the number of connected
components, ci, in the graph.
2. Select a hitting set hi whose number of connected
components, ci, is minimum.</p>
        <p>In the previous example, we find, the following candidate
sets to contract (5):</p>
        <p>f1, 3g and f2, 4g.</p>
        <p>Thus contraction is given by one of these alternatives, and
not the inferior possibilities f1; 4g or f2; 3g.</p>
        <p>
          For time complexity, each hitting set must be considered;
and for a hitting set the number of connected components
determined. This latter problem can be solved via
          <xref ref-type="bibr" rid="ref20">Tarjan’s (1972</xref>
          ) algorithm, which is O(jV j + jEj) for a graph
G = (V; E). Another pass through the hitting sets finds the
minimum, for overall time complexity O(m (jV j + jEj)).
Since all we are doing is selecting a particular hitting set, as
given in Definition 5 the resulting contraction operator is a
smooth kernel contraction, as given in Theorem 1.
        </p>
        <p>Clearly this procedure can be enhanced or modified.
A more sophisticated notion of localization could be
employed, for example, or the localization heuristic could be
applied just to the hitting sets of minimum cardinality (and
so help serve as a tie-breaker).</p>
      </sec>
      <sec id="sec-4-2">
        <title>Specificity</title>
        <p>The assumption underlying this heuristic is that in
choosing which GCIs to remove, it is preferable to remove those
involving more specific concepts over those involving less
specific concepts: Arguably more general concepts are more
entrenched; as well, removing a more general concept
subsumption may be expected to be more disruptive to the TBox
as a whole.</p>
        <p>The subsumption hierarchy implicit in a normalized E L
TBox categorizes concepts into different levels of generality.
Consider the relations:</p>
        <p>A v B; B v C; C v D .</p>
        <p>Suppose we would like to contract the TBox by the implied
relation A v D. We have three options, corresponding to
removing one of the three given subsumptions. Removing
any of the three GCIs would guarantee minimum change.
However, we can say that A is more specific than B and B
is more specific than C. Removing A v B, in this case, is
more reasonable than removing C v D, because A v B
involves concepts that are more specific than the ones involved
in C v D. This also means that removing C v D results
in all individuals that are Cs no longer believed to be Ds,
while removing A v B only affects the subset of the Cs
which are As.</p>
        <p>To begin, for a GCI A v B, we say that A is a child
concept relative to B. Similarly, for A1 u A2 v B, we say that
A1 and A2 are both child concepts relative to B. Moreover,
we ignore the structure of a concept expression of the form
9R:C, which is to say, we treat a concept expression 9R:C
as if it were a primitive concept name.</p>
        <p>The approach assigns weights to every GCI, where a
weight reflects its relative generality. Then we select the
kernels with minimum overall weight. A GCI A v B gets
weight 0 if A has no children; and a GCI A v B gets weight
i+1 if the maximum weight of the GCIs of form X v A is i.
Similarly, a GCI A1 u A2 v B gets weight i + 1 if the
maximum weight of the GCIs of form X v Aj for j 2 f1; 2g
is i. For simplicity, we assume that the TBox is acyclic – so
we avoid the problems that will be caused by loops.</p>
        <p>In more detail, let A and B be concept expressions
appearing in a normalized TBox.
1. If A has no children, assign A v B a weight of 0. If A1,</p>
        <p>A2 have no children, assign A1 u A2 v B a weight of 0.
2. For GCI A v B (A1 u A2 v B), assign the GCI a weight
of i + 1 if the maximum weight of GCIs of the form X v
A (resp. X v Aj j 2 f1; 2g) is i.</p>
        <p>Then, given that a set of hitting sets H = fh1; : : : ; hng
has been found, choose the preferred hitting set hi where the
sum of the weights of the GCIs in hi is a minimum.</p>
        <p>As a simple example, consider the previous example:
A v B has weight = 0
B v C has weight = 1</p>
        <p>C v D has weight = 2
In this case, the hitting set consisting of the GCI A v B is
selected.</p>
        <p>The time complexity of the underlying algorithm is
polynomial. Let m be the size of the normalized TBox T , and n
the number of hitting sets. Determining the depth labeling
of GCIs in T takes linear time, and determining the weights
of the hitting sets is O(n k) where k is the maximum size of
a hitting set. Hence the overall complexity is O(m + (n k)).
As with localization, the resulting operator is a smooth
kernel contraction.</p>
        <p>Again, the basic approach to specificity can be enhanced
or modified. For example, one could just look at hitting
sets of minimum cardinality, and so the specificity heuristic
would be a tie-breaker for equal-sized hitting sets. As well,
there are other ways of determining overall specificity of a
set of GCIs. One appealing candidate is to choose the hitting
set hi whose maximally-weighted GCI is a minimum. That
is: rank the hitting sets by the maximum weight assigned
to a contained GCI, and then choose the minimum ranked
hitting set in the resulting ranking.</p>
        <p>Last, it might be objected that since there may be an
exponential number of hitting sets, any approach is bound to be
intractable in the worst case. This difficulty can be
pragmatically addressed by employing an anytime algorithm for
generating an acceptable candidate: simply run through hitting
sets, keeping track of the best candidate found so far, until
either a “good enough” candidate is found, a time bound is
hit, or some other stopping criterion is met.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Conclusion</title>
      <p>In this paper we have studied belief contraction in the
description logic E L. The overall goal is to apply the formal
theory of belief base change in a practical setting. To this
end, kernel contraction seems to most appropriate approach
for contraction. We chose to work with E L first, because
it used in large-scale knowledge bases and second, because
the E L constructs for structuring knowledge allow the
specification of heuristics to choose intuitive, commonsense
outcomes for a contraction. We first examined ways of
finding (or approximating) hitting sets. We also proposed two
heuristics, based on localization and specificity, for
determining suitable hitting sets for contraction. Our main
contraction operators are shown to conform to Hansson’s
postulates for smooth base contraction. All of our algorithms are
efficient, being polynomial in (one or both) of the size of the
TBox or the number of hitting sets. The bottleneck then lies
in the number of hitting sets, which may be exponential in
the size of the TBox.</p>
      <p>For future work there are two main ways to go. First,
the present heuristics can be further explored, elaborated on,
and variants developed; as well, these approaches should be
implemented to determine what works best in practice.
Second, is to explore belief change in a more expressive
description logic. To this end, there are enhancements to E L
that preserve the polynomial-time features, but also bring
in a notion of consistency (by allowing the concept ?), for
which the present approach should apply without change.
This would allow a definition of revision in E L, in terms
of the Levi Identity. As well, belief change in significantly
more expressive DLs, such as in the ALC family, can also
be addressed.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgements</title>
      <p>We thank the three reviewers for their very helpful
comments. Financial support is gratefully acknowledged from
the Natural Sciences and Engineering Research Council of
Canada.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <surname>Abreu</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          , and van Gemund,
          <string-name>
            <surname>A. J. C.</surname>
          </string-name>
          <year>2009</year>
          .
          <article-title>A low-cost approximate minimal hitting set algorithm and its application to model-based diagnosis</article-title>
          . In Bulitko, V., and
          <string-name>
            <surname>Beck</surname>
          </string-name>
          , J. C., eds., Eighth Symposium on Abstraction, Reformulation, and
          <string-name>
            <surname>Approximation</surname>
          </string-name>
          (SARA
          <year>2009</year>
          ). AAAI Press.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <surname>Alchourro´n</surname>
          </string-name>
          , C.; Ga¨rdenfors, P.; and
          <string-name>
            <surname>Makinson</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <year>1985</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <article-title>On the logic of theory change: Partial meet contraction and revision functions</article-title>
          .
          <source>Journal of Symbolic Logic</source>
          <volume>50</volume>
          (
          <issue>2</issue>
          ):
          <fpage>510</fpage>
          -
          <lpage>530</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <surname>Baader</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Calvanese</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>McGuiness</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Nardi</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Patel-Schneider</surname>
          </string-name>
          , P., eds.
          <year>2007</year>
          .
          <article-title>The Description Logic Handbook</article-title>
          . Cambridge: Cambridge University Press, second edition.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <surname>Baader</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ; Pen˜aloza, R.; and
          <string-name>
            <surname>Suntisrivaraporn</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <article-title>Pinpointing in the description logic E L</article-title>
          .
          <source>In Proceedings of the 2007 International Workshop on Description Logics (DL2007)</source>
          , volume
          <volume>250</volume>
          <source>of CEUR-WS</source>
          ,
          <fpage>171</fpage>
          -
          <lpage>178</lpage>
          . Brixen/Bressanone, Italy: Bozen-Bolzano University Press.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <surname>Baader</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <year>2011</year>
          .
          <article-title>What's new in description logics</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <source>Informatik-Spektrum</source>
          <volume>34</volume>
          (
          <issue>5</issue>
          ):
          <fpage>434</fpage>
          -
          <lpage>442</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <surname>Booth</surname>
          </string-name>
          , R.; Meyer, T.; and
          <string-name>
            <surname>Varzinczak</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          <year>2009a</year>
          .
          <article-title>First steps in E L contraction</article-title>
          .
          <source>In Workshop on Automated Reasoning about Context and Ontology Evolution (ARCOE-09)</source>
          ,
          <fpage>16</fpage>
          -
          <lpage>18</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <surname>Booth</surname>
          </string-name>
          , R.; Meyer, T.; and
          <string-name>
            <surname>Varzinczak</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          <year>2009b</year>
          .
          <article-title>Next steps in propositional Horn contraction</article-title>
          .
          <source>In Proceedings of the International Joint Conference on Artificial Intelligence</source>
          ,
          <fpage>702</fpage>
          -
          <lpage>707</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <surname>Dawood</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <year>2017</year>
          .
          <article-title>An algorithmic study of kernel contraction in E L. Master's thesis</article-title>
          , Simon Fraser University.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <surname>Delgrande</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <year>2008</year>
          .
          <article-title>Horn clause belief change: Contraction functions</article-title>
          . In Brewka, G., and
          <string-name>
            <surname>Lang</surname>
          </string-name>
          , J., eds.,
          <source>Proceedings of the Eleventh International Conference on the Principles of Knowledge Representation and Reasoning</source>
          ,
          <volume>156</volume>
          -
          <fpage>165</fpage>
          . Sydney, Australia: AAAI Press.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <string-name>
            <surname>Ga</surname>
          </string-name>
          ¨rdenfors, P.
          <year>1988</year>
          .
          <article-title>Knowledge in Flux: Modelling the Dynamics of Epistemic States</article-title>
          . Cambridge, MA: The MIT Press.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <string-name>
            <surname>Garey</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Johnson</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <year>1979</year>
          .
          <article-title>Computers and Intractability: A Guide to the Theory of NP-Completeness.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <string-name>
            <surname>Hansson</surname>
            ,
            <given-names>S. O.</given-names>
          </string-name>
          <year>1994</year>
          .
          <article-title>Kernel contraction</article-title>
          .
          <source>Journal of Symbolic Logic</source>
          <volume>59</volume>
          (
          <issue>3</issue>
          ):
          <fpage>845</fpage>
          -
          <lpage>859</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <string-name>
            <surname>Hansson</surname>
            ,
            <given-names>S. O.</given-names>
          </string-name>
          <year>1999</year>
          .
          <article-title>A Textbook of Belief Dynamics</article-title>
          . Applied Logic Series. Kluwer Academic Publishers.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <string-name>
            <surname>Liao</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <year>2014</year>
          . Kernel contraction in E L.
          <article-title>Master's thesis</article-title>
          , School of Computing Science, Simon Fraser University.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          <string-name>
            <surname>Peppas</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <year>2008</year>
          .
          <article-title>Belief revision</article-title>
          . In van Harmelen,
          <string-name>
            <given-names>F.</given-names>
            ;
            <surname>Lifschitz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            ; and
            <surname>Porter</surname>
          </string-name>
          , B., eds.,
          <source>Handbook of Knowledge Representation. Elsevier Science</source>
          .
          <volume>317</volume>
          -
          <fpage>359</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          <string-name>
            <surname>Spackman</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <year>2000</year>
          .
          <article-title>Managing clinical terminology hierarchies using algorithmic calculation of subsumption: Experience with SNOMED-RT</article-title>
          .
          <source>J. of the Amer. Med</source>
          . Inf. Assn.
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          <string-name>
            <surname>Tarjan</surname>
            ,
            <given-names>R. E.</given-names>
          </string-name>
          <year>1972</year>
          .
          <article-title>Depth-first search and linear graph algorithms</article-title>
          .
          <source>SIAM J. Comput</source>
          .
          <volume>1</volume>
          (
          <issue>2</issue>
          ):
          <fpage>146</fpage>
          -
          <lpage>160</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          <string-name>
            <surname>Zhuang</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Pagnucco</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <year>2009</year>
          .
          <article-title>Belief contraction in the description logic E L</article-title>
          .
          <source>In Proceedings of the 22nd International Workshop on Description Logics (DL2009)</source>
          , volume
          <volume>477</volume>
          <source>of CEUR-WS.</source>
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          <string-name>
            <surname>Zhuang</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ; and Qi,
          <string-name>
            <surname>G.</surname>
          </string-name>
          <year>2016</year>
          .
          <article-title>DL-Lite contraction and revision</article-title>
          .
          <source>Journal of Artificial Intelligence Research</source>
          <volume>56</volume>
          :
          <fpage>328</fpage>
          -
          <lpage>378</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>