<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Concept Abduction for Description Logics</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Birte Glimm</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Yevgeny Kazakov</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Michael Welt</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Ulm University</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>We present two alternative algorithms for computing (all or some) solutions to the concept abduction problem: one algorithm is based on Reiter's hitting set tree algorithm, whereas the other one relies on a SAT encoding. In contrast to previous work, the algorithms do not rely on a refutation-based calculus and, hence, can be used also with eficient reasoners for tractable DLs such as ℰℒ and its extensions. An adaptation to other forms of (logic-based) abduction, e.g., to ABox abduction, is also possible.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Description Logics</kwd>
        <kwd>abduction</kwd>
        <kwd>ontologies</kwd>
        <kwd>reasoning</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>but directly derive (logical) consequences, which prevents the use of existing approaches to
abduction. We aim at closing this gap and, inspired by the work of Kazakov and Glimm [10] for
computing justification (i.e., minimal subsets of axioms from a knowledge base that are needed
for a given entailment), we propose an alternative approach to eficiently compute some or
all (minimal) explanations. The approach is based on Reiter’s minimal hitting set algorithm
[11] without relying on a refutation-based calculus to guide the algorithm. An encoding of
the problem into SAT allows for better managing the combinatorial problem of selecting what
constitutes an explanation. As our goal is the development of an approach for DLs such as
ℰ ℒ, where the main reasoning task involves determining concept subsumptions, we focus on
concept abduction [12] as introduced next. The accompanying technical report gives further
details, examples, and complete proofs [13].</p>
    </sec>
    <sec id="sec-2">
      <title>2. Preliminaries</title>
      <p>We use the standard syntax and semantics of DLs (see, e.g., [14]). Note that the DL ℰ ℒ only
allows for using the top concept, conjunctions, and existential restrictions as concepts.
We now introduce the basic ideas of (concept) abduction (see also [12, 15]).</p>
      <p>Definition 1. A concept abduction problem  is a tuple ⟨, ℋ, ⟩ with  a knowledge base,
ℋ, the hypotheses, a set of atomic concepts, and , the observation, a single atomic concept.1
A set  = {1, . . . , } ⊆ ℋ is an explanation for  if  |= 1 ⊓ . . . ⊓  ⊑ . Such an 
is explanatory if 1 ⊓ . . . ⊓  ⊑  ̸∈ ,  is satisfiable if 1 ⊓ . . . ⊓  is satisfiable w.r.t.
,  is relevant if ∅ ̸|= 1 ⊓ . . . ⊓  ⊑ , and  (syntactically) minimal if there is no other
explanation ′ for  such that ′ ⊊ .</p>
      <p>Note that the given set of hypothesis allows for restricting the set of concepts that can be
used in explanations, but may also contain all concepts used in the knowledge base.</p>
      <p>In the remainder, we are interested in finding explanations that are explanatory, satisfiable,
relevant, and (syntactically) minimal. Abusing notation, for a set of concepts  = {1, . . . , },
we also write  |=  ⊑  instead of  |= 1 ⊓ . . . ⊓  ⊑ .</p>
      <p>Example 1. Consider the knowledge base  containing the axioms</p>
      <p>
        ⊑ ∃.
∃. ⊑ 
(
        <xref ref-type="bibr" rid="ref1">1</xref>
        )
(
        <xref ref-type="bibr" rid="ref2">2</xref>
        )
 ⊓ ′ ⊑ 
 ⊓  ⊑ ⊥
(
        <xref ref-type="bibr" rid="ref3">3</xref>
        )
(
        <xref ref-type="bibr" rid="ref4">4</xref>
        )
and the abduction problem  = ⟨, ℋ, ⟩ with ℋ = {, , , ′} and  = . From Axiom (
        <xref ref-type="bibr" rid="ref1">1</xref>
        )
and Axiom (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ), we have  |=  ⊑ . Hence, together with Axiom (
        <xref ref-type="bibr" rid="ref3">3</xref>
        ), we find that  |=  ⊓ ′ ⊑
 = . Among others, we have the following explanations for :
1Note that requiring the observation to be an atomic concept is without loss of generality since for an observation
in form of a complex concept , we can simply introduce an axiom  ≡  and use  as observation. We can
proceed analogously for hypotheses.
      </p>
      <p>Algorithm 1: Finding one explanation</p>
      <p>
        Minimize(, ℋ, ): compute a minimal explanation for the concept abduction problem
⟨, ℋ, ⟩
input : a knowledge base , a set of hypotheses ℋ, and an observation  such that
 |= ℋ ⊑ 
output : a minimal explanation  ⊆ ℋ such that  |=  ⊑  (cf. Definition 1)
1  ← ℋ ;
2 for  ∈ ℋ do
3 if  |= ( ∖ {}) ⊑  then
4  ←  ∖ {};
5 return ;
Among these only 1 is explanatory, relevant, satisfiable, and minimal. We have that 2 is not
explanatory since  ⊓ ′ ⊑  ∈ , and 3 is not satisfiable (due to Axiom (
        <xref ref-type="bibr" rid="ref4">4</xref>
        )). Finally, 4 is
not minimal since 1 ⊊ 4. Note that {} is not an explanation since  ∈/ ℋ, but even if it were,
the explanation would not be relevant since ∅ |=  ⊑ .
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Computing Abductive Explanations</title>
      <p>Before proposing our approach, we point out that the number of (minimal) explanations for a
given observation may be exponential in the size of the knowledge base. We start by showing
how to compute one explanation before generalizing the approach to compute all explanations,
ifrst, using hitting set trees and, then, via a SAT encoding.</p>
      <sec id="sec-3-1">
        <title>3.1. Computing One Abductive Explanation</title>
        <p>Given a concept abduction problem  = ⟨, ℋ, ⟩, a naive algorithm for computing one
explanation  such that  |=  ⊑  is relatively easy. If  ̸|= ℋ ⊑ , there is no satisfiable
explanation and we can stop. Otherwise, we start from  = ℋ and repeatedly remove concepts
from  as long as this does not break the entailment  |=  ⊑ . At a certain point, no concept
can be removed without breaking the entailment, which implies that  is a minimal explanation
w.r.t. . Algorithm 1 summarizes this idea. Note that Algorithm 1 makes calls to a reasoning
service for concept subsumption checking without relying on a particular kind of procedure
(e.g., tableau-based as well as consequence-based procedures may be used). It remains to check
whether the returned  is explanatory, satisfiable, and relevant, which requires at most two
further subsumption checks and a check for set containment.</p>
        <p>The correctness of Algorithm 1 relies on the fact that a conjunction with additional conjuncts
is more specific:
Lemma 1. Let  be a DL knowledge base,  and ′ two sets of atomic concepts such that ′ ⊆ ,
and  an atomic concept. Then  |= ′ ⊑  implies  |=  ⊑ .</p>
        <p>Proof Sketch. Intuitively, the interpretation of a conjunction is more restricted the more
conjuncts are contained in the conjunction.</p>
        <p>Note that Lemma 1 only relies on the semantics of conjunctions and subsumption, which is
shared among all DLs. We show that Algorithm 1 is correct under the assumed semantics.
Theorem 1. Let  be the output of Algorithm 1 for the input , ℋ, and  such that  |=  ⊑ .
Then  is a minimal explanation for the abduction problem ⟨, ℋ, ⟩.</p>
        <p>Proof Sketch. Since  |= ℋ ⊑  by assumption,  is initialized such that  |=  ⊑  (in
Line 1). Since we only remove a concept from , if the subsumption still holds in Line 4,
 |=  ⊑  is preserved. We can further show that  is minimal explanation as unneeded
conjuncts are step by step removed.</p>
        <p>Finally, observe that a run of Algorithm 1 requires exactly  subsumption tests, where
 = ||ℋ||, which is bounded by the number of concepts in  (two further subsumption tests
and a set containment test are needed for checking whether  is relevant, satisfiable, and
explanatory). Hence, the complexity of computing one minimal explanation is bounded by
a linear function over the complexity of subsumption checking. In particular, for tractable
languages such as ℰ ℒ and its tractable extensions [16], one (minimal) explanation for the
concept abduction problem can be computed in polynomial time, which is worst-case optimal
[12].2 For DLs where subsumption checking is ExpTime-complete such as ℒ [17, 18], the
algorithm requires worst-case exponential time.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Computing All Abductive Explanations</title>
        <p>An explanation for a concept abduction problem ⟨, ℋ, ⟩ consists of concepts that are
subsumed by the observation. As we have seen in Example 1, there can be several diferent
explanations. To find all explanations, it is, therefore, necessary to change a concept in every
explanation of ⟨, ℋ, ⟩ and the question of how to compute all (minimal) explanations arises.</p>
        <p>Note first, that the output of Algorithm 1 depends on the order in which the axioms in 
are enumerated in the for-loop (Line 2). Diferent orders of the concepts can result in diferent
removals and, consequently, diferent explanations.</p>
        <p>Lemma 2. For each (minimal) explanation  of a concept abduction problem ⟨, ℋ, ⟩, there
exists some order of concepts in ℋ for which Algorithm 1 with the input , ℋ, and  returns .
Proof Sketch. Assume, to the contrary of what is to be shown, that a minimal explanation  is
not returned. We can show that processing first concepts in ℋ ∖  and then concepts from 
leads to a contradiction as desired.
2Note that although subsumption checking for ℰℒ++ is tractable, the complexity of checking whether an explanation
exists is found to be NP-complete by Bienvenu [12]. This apparent contradiction is because unsatisfiable explanations
(which can only occur in logics that can express unsatisfiable concepts/ ⊥, as ℰℒ++) are excluded in the definition
of Bienvenu as explanations. Our algorithm, however, might terminate with an unsatisfiable explanation. If that
has to be avoided, some form of backtracking for such cases would be needed, which indeed then leads to a higher
complexity of the problem.

′
⊥</p>
        <p>⊥
′
⊥</p>
        <p>′</p>
        <p>The property stated in Lemma 2 means that for computing all explanations for  = ⟨, ℋ, ⟩,
it is suficient to run Algorithm 1 for all possible orders of concepts in ℋ. As the number of
explanations can be exponential in , the exponential behavior of an algorithm for computing
all explanations cannot be avoided in general. Unfortunately, the described algorithm is not
very practical since it performs exponentially many subsumption tests for all inputs, even if, e.g.,
 has just one explanation. This is because this algorithm is not goal-directed: the computation
of each next explanation does not depend on the explanations computed before.</p>
        <p>Hence, the question arises, how we can we find a more goal-directed algorithm. Suppose that
we have computed an explanation 1 using Algorithm 1. The next explanation 2 must be
diferent from 1, so 2 should miss at least one axiom from 1. Hence the next explanation
2 can be found by finding 1 ∈ 1 such that  |= (ℋ ∖ {1}) ⊑  and calling Algorithm 1
for the input , ℋ ∖ {1}, and . The next explanation 3, similarly, should miss something
from 1 and something from 2, so it can be found by finding some 1 ∈ 1 and 2 ∈ 2
such that  |= (ℋ ∖ {1, 2}) ⊑  and calling Algorithm 1 for the input , (ℋ ∖ {1, 2}),
and . In general, when explanations  (1 ≤  ≤ ) are computed, the next explanation
can be found by calling Algorithm 1 for the input , (ℋ ∖ { | 1 ≤  ≤ }), and  such
that  ∈  (1 ≤  ≤ ) and  |= (ℋ ∖ { | 1 ≤  ≤ }) ⊑ . Enumeration of subsets
ℋ ∖ { | 1 ≤  ≤ } can be organized using a data structure called a hitting set tree.
Definition 2. Let  = ⟨, ℋ, ⟩ be a concept abduction problem. A hitting set tree (short:
HS-tree) for  is a labeled tree  = (, , ) with  ̸= ∅ such that:
1. each non-leaf node  ∈  is labeled with an explanation () =  for  and, for each
 ∈ ,  has an outgoing edge ⟨, ⟩ ∈  with label (, ) = , and
2. each leaf node  ∈  is labeled by a special symbol () = ⊥.</p>
        <p>For each  ∈  , let () be the set of edge labels appearing on the path from  to the root node of
 . Then the following properties should additionally hold:
3. for each non-leaf node  ∈  , we have () ∩ () = ∅, and
4. for each leaf node  ∈  , we have  ̸|= (ℋ ∖ ()) ⊑ .
Example 2. Consider the concept abduction problem  = ⟨, ℋ, ⟩ with  = {′ ⊑ , ′ ⊓
 ⊑ ⊥,  ⊓  ⊑ }, ℋ = {, ′, , }, and  = , which has the (minimal) explanations
1 = {′, } (not satisfiable), 2 = {, } (not explanatory), and 3 = {′, } (satisfiable,
explanatory, and relevant). Figure 2 shows an HS-trees for  , where the explanation 2 = {, }
labels two diferent nodes of the tree.</p>
        <p>We next prove that every HS-tree must contain every explanation at least once.
Lemma 3. Let  = (, , ) be an HS-tree for the concept abduction problem  = ⟨, ℋ, ⟩.
Then, for each explanation  for  , there exists a node  ∈  such that () = .
Proof Sketch. Using Lemma 1 and Conditions 1, 3 and 4 of Definition 2, we can show that a
node  ∈  for which  () is maximal (w.r.t. set inclusion) and  () ∩  = ∅ is such that
() = .</p>
        <p>We next show that each HS-tree  = (, , ) for a concept abduction problem  =
⟨, ℋ, ⟩ has at most exponentially many nodes in the number of concepts in ℋ.
Lemma 4. Every HS-tree  for a concept abduction problem  = ⟨, ℋ, ⟩ has at most
∑︀0≤ ≤   nodes, where  is the number of concepts in ℋ.</p>
        <p>Proof Sketch. By analyzing Conditions 1 and 3 of Definition 2, we can show that both the depth
and the branching factor of  is bounded by ||ℋ||, which gives the desired bound.</p>
        <p>For constructing an HS-tree  = (, , ) for a concept abduction problem  = ⟨, ℋ, ⟩,
we can use Reiter’s Hitting Set Tree algorithm (or short: HST-algorithm) [11, 19]: We start by
creating the root node 0 ∈  . Then we repeatedly assign labels of nodes and edges as follows.
For each  ∈  , if () was not yet assigned, we calculate  (). If  ̸|= (ℋ ∖  ()) ⊑ , we
label () = ⊥ according to Condition 4 of Definition 2. Otherwise, we compute an explanation
 for  |= (ℋ ∖  ()) ⊑  using Algorithm 1 and set () = . Note that  satisfies
Condition 3 of Definition 2 since  ⊆ (ℋ ∖  ()). Next, for each  ∈ , we create a successor
node  of  and label (, ) = . This ensures that Condition 1 of Definition 2 is satisfied for
. Since, by Lemma 4,  has a bounded number of nodes, this process eventually terminates.</p>
        <p>Algorithm 2: Computing all explanations by the Hitting Set Tree algorithm
ComputeExplanationsHST(, ℋ, ): compute all explanations for ⟨, ℋ, ⟩
input : a knowledge base , a set of hypotheses ℋ, and an observation  such that
 |= ℋ ⊑ 
output : the set of all (minimal) subsets  ⊆ ℋ such that  |=  ⊑ 
1  ← ∅ ;
2  ← {∅} ;
3 while  ̸= ∅ do
4  ← choose  ∈ ;
5  ←  ∖ {};
6 if  |= ℋ ∖  ⊑  then
7  ← Minimize(, ℋ ∖ , );
8  ←  ∪ {};
9 for  ∈  do
10  ←  ∪ { ∪ {}};
11 return ;
Note that unlike the algorithm sketched in Lemma 2, the input for each call of Algorithm 1 now
depends on the results returned by the previous calls.</p>
        <p>
          The main idea of the HST-algorithm is to systematically compute two kinds of sets: (
          <xref ref-type="bibr" rid="ref1">1</xref>
          )
explanations  for the concept abduction problem ⟨, ℋ, ⟩ and (
          <xref ref-type="bibr" rid="ref2">2</xref>
          ) sets  that contain one element
from each explanation  on a branch. The name of the algorithm comes from the notion of a
hitting set, which characterizes the latter sets.
        </p>
        <p>Definition 3. Let  be a set of sets of some elements. A set  is a hitting set for  if  ∩  ̸= ∅
for each  ∈  . A hitting set  for  is minimal if every ′ ⊊  is not a hitting set for  .</p>
        <p>Intuitively, a hitting set for  is a set  that contains at least one element from every set
 ∈  . An HS-tree is then a tree  = (, , ) such that for each  ∈  , () is a hitting set
of the set of explanations on the path from  to the root of  . The leaf nodes  of  are labeled
by hitting sets () such  ̸|= (ℋ ∖ ()) ⊑ . Intuitively, the set () represents a set such
that the removal of () from ℋ breaks the subsumption.</p>
        <p>We get following bound on the number of subsumption tests performed by the HST-algorithm:
Lemma 5. An HS-tree for a concept abduction problem  = ⟨, ℋ, ⟩ can be constructed using
at most ∑︀1≤ ≤ +1  subsumption tests, where  is the number of concepts in ℋ.
Proof. We call Algorithm 1 exactly once per node. Combined with Lemma 4, we get the desired
bound on the number of subsumption tests performed by the HST-algorithm.</p>
        <p>The HST-algorithm can further be optimized in several ways. First, it is not necessary to
store the complete HS-tree in memory. For computing an explanation  at each node , it is
suficient to know just the set (). For each successor  of  associated with some  ∈ (),
the set () can be computed as () = () ∪ {}. Hence, it is possible to compute all
explanations by recursively processing and creating the sets () as shown in Algorithm 2. The
algorithm saves all explanations in a set , which is initially empty (Line 1). The explanations
are computed by processing the sets (); the sets that are not yet processed are stored in the
queue , which initially contains (0) = ∅ for the root node 0 (Line 2). The elements of 
are then repeatedly processed in a loop (Lines 3–10) until  becomes empty. First, we choose
any  ∈  (Line 4) and remove it from  (Line 5). Then, we test whether  |= (ℋ ∖ ) ⊑ 
(Line 6). If the subsumption holds, this means that the corresponding node  of the HS-tree
with () =  is not a leaf node. We then compute an explanation  using Algorithm 1 and
add it to  (Lines 7–8). Further, for each  ∈ , we create the set () = () ∪ {} for the
corresponding successor node  of  and add () to  for later processing (Lines 9–10). If
the subsumption  |= (ℋ ∖ ) ⊑  does not hold, we have reached a leaf of the HS-tree and
no further children of this node should be created.</p>
        <p>Lemma 6. Given , ℋ, and  as input, Algorithm 2 returns all explanations for the concept
abduction problem  = ⟨, ℋ, ⟩.</p>
        <p>Proof Sketch. We can prove the claim by showing that the following invariant always holds
in the main loop (Lines 3–10): If  is an explanation for , then either  ∈  or there exists
 ∈  such that  ⊆ ℋ ∖ .</p>
        <p>Note that, as before, checking whether a returned subsumption is explanatory, satisfiable,
and relevant requires at most two subsumption and a set membership check.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Computing Abductive Explanations using SAT Solvers</title>
        <p>
          The main idea of the HST-algorithm is to systematically compute two kinds of sets: (
          <xref ref-type="bibr" rid="ref1">1</xref>
          )
explanations  for a concept abduction problem ⟨, ℋ, ⟩ and (
          <xref ref-type="bibr" rid="ref2">2</xref>
          ) hitting sets  that contain one
element from each explanation  on a branch. Intuitively, for a leaf node , () represents a
set such that the removal of () from ℋ breaks the subsumption, which means that every
explanation on the path to  is also a minimal hitting set for (). This property is called the
hitting set duality and it takes a prominent place in the HST-algorithm. We can, however, also
use this property as the basis of a direct algorithm for computing abductive explanations.
        </p>
        <p>
          Suppose that we have already computed some set  of explanations and some set  of hitting
sets for the concept abduction problem  = ⟨, ℋ, ⟩. How can we find a new explanation?
As mentioned, each new explanation must be a hitting set for  , i.e., it should contain one
concept from every set in  . Furthermore, it should be diferent from any of the previously
computed explanations, i.e., it should miss one axiom from every  ∈ . Suppose we have
found a subset  ⊆ ℋ satisfying these two requirements:
∀ ∈  :  ∩  ̸= ∅,
(
          <xref ref-type="bibr" rid="ref5">5</xref>
          )
∀ ∈  :  ∖  ̸= ∅. (
          <xref ref-type="bibr" rid="ref6">6</xref>
          )
If  |=  ⊑ , then, using Algorithm 1, we can extract a minimal subset ′ ⊆  such
that  |= ′ ⊑ . Note that ′ still misses at least one axiom from each  ∈  since (
          <xref ref-type="bibr" rid="ref6">6</xref>
          ) is
preserved under removal of concepts from  . Therefore, ′ is a new explanation for . If
 ̸|=  ⊑ , then, similarly, by adding concepts  ∈ ℋ to  preserving  ̸|=  ⊑ , we
Algorithm 3: Maximizing non-subsumption
        </p>
        <p>
          Maximize(, ℋ, , ): compute a maximal subset  ⊆ ℋ such that  ⊆  and
 ̸|=  ⊑ 
input : a knowledge base , a set of hypotheses ℋ, a subset  ⊆ ℋ , and an
observation  such that  ̸|=  ⊑ 
output :  such that  ⊆  ⊆ ℋ such that  ̸|=  ⊑  but  |=  ′ ⊑  for every
 ′ with  ⊊  ′ ⊆ ℋ
1  ←  ;
2 for  ∈ ℋ ∖  do
3 if  ̸|= ( ∪ {}) ⊑  then
4  ←  ∪ {};
5 return  ;
can find a maximal superset  of  ( ⊆  ⊆ ℋ ) such that  ̸|=  ⊑ : see Algorithm 3.
Note that (
          <xref ref-type="bibr" rid="ref5">5</xref>
          ) is preserved under additions of elements to  . Thus, using any set  satisfying
(
          <xref ref-type="bibr" rid="ref5">5</xref>
          ) and (
          <xref ref-type="bibr" rid="ref6">6</xref>
          ) we can find either a new explanation or a new minimal hitting set.
        </p>
        <p>
          The question arises, how we can find a set  satisfying Conditions (
          <xref ref-type="bibr" rid="ref5">5</xref>
          ) and (
          <xref ref-type="bibr" rid="ref6">6</xref>
          ). These
conditions require solving a rather complex combinatorial problem, for which propositional
(SAT) solvers ofer a convenient and efective way of solving such problems. In the following,
we describe a propositional encoding of Conditions (
          <xref ref-type="bibr" rid="ref5">5</xref>
          ) and (
          <xref ref-type="bibr" rid="ref6">6</xref>
          ).
        </p>
        <p>
          To formulate the propositional encoding, we assign to each concept  ∈ ℋ a fresh
propositional variable . Then, every interpretation ℐ determines a set  =  (ℐ) = { ∈ ℋ | ℐ =
1} of concepts whose corresponding propositional variable is true. We construct a propositional
formula  such that  ℐ = 1 if and only if  (ℐ) satisfies (
          <xref ref-type="bibr" rid="ref5">5</xref>
          ) and (
          <xref ref-type="bibr" rid="ref6">6</xref>
          ) for the given sets  of
explanations and  of minimal hitting sets. Thus, to find a subset  satisfying (
          <xref ref-type="bibr" rid="ref5">5</xref>
          ) and (
          <xref ref-type="bibr" rid="ref6">6</xref>
          ), it is
suficient to find a model ℐ of  and compute  (ℐ). We define  as follows:
 =  (,  ) = ⋀︁ ⋁︁ ¬ ∧ ⋀︁ ⋁︁ .
        </p>
        <p>
          ∈ ∈
∈ ∈
(
          <xref ref-type="bibr" rid="ref7">7</xref>
          )
Example 3. Let  be the ontology from Example 1. We assign the propositional variables
, ,  , ′ to the concepts , , , ′, respectively. Let  = {{, ′}, {, ′}} be the set
of explanations 1 and 2 from Example 1 and  a set containing {, }, i.e., assume we have
constructed the left-most branch on the HST on the left-hand side of Figure 1. Then according to (
          <xref ref-type="bibr" rid="ref7">7</xref>
          )
we have:
        </p>
        <p>=  (,  ) = (¬ ∨ ¬′ ) ∧ (¬ ∨ ¬′ ) ∧ ( ∨  ).
 has a model ℐ with ℐ = 1 and ℐ = ℐ = ℐ′ = 0, which gives  (ℐ) = {}.</p>
        <p>
          Once the set  determined by a model ℐ of  is found, we can extract either a new explanation
 or a new minimal hitting set  from  by minimizing the subsumee using Algorithm 1
or maximizing non-subsumption using Algorithm 3. After that, we can update  according
to (
          <xref ref-type="bibr" rid="ref7">7</xref>
          ) and compute a new model of  , if there exist any. Once  is unsatisfiable,  contains all
explanations for  and  contains all minimal hitting sets.
        </p>
        <p>Algorithm 4: Computing all explanations using a SAT solver</p>
        <p>ComputeExplanationsSAT(, ℋ, ): compute all explanations for the concept
abduction problem  = ⟨, ℋ, ⟩
input : a knowledge base , a set of hypotheses ℋ, and an observation  such that
 |= ℋ ⊑ 
output : the set of all minimal subsets  ⊆ ℋ such that  |=  ⊑</p>
        <p>Maximize(ℋ, ℋ, , );
 ∧ ⋁︀{ |  ∈ ℋ ∖  };
 ←
 ←
13 return ;</p>
        <p>
          Algorithm 4 summarizes the described procedure for computing all explanations for a concept
abduction problem  = ⟨, ℋ, ⟩ using a SAT solver. We start by creating an empty set  of
explanations (Line 1) and a formula  that is always true (Line 2). Then, in a loop (Lines 3–12), as
long as  is satisfiable (which is checked using a SAT solver), we take any model ℐ of  (Line 4),
extract the corresponding set  =  (ℐ) that it defines (Line 5), and check the subsumption
 |=  ⊑ . If the entailment holds, using Algorithm 1 we compute an explanation for
 |=  ⊑  (Line 7), which, by Lemma 1, is an explanation for . This explanation is then
added to  (Line 8) and  is extended with a new conjunct for this explanation according to (
          <xref ref-type="bibr" rid="ref7">7</xref>
          )
(Line 9). If the entailment does not hold, we compute a maximal superset  of  such that
 ̸|=  ⊑  using Algorithm 3 (Line 11) and extend  with the corresponding conjunct for
the new minimal hitting set  = ℋ ∖  according to (
          <xref ref-type="bibr" rid="ref7">7</xref>
          ) (Line 12). As soon as  becomes
unsatisfiable, we return the set  of computed explanations (Line 13).
        </p>
        <p>Example 4. Consider the concept abduction problem  = ⟨, ℋ, ⟩ from Example 1 and
propositional encoding of concepts in ℋ from Example 3. The following table shows a run of
Algorithm 4 for the inputs , ℋ, and . Every row in this table corresponds to one iteration of the
while-loop (Lines 3–12). The first column gives the value of the interpretation ℐ for  computed
in this iteration. The second column shows the value of  computed for this interpretation and
whether the entailment  |=  ⊑  holds. The third column shows the result of minimizing the
subsumee or maximizing the non-subsumption using Algorithms 1 and 3. The last column shows
the conjunct that is added to  for the corresponding explanation or minimal hitting set.
 |= min( ) ⊑ /
 ̸|= max( ) ⊑ 
 ̸|= {, } ⊑ 
 ̸|= {, } ⊑ 
 |= {, } ⊑ 
 |= {, ′} ⊑ 
 |= {, ′} ⊑ 
 ̸|= {, ′} ⊑</p>
        <p>∨ ′
 ∨ ′
¬ ∨ ¬
¬ ∨ ¬′
¬ ∨ ¬′
 ∨</p>
        <p>We briefly discuss similarities and diferences between Algorithm 2 and Algorithm 4: Both
algorithms systematically explore subsets of ℋ and minimize the subsumee from such subset to
compute explanations. Algorithm 2 constructs such subsets (ℋ ∖ ) manually by removing one
axiom appearing in the previously computed explanation (if there is any) in all possible ways.
Algorithm 4 enumerates such subsets  with the help of a SAT solver. The main diference is
that Algorithm 2 may encounter the same subsets many times (on diferent branches), whereas
the propositional encoding in Algorithm 4 ensures that such subsets never repeat. Of course,
an iteration of Algorithm 2 cannot be directly compared to an iteration of Algorithm 4. Both
iterations use at most one call to Algorithm 1, but Algorithm 4 may also require a call to
Algorithm 3, as well as checking satisfiability of  . The latter requires solving an NP-complete
problem, for which no polynomial algorithm is known so far. In order to check satisfiability
of  , a SAT solver usually tries several (in worst-case exponentially many) propositional
interpretations until a model of  is found. As each such interpretation ℐ corresponds to a
subset  (ℐ) ⊆ ℋ , this process can be compared to the enumeration of subsets in Algorithm 2.
However, a SAT solver usually implements a number of sophisticated optimizations, which
make the search for models very eficient in practice, whereas the subset enumeration strategy
used by Algorithm 2 is rather simplistic. Hence Algorithm 4 is likely to win in speed. On
the other hand, Algorithm 4 requires saving all explanations (and minimal hitting sets) in the
propositional formula  , which might result in a formula of exponential size, if the number
of such explanations or hitting sets is exponential. In this regard, Algorithm 2 could be more
memory eficient since saving (all) explanations is optional (see the discussion at the end of
Section 3.2). Hence both algorithms have their own advantages and disadvantages.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Conclusions</title>
      <p>We have presented two alternative algorithms for computing (all or some) solutions to the
concept abduction problem: one algorithm is based on Reiter’s hitting set tree algorithm,
whereas the other one relies on a SAT encoding. It remains to be analyzed how these algorithms
behave in practice and how these algorithms difer on diferent real-world ontologies, where an
important aspect is also finding eficient incremental SAT solvers.</p>
      <p>In contrast to previous work, the algorithms do not rely on a refutation-based calculus and,
hence, can be used also with eficient reasoners for tractable DLs such as ℰ ℒ and its extensions.
Another direction for future work is extending the approach to other forms of (logic-based)
abduction, e.g., to ABox abduction and a comparison with other existing approaches, which are
mostly focusing on ABox abduction.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>C. S.</given-names>
            <surname>Peirce</surname>
          </string-name>
          , Deduction, Induction, and
          <string-name>
            <surname>Hypothesis</surname>
          </string-name>
          ,
          <source>Popular Science Monthly</source>
          <volume>13</volume>
          (
          <year>1878</year>
          )
          <fpage>470</fpage>
          -
          <lpage>482</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>T.</given-names>
            <surname>Eiter</surname>
          </string-name>
          ,
          <string-name>
            <surname>G. Gottlob,</surname>
          </string-name>
          <article-title>The complexity of logic-based abduction</article-title>
          ,
          <source>J. ACM</source>
          <volume>42</volume>
          (
          <year>1995</year>
          )
          <fpage>3</fpage>
          -
          <lpage>42</lpage>
          . URL: https://doi.org/10.1145/200836.200838.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>M.</given-names>
            <surname>Denecker</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. C.</given-names>
            <surname>Kakas</surname>
          </string-name>
          ,
          <article-title>Abduction in logic programming</article-title>
          , in: A.
          <string-name>
            <surname>C. Kakas</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Sadri</surname>
          </string-name>
          (Eds.), Computational Logic:
          <article-title>Logic Programming and Beyond, Essays in Honour of Robert A</article-title>
          .
          <string-name>
            <surname>Kowalski</surname>
          </string-name>
          ,
          <string-name>
            <surname>Part</surname>
            <given-names>I</given-names>
          </string-name>
          , volume
          <volume>2407</volume>
          of Lecture Notes in Computer Science, Springer,
          <year>2002</year>
          , pp.
          <fpage>402</fpage>
          -
          <lpage>436</lpage>
          . URL: https://doi.org/10.1007/3-540-45628-7_
          <fpage>16</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>C.</given-names>
            <surname>Elsenbroich</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Kutz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>U.</given-names>
            <surname>Sattler</surname>
          </string-name>
          ,
          <article-title>A case for abductive reasoning over ontologies</article-title>
          , in: B.
          <string-name>
            <surname>C. Grau</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Hitzler</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Shankey</surname>
          </string-name>
          , E. Wallace (Eds.),
          <source>Proceedings of the OWLED*06 Workshop on OWL: Experiences and Directions</source>
          , Athens, Georgia, USA, November
          <volume>10</volume>
          -
          <issue>11</issue>
          ,
          <year>2006</year>
          , volume
          <volume>216</volume>
          <source>of CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2006</year>
          . URL: http: //ceur-ws.
          <source>org/</source>
          Vol-
          <volume>216</volume>
          /submission_25.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>S.</given-names>
            <surname>Klarman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>U.</given-names>
            <surname>Endriss</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Schlobach</surname>
          </string-name>
          ,
          <article-title>Abox abduction in the description logic ALC</article-title>
          ,
          <source>J. Autom. Reason</source>
          .
          <volume>46</volume>
          (
          <year>2011</year>
          )
          <fpage>43</fpage>
          -
          <lpage>80</lpage>
          . URL: https://doi.org/10.1007/s10817-010-9168-z.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>K.</given-names>
            <surname>Halland</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Britz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Klarman</surname>
          </string-name>
          ,
          <article-title>Tbox abduction in ALC using a DL tableau</article-title>
          , in: M.
          <string-name>
            <surname>Bienvenu</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Ortiz</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Rosati</surname>
          </string-name>
          , M. Simkus (Eds.),
          <source>Informal Proceedings of the 27th International Workshop on Description Logics</source>
          , Vienna, Austria,
          <source>July 17-20</source>
          ,
          <year>2014</year>
          , volume
          <volume>1193</volume>
          <source>of CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2014</year>
          , pp.
          <fpage>556</fpage>
          -
          <lpage>566</lpage>
          . URL: http: //ceur-ws.
          <source>org/</source>
          Vol-
          <volume>1193</volume>
          /paper_42.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J.</given-names>
            <surname>Pukancová</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Homola</surname>
          </string-name>
          ,
          <article-title>Abox abduction for description logics: The case of multiple observations</article-title>
          , in: M.
          <string-name>
            <surname>Ortiz</surname>
          </string-name>
          , T. Schneider (Eds.),
          <source>Proceedings of the 31st International Workshop on Description Logics co-located with 16th International Conference on Principles of Knowledge Representation and Reasoning (KR</source>
          <year>2018</year>
          ), Tempe, Arizona,
          <string-name>
            <surname>US</surname>
          </string-name>
          , October 27th - to - 29th,
          <year>2018</year>
          , volume
          <volume>2211</volume>
          <source>of CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2018</year>
          . URL: http://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>2211</volume>
          /paper-31.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>P.</given-names>
            <surname>Koopmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Del-Pinto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Tourret</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. A.</given-names>
            <surname>Schmidt</surname>
          </string-name>
          ,
          <article-title>Signature-based abduction for expressive description logics</article-title>
          , in: D.
          <string-name>
            <surname>Calvanese</surname>
          </string-name>
          , E. Erdem, M. Thielscher (Eds.),
          <source>Proceedings of the 17th International Conference on Principles of Knowledge Representation and Reasoning</source>
          , KR 2020, Rhodes, Greece,
          <source>September 12-18</source>
          ,
          <year>2020</year>
          ,
          <year>2020</year>
          , pp.
          <fpage>592</fpage>
          -
          <lpage>602</lpage>
          . URL: https://doi.org/10.24963/kr.2020/59.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>W.</given-names>
            <surname>Del-Pinto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. A.</given-names>
            <surname>Schmidt</surname>
          </string-name>
          ,
          <article-title>Abox abduction via forgetting in ALC</article-title>
          , in: The
          <source>ThirtyThird AAAI Conference on Artificial Intelligence</source>
          ,
          <source>AAAI</source>
          <year>2019</year>
          ,
          <source>The Thirty-First Innovative Applications of Artificial Intelligence Conference</source>
          ,
          <string-name>
            <surname>IAAI</surname>
          </string-name>
          <year>2019</year>
          ,
          <source>The Ninth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI</source>
          <year>2019</year>
          , Honolulu, Hawaii, USA, January 27 - February 1,
          <year>2019</year>
          , AAAI Press,
          <year>2019</year>
          , pp.
          <fpage>2768</fpage>
          -
          <lpage>2775</lpage>
          . URL: https://doi.org/10. 1609/aaai.v33i01.
          <fpage>33012768</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>B.</given-names>
            <surname>Glimm</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Kazakov</surname>
          </string-name>
          ,
          <article-title>Classical algorithms for reasoning and explanation in description logics</article-title>
          , in: M.
          <string-name>
            <surname>Krötzsch</surname>
          </string-name>
          , D. Stepanova (Eds.),
          <source>Reasoning Web. Explainable Artificial Intelligence - 15th International Summer School</source>
          <year>2019</year>
          , Bolzano, Italy,
          <source>September 20-24</source>
          ,
          <year>2019</year>
          , Tutorial Lectures, volume
          <volume>11810</volume>
          of Lecture Notes in Computer Science, Springer,
          <year>2019</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>64</lpage>
          . URL: https://doi.org/10.1007/978-3-
          <fpage>030</fpage>
          -31423-
          <issue>1</issue>
          _
          <fpage>1</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>R.</given-names>
            <surname>Reiter</surname>
          </string-name>
          ,
          <article-title>A theory of diagnosis from first principles</article-title>
          ,
          <source>Artificial Intelligence</source>
          <volume>32</volume>
          (
          <year>1987</year>
          )
          <fpage>57</fpage>
          -
          <lpage>95</lpage>
          . URL: https://www.sciencedirect.com/science/article/pii/0004370287900622. doi:https: //doi.org/10.1016/
          <fpage>0004</fpage>
          -
          <lpage>3702</lpage>
          (
          <issue>87</issue>
          )
          <fpage>90062</fpage>
          -
          <lpage>2</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>M.</given-names>
            <surname>Bienvenu</surname>
          </string-name>
          ,
          <article-title>Complexity of abduction in the EL family of lightweight description logics</article-title>
          , in: G. Brewka, J. Lang (Eds.),
          <source>Principles of Knowledge Representation and Reasoning: Proceedings of the Eleventh International Conference, KR</source>
          <year>2008</year>
          , Sydney, Australia,
          <source>September 16-19</source>
          ,
          <year>2008</year>
          , AAAI Press,
          <year>2008</year>
          , pp.
          <fpage>220</fpage>
          -
          <lpage>230</lpage>
          . URL: http://www.aaai.org/Library/KR/2008/ kr08-
          <fpage>022</fpage>
          .php.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>M. W. Birte</surname>
            <given-names>Glimm</given-names>
          </string-name>
          , Yevgeny Kazakov,
          <source>Concept Abduction for Description Logics - Technical Report, Technical Report</source>
          , Ulm University,
          <source>Institute of Artificial Intelligence</source>
          ,
          <year>2022</year>
          . https://www.uni-ulm.de/fileadmin/website_uni_ulm/iui.inst.090/ Publikationen/2022/GlKW22a_report.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>M.</given-names>
            <surname>Krötzsch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Simančík</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Horrocks</surname>
          </string-name>
          ,
          <article-title>A description logic primer</article-title>
          ,
          <source>CoRR abs/1201</source>
          .4089 (
          <year>2012</year>
          ). Available at http://arxiv.org/pdf/1201.4089.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>T.</given-names>
            <surname>Eiter</surname>
          </string-name>
          ,
          <string-name>
            <surname>G. Gottlob,</surname>
          </string-name>
          <article-title>The complexity of logic-based abduction</article-title>
          ,
          <source>J. ACM</source>
          <volume>42</volume>
          (
          <year>1995</year>
          )
          <fpage>3</fpage>
          -
          <lpage>42</lpage>
          . URL: https://doi.org/10.1145/200836.200838. doi:
          <volume>10</volume>
          .1145/200836.200838.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>F.</given-names>
            <surname>Baader</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Brandt</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>Lutz, Pushing the ℰ ℒ envelope further</article-title>
          ,
          <source>in: Proc. 5th Workshop on OWL: Experiences and Directions (OWLED'08)</source>
          , volume
          <volume>496</volume>
          ,
          <string-name>
            <surname>CEUR</surname>
          </string-name>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>K.</given-names>
            <surname>Schild</surname>
          </string-name>
          ,
          <article-title>A correspondence theory for terminological logics: Preliminary report</article-title>
          , in: J.
          <string-name>
            <surname>Mylopoulos</surname>
          </string-name>
          , R. Reiter (Eds.),
          <source>Proc. 12th Int. Joint Conf. on Artificial Intelligence (IJCAI'91)</source>
          , Morgan Kaufmann,
          <year>1991</year>
          , pp.
          <fpage>466</fpage>
          -
          <lpage>471</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>F. M.</given-names>
            <surname>Donini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Massacci</surname>
          </string-name>
          , EXPTIME tableaux for ℒ,
          <source>J. of Artificial Intelligence</source>
          <volume>124</volume>
          (
          <year>2000</year>
          )
          <fpage>87</fpage>
          -
          <lpage>138</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>R.</given-names>
            <surname>Greiner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. A.</given-names>
            <surname>Smith</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. W.</given-names>
            <surname>Wilkerson</surname>
          </string-name>
          ,
          <article-title>Readings in model-based diagnosis</article-title>
          , Morgan Kaufmann Publishers Inc.,
          <year>1992</year>
          , pp.
          <fpage>49</fpage>
          -
          <lpage>53</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>