=Paper=
{{Paper
|id=Vol-3263/paper-11
|storemode=property
|title=Concept Abduction for Description Logics
|pdfUrl=https://ceur-ws.org/Vol-3263/paper-11.pdf
|volume=Vol-3263
|authors=Birte Glimm,Yevgeny Kazakov,Michael Welt
|dblpUrl=https://dblp.org/rec/conf/dlog/GlimmKW22
}}
==Concept Abduction for Description Logics==
Concept Abduction for Description Logics Birte Glimm1 , Yevgeny Kazakov1 and Michael Welt1 1 Ulm University, Germany Abstract We present two alternative algorithms for computing (all or some) solutions to the concept abduction problem: one algorithm is based on Reiterβs hitting set tree algorithm, whereas the other one relies on a SAT encoding. In contrast to previous work, the algorithms do not rely on a refutation-based calculus and, hence, can be used also with efficient reasoners for tractable DLs such as β°β and its extensions. An adaptation to other forms of (logic-based) abduction, e.g., to ABox abduction, is also possible. Keywords Description Logics, abduction, ontologies, reasoning 1. Introduction The reasoning service of abduction, as originally introduced by Peirce [1], is a form of rea- soning that is used to generate possible explanations for a given observation. In logic-based abduction, the objective is to find a set of hypotheses (the explanation) which are consistent with a background theory and which logically entail the observation when taken together with the background knowledge. Often some minimality criterion, e.g., set inclusion or cardinality, is used to select preferred explanations. Abduction has numerous applications in artificial intelligence, e.g., in diagnosis, planning, or natural language understanding and abductive reasoning has intensively been studied in the context of propositional logic (see, e.g., [2]) and also in the context of answer set programming (see, e.g., [3]). For logic-based abduction over ontologies, approaches started to emerge more recently (see, e.g., [4, 5, 6, 7]). With an exception of recent approaches based on forgetting [8, 9], practical approaches for logic-based abduction over ontologies are based on refutation-based calculi such as tableau. Such approaches are based on the observation that the knowledge base π¦ does not yet entail the observation π, i.e., π¦ βͺ {Β¬π} is satisfiable, whereas an explanation πΈ has to be such that π¦ βͺ πΈ βͺ {Β¬π} is unsatisfiable since it is desired that π¦ βͺ πΈ |= π. Hence, open branches in a tableau for π¦ βͺ {Β¬π} can βguideβ the process of finding explanations as explanations have to be such that they lead to closing the open branches. While refutation-based calculi are dominantly used for reasoning in expressive Description Logics (DLs), tractable reasoning procedures for DLs such as β°β are not based on refutation, DL 2022: 35th International Workshop on Description Logics, August 7β10, 2022, Haifa, Israel $ birte.glimm@uni-ulm.de (B. Glimm); yevgeny.kazakov@uni-ulm.de (Y. Kazakov); michael.welt@uni-ulm.de (M. Welt) Β https://www.uni-ulm.de/in/ki/glimm/ (B. Glimm); https://www.uni-ulm.de/in/ki/kazakov/ (Y. Kazakov) 0000-0002-6331-4176 (B. Glimm); 0000-0002-3011-8646 (Y. Kazakov) Β© 2022 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0). CEUR Workshop Proceedings http://ceur-ws.org ISSN 1613-0073 CEUR Workshop Proceedings (CEUR-WS.org) but directly derive (logical) consequences, which prevents the use of existing approaches to abduction. We aim at closing this gap and, inspired by the work of Kazakov and Glimm [10] for computing justification (i.e., minimal subsets of axioms from a knowledge base that are needed for a given entailment), we propose an alternative approach to efficiently compute some or all (minimal) explanations. The approach is based on Reiterβs minimal hitting set algorithm [11] without relying on a refutation-based calculus to guide the algorithm. An encoding of the problem into SAT allows for better managing the combinatorial problem of selecting what constitutes an explanation. As our goal is the development of an approach for DLs such as β°β, where the main reasoning task involves determining concept subsumptions, we focus on concept abduction [12] as introduced next. The accompanying technical report gives further details, examples, and complete proofs [13]. 2. Preliminaries We use the standard syntax and semantics of DLs (see, e.g., [14]). Note that the DL β°β only allows for using the top concept, conjunctions, and existential restrictions as concepts. We now introduce the basic ideas of (concept) abduction (see also [12, 15]). Definition 1. A concept abduction problem π« is a tuple β¨π¦, β, πͺβ© with π¦ a knowledge base, β, the hypotheses, a set of atomic concepts, and π, the observation, a single atomic concept.1 A set πΈ = {π΄1 , . . . , π΄π } β β is an explanation for π« if π¦ |= π΄1 β . . . β π΄π β π. Such an πΈ is explanatory if π΄1 β . . . β π΄π β π ΜΈβ π¦, πΈ is satisfiable if π΄1 β . . . β π΄π is satisfiable w.r.t. π¦, πΈ is relevant if β ΜΈ|= π΄1 β . . . β π΄π β π, and πΈ (syntactically) minimal if there is no other explanation πΈ β² for π« such that πΈ β² β πΈ. Note that the given set of hypothesis allows for restricting the set of concepts that can be used in explanations, but may also contain all concepts used in the knowledge base. In the remainder, we are interested in finding explanations that are explanatory, satisfiable, relevant, and (syntactically) minimal. Abusing notation, for a set of concepts πΈ = {π΄1 , . . . , π΄π }, we also write π¦ |= πΈ β π instead of π¦ |= π΄1 β . . . β π΄π β π. Example 1. Consider the knowledge base π¦ containing the axioms π΄ β βπ.π΅ (1) πΆ β πΆβ² β π· (3) βπ.π΅ β πΆ (2) π΄βπ΅ ββ₯ (4) and the abduction problem π« = β¨π¦, β, πͺβ© with β = {π΄, π΅, πΆ, πΆ β² } and π = π·. From Axiom (1) and Axiom (2), we have π¦ |= π΄ β πΆ. Hence, together with Axiom (3), we find that π¦ |= π΄ β πΆ β² β π· = π. Among others, we have the following explanations for π«: πΈ1 = {π΄, πΆ β² } πΈ2 = {πΆ, πΆ β² } πΈ3 = {π΄, π΅} πΈ4 = {π΄, πΆ, πΆ β² } 1 Note that requiring the observation to be an atomic concept is without loss of generality since for an observation in form of a complex concept πΆ, we can simply introduce an axiom π β‘ πΆ and use π as observation. We can proceed analogously for hypotheses. Algorithm 1: Finding one explanation Minimize(π¦, β, π): compute a minimal explanation for the concept abduction problem β¨π¦, β, πͺβ© input : a knowledge base π¦, a set of hypotheses β, and an observation π such that π¦ |= β β π output : a minimal explanation πΈ β β such that π¦ |= πΈ β π (cf. Definition 1) 1 πΈ β β; 2 for π΄ β β do 3 if π¦ |= (πΈ β {π΄}) β π then 4 πΈ β πΈ β {π΄}; 5 return πΈ; Among these only πΈ1 is explanatory, relevant, satisfiable, and minimal. We have that πΈ2 is not explanatory since πΆ β πΆ β² β π· β π¦, and πΈ3 is not satisfiable (due to Axiom (4)). Finally, πΈ4 is not minimal since πΈ1 β πΈ4 . Note that {π·} is not an explanation since π· β / β, but even if it were, the explanation would not be relevant since β |= π· β π·. 3. Computing Abductive Explanations Before proposing our approach, we point out that the number of (minimal) explanations for a given observation may be exponential in the size of the knowledge base. We start by showing how to compute one explanation before generalizing the approach to compute all explanations, first, using hitting set trees and, then, via a SAT encoding. 3.1. Computing One Abductive Explanation Given a concept abduction problem π« = β¨π¦, β, πͺβ©, a naive algorithm for computing one explanation πΈ such that π¦ |= πΈ β π is relatively easy. If π¦ ΜΈ|= β β π, there is no satisfiable explanation and we can stop. Otherwise, we start from πΈ = β and repeatedly remove concepts from πΈ as long as this does not break the entailment π¦ |= πΈ β π. At a certain point, no concept can be removed without breaking the entailment, which implies that πΈ is a minimal explanation w.r.t. π«. Algorithm 1 summarizes this idea. Note that Algorithm 1 makes calls to a reasoning service for concept subsumption checking without relying on a particular kind of procedure (e.g., tableau-based as well as consequence-based procedures may be used). It remains to check whether the returned πΈ is explanatory, satisfiable, and relevant, which requires at most two further subsumption checks and a check for set containment. The correctness of Algorithm 1 relies on the fact that a conjunction with additional conjuncts is more specific: Lemma 1. Let π¦ be a DL knowledge base, πΈ and πΈ β² two sets of atomic concepts such that πΈ β² β πΈ, and π an atomic concept. Then π¦ |= πΈ β² β π implies π¦ |= πΈ β π. Proof Sketch. Intuitively, the interpretation of a conjunction is more restricted the more con- juncts are contained in the conjunction. Note that Lemma 1 only relies on the semantics of conjunctions and subsumption, which is shared among all DLs. We show that Algorithm 1 is correct under the assumed semantics. Theorem 1. Let πΈ be the output of Algorithm 1 for the input π¦, β, and π such that π¦ |= πΈ β π. Then πΈ is a minimal explanation for the abduction problem β¨π¦, β, πͺβ©. Proof Sketch. Since π¦ |= β β π by assumption, πΈ is initialized such that π¦ |= πΈ β π (in Line 1). Since we only remove a concept from πΈ, if the subsumption still holds in Line 4, π¦ |= πΈ β π is preserved. We can further show that πΈ is minimal explanation as unneeded conjuncts are step by step removed. Finally, observe that a run of Algorithm 1 requires exactly π subsumption tests, where π = ||β||, which is bounded by the number of concepts in π¦ (two further subsumption tests and a set containment test are needed for checking whether πΈ is relevant, satisfiable, and explanatory). Hence, the complexity of computing one minimal explanation is bounded by a linear function over the complexity of subsumption checking. In particular, for tractable languages such as β°β and its tractable extensions [16], one (minimal) explanation for the concept abduction problem can be computed in polynomial time, which is worst-case optimal [12].2 For DLs where subsumption checking is ExpTime-complete such as πβπ [17, 18], the algorithm requires worst-case exponential time. 3.2. Computing All Abductive Explanations An explanation for a concept abduction problem β¨π¦, β, πͺβ© consists of concepts that are sub- sumed by the observation. As we have seen in Example 1, there can be several different explanations. To find all explanations, it is, therefore, necessary to change a concept in every explanation of β¨π¦, β, πͺβ© and the question of how to compute all (minimal) explanations arises. Note first, that the output of Algorithm 1 depends on the order in which the axioms in π¦ are enumerated in the for-loop (Line 2). Different orders of the concepts can result in different removals and, consequently, different explanations. Lemma 2. For each (minimal) explanation πΈ of a concept abduction problem β¨π¦, β, πͺβ©, there exists some order of concepts in β for which Algorithm 1 with the input π¦, β, and π returns πΈ. Proof Sketch. Assume, to the contrary of what is to be shown, that a minimal explanation πΈπ is not returned. We can show that processing first concepts in β β πΈπ and then concepts from πΈπ leads to a contradiction as desired. 2 Note that although subsumption checking for β°β++ is tractable, the complexity of checking whether an explanation exists is found to be NP-complete by Bienvenu [12]. This apparent contradiction is because unsatisfiable explanations (which can only occur in logics that can express unsatisfiable concepts/β₯, as β°β++) are excluded in the definition of Bienvenu as explanations. Our algorithm, however, might terminate with an unsatisfiable explanation. If that has to be avoided, some form of backtracking for such cases would be needed, which indeed then leads to a higher complexity of the problem. πΈ1 = {π΄, πΆ β² } πΈ3 = {π΄, π΅} π΄ πΆβ² π΄ π΅ πΈ2 = {πΆ, πΆ β² } πΈ3 = {π΄, π΅} πΈ2 = {πΆ, πΆ β² } πΈ1 = {π΄, πΆ β² } πΆ πΆβ² π΄ π΅ πΆ πΆβ² π΄ πΆβ² β₯ β₯ β₯ β₯ β₯ β₯ β₯ β₯ Figure 1: Two HS-trees for the concept abduction problem from Example 1 The property stated in Lemma 2 means that for computing all explanations for π« = β¨π¦, β, πͺβ©, it is sufficient to run Algorithm 1 for all possible orders of concepts in β. As the number of explanations can be exponential in π, the exponential behavior of an algorithm for computing all explanations cannot be avoided in general. Unfortunately, the described algorithm is not very practical since it performs exponentially many subsumption tests for all inputs, even if, e.g., π« has just one explanation. This is because this algorithm is not goal-directed: the computation of each next explanation does not depend on the explanations computed before. Hence, the question arises, how we can we find a more goal-directed algorithm. Suppose that we have computed an explanation πΈ1 using Algorithm 1. The next explanation πΈ2 must be different from πΈ1 , so πΈ2 should miss at least one axiom from πΈ1 . Hence the next explanation πΈ2 can be found by finding π΄1 β πΈ1 such that π¦ |= (β β {π΄1 }) β π and calling Algorithm 1 for the input π¦, β β {π΄1 }, and π. The next explanation πΈ3 , similarly, should miss something from πΈ1 and something from πΈ2 , so it can be found by finding some π΄1 β πΈ1 and π΄2 β πΈ2 such that π¦ |= (β β {π΄1 , π΄2 }) β π and calling Algorithm 1 for the input π¦, (β β {π΄1 , π΄2 }), and π. In general, when explanations πΈπ (1 β€ π β€ π) are computed, the next explanation can be found by calling Algorithm 1 for the input π¦, (β β {π΄π | 1 β€ π β€ π}), and π such that π΄π β πΈπ (1 β€ π β€ π) and π¦ |= (β β {π΄π | 1 β€ π β€ π}) β π. Enumeration of subsets β β {π΄π | 1 β€ π β€ π} can be organized using a data structure called a hitting set tree. Definition 2. Let π« = β¨π¦, β, πͺβ© be a concept abduction problem. A hitting set tree (short: HS-tree) for π« is a labeled tree π = (π, πΈ, πΏ) with π ΜΈ= β such that: 1. each non-leaf node π£ β π is labeled with an explanation πΏ(π£) = πΈ for π« and, for each π΄ β πΈ, π£ has an outgoing edge β¨π£, π€β© β πΈ with label πΏ(π£, π€) = π΄, and 2. each leaf node π£ β π is labeled by a special symbol πΏ(π£) = β₯. For each π£ β π , let π»(π£) be the set of edge labels appearing on the path from π£ to the root node of π . Then the following properties should additionally hold: 3. for each non-leaf node π£ β π , we have πΏ(π£) β© π»(π£) = β , and 4. for each leaf node π£ β π , we have π¦ ΜΈ|= (β β π»(π£)) β π. Figure 1 shows an example of two different HS-trees for the concept abduction problem from Example 1. The following example shows that an explanation might occur more than once in an HS-Tree: πΈ1 = {π΄β² , πΆ} π΄β² πΆ πΈ2 = {π΄, π΅} πΈ2 = {π΄, π΅} π΄ π΅ π΄ π΅ β₯ β₯ πΈ3 = {π΄β² , π΅} β₯ π΄β² π΅ β₯ β₯ Figure 2: An HS-tree for the concept abduction problem from Example 2 Example 2. Consider the concept abduction problem π« = β¨π¦, β, πͺβ© with π¦ = {π΄β² β π΄, π΄β² β πΆ β β₯, π΄ β π΅ β π·}, β = {π΄, π΄β² , π΅, πΆ}, and π = π·, which has the (minimal) explanations πΈ1 = {π΄β² , πΆ} (not satisfiable), πΈ2 = {π΄, π΅} (not explanatory), and πΈ3 = {π΄β² , π΅} (satisfiable, explanatory, and relevant). Figure 2 shows an HS-trees for π«, where the explanation πΈ2 = {π΄, π΅} labels two different nodes of the tree. We next prove that every HS-tree must contain every explanation at least once. Lemma 3. Let π = (π, πΈ, πΏ) be an HS-tree for the concept abduction problem π« = β¨π¦, β, πͺβ©. Then, for each explanation πΈ for π«, there exists a node π£ β π such that πΏ(π£) = πΈ. Proof Sketch. Using Lemma 1 and Conditions 1, 3 and 4 of Definition 2, we can show that a node π£ β π for which π»(π£) is maximal (w.r.t. set inclusion) and π»(π£) β© πΈ = β is such that πΏ(π£) = πΈ. We next show that each HS-tree π = (π, πΈ, πΏ) for a concept abduction problem π« = β¨π¦, β, πͺβ© has at most exponentially many nodes in the number of concepts in β. Lemma 4. Every HS-tree π for a concept abduction problem π« = β¨π¦, β, πͺβ© has at most π nodes, where π is the number of concepts in β. βοΈ 0β€πβ€π π Proof Sketch. By analyzing Conditions 1 and 3 of Definition 2, we can show that both the depth and the branching factor of π is bounded by ||β||, which gives the desired bound. For constructing an HS-tree π = (π, πΈ, πΏ) for a concept abduction problem π« = β¨π¦, β, πͺβ©, we can use Reiterβs Hitting Set Tree algorithm (or short: HST-algorithm) [11, 19]: We start by creating the root node π£0 β π . Then we repeatedly assign labels of nodes and edges as follows. For each π£ β π , if πΏ(π£) was not yet assigned, we calculate π»(π£). If π¦ ΜΈ|= (β β π»(π£)) β π, we label πΏ(π£) = β₯ according to Condition 4 of Definition 2. Otherwise, we compute an explanation πΈ for π¦ |= (β β π»(π£)) β π using Algorithm 1 and set πΏ(π£) = πΈ. Note that πΈ satisfies Condition 3 of Definition 2 since πΈ β (β β π»(π£)). Next, for each π΄ β πΈ, we create a successor node π€ of π£ and label πΏ(π£, π€) = π΄. This ensures that Condition 1 of Definition 2 is satisfied for π£. Since, by Lemma 4, π» has a bounded number of nodes, this process eventually terminates. Algorithm 2: Computing all explanations by the Hitting Set Tree algorithm ComputeExplanationsHST(π¦, β, π): compute all explanations for β¨π¦, β, πͺβ© input : a knowledge base π¦, a set of hypotheses β, and an observation π such that π¦ |= β β π output : the set of all (minimal) subsets πΈ β β such that π¦ |= πΈ β π 1 π β β ; 2 π β {β }; 3 while π ΜΈ= β do 4 π» β choose π» β π; 5 π β π β {π»}; 6 if π¦ |= β β π» β π then 7 πΈ β Minimize(π¦, β β π», π); 8 π β π βͺ {πΈ}; 9 for π΄ β πΈ do 10 π β π βͺ {π» βͺ {π΄}}; 11 return π; Note that unlike the algorithm sketched in Lemma 2, the input for each call of Algorithm 1 now depends on the results returned by the previous calls. The main idea of the HST-algorithm is to systematically compute two kinds of sets: (1) expla- nations πΈ for the concept abduction problem β¨π¦, β, πͺβ© and (2) sets π» that contain one element from each explanation πΈ on a branch. The name of the algorithm comes from the notion of a hitting set, which characterizes the latter sets. Definition 3. Let π be a set of sets of some elements. A set π» is a hitting set for π if π» β© π ΜΈ= β for each π β π . A hitting set π» for π is minimal if every π» β² β π» is not a hitting set for π . Intuitively, a hitting set for π is a set π» that contains at least one element from every set π β π . An HS-tree is then a tree π = (π, πΈ, πΏ) such that for each π£ β π , π»(π£) is a hitting set of the set of explanations on the path from π£ to the root of π . The leaf nodes π£ of π are labeled by hitting sets π»(π£) such π¦ ΜΈ|= (β β π»(π£)) β π. Intuitively, the set π»(π£) represents a set such that the removal of π»(π£) from β breaks the subsumption. We get following bound on the number of subsumption tests performed by the HST-algorithm: LemmaβοΈ 5. An HS-tree for a concept abduction problem π« = β¨π¦, β, πͺβ© can be constructed using at most 1β€πβ€π+1 ππ subsumption tests, where π is the number of concepts in β. Proof. We call Algorithm 1 exactly once per node. Combined with Lemma 4, we get the desired bound on the number of subsumption tests performed by the HST-algorithm. The HST-algorithm can further be optimized in several ways. First, it is not necessary to store the complete HS-tree in memory. For computing an explanation πΈ at each node π£, it is sufficient to know just the set π»(π£). For each successor π€ of π£ associated with some π΄ β π»(π£), the set π»(π€) can be computed as π»(π€) = π»(π£) βͺ {π΄}. Hence, it is possible to compute all explanations by recursively processing and creating the sets π»(π£) as shown in Algorithm 2. The algorithm saves all explanations in a set π, which is initially empty (Line 1). The explanations are computed by processing the sets π»(π£); the sets that are not yet processed are stored in the queue π, which initially contains π»(π£0 ) = β for the root node π£0 (Line 2). The elements of π are then repeatedly processed in a loop (Lines 3β10) until π becomes empty. First, we choose any π» β π (Line 4) and remove it from π (Line 5). Then, we test whether π¦ |= (β β π») β π (Line 6). If the subsumption holds, this means that the corresponding node π£ of the HS-tree with π»(π£) = π» is not a leaf node. We then compute an explanation πΈ using Algorithm 1 and add it to π (Lines 7β8). Further, for each π΄ β πΈ, we create the set π»(π€) = π»(π£) βͺ {π΄} for the corresponding successor node π€ of π£ and add π»(π€) to π for later processing (Lines 9β10). If the subsumption π¦ |= (β β π») β π does not hold, we have reached a leaf of the HS-tree and no further children of this node should be created. Lemma 6. Given π¦, β, and π as input, Algorithm 2 returns all explanations for the concept abduction problem π« = β¨π¦, β, πͺβ©. Proof Sketch. We can prove the claim by showing that the following invariant always holds in the main loop (Lines 3β10): If πΈ is an explanation for π«, then either πΈ β π or there exists π» β π such that πΈ β β β π». Note that, as before, checking whether a returned subsumption is explanatory, satisfiable, and relevant requires at most two subsumption and a set membership check. 3.3. Computing Abductive Explanations using SAT Solvers The main idea of the HST-algorithm is to systematically compute two kinds of sets: (1) expla- nations πΈ for a concept abduction problem β¨π¦, β, πͺβ© and (2) hitting sets π» that contain one element from each explanation πΈ on a branch. Intuitively, for a leaf node π£, π»(π£) represents a set such that the removal of π»(π£) from β breaks the subsumption, which means that every explanation on the path to π£ is also a minimal hitting set for π»(π£). This property is called the hitting set duality and it takes a prominent place in the HST-algorithm. We can, however, also use this property as the basis of a direct algorithm for computing abductive explanations. Suppose that we have already computed some set π of explanations and some set π of hitting sets for the concept abduction problem π« = β¨π¦, β, πͺβ©. How can we find a new explanation? As mentioned, each new explanation must be a hitting set for π , i.e., it should contain one concept from every set in π . Furthermore, it should be different from any of the previously computed explanations, i.e., it should miss one axiom from every πΈ β π. Suppose we have found a subset π β β satisfying these two requirements: βπ β π : π β© π ΜΈ= β , (5) βπΈ β π : πΈ β π ΜΈ= β . (6) If π¦ |= π β π, then, using Algorithm 1, we can extract a minimal subset πΈ β² β π such that π¦ |= πΈ β² β π. Note that πΈ β² still misses at least one axiom from each πΈ β π since (6) is preserved under removal of concepts from π . Therefore, πΈ β² is a new explanation for π«. If π¦ ΜΈ|= π β π, then, similarly, by adding concepts π΄ β β to π preserving π¦ ΜΈ|= π β π, we Algorithm 3: Maximizing non-subsumption Maximize(π¦, β, π, π): compute a maximal subset π β β such that π β π and π¦ ΜΈ|= π β π input : a knowledge base π¦, a set of hypotheses β, a subset π β β, and an observation π such that π¦ ΜΈ|= π β π output : π such that π β π β β such that π¦ ΜΈ|= π β π but π¦ |= π β² β π for every π β² with π β π β² β β 1 π β π; 2 for π΄ β β β π do 3 if π¦ ΜΈ|= (π βͺ {π΄}) β π then 4 π β π βͺ {π΄}; 5 return π ; can find a maximal superset π of π (π β π β β) such that π¦ ΜΈ|= π β π: see Algorithm 3. Note that (5) is preserved under additions of elements to π . Thus, using any set π satisfying (5) and (6) we can find either a new explanation or a new minimal hitting set. The question arises, how we can find a set π satisfying Conditions (5) and (6). These conditions require solving a rather complex combinatorial problem, for which propositional (SAT) solvers offer a convenient and effective way of solving such problems. In the following, we describe a propositional encoding of Conditions (5) and (6). To formulate the propositional encoding, we assign to each concept π΄ β β a fresh proposi- tional variable ππ΄ . Then, every interpretation β determines a set π = π (β) = {π΄ β β | πβπ΄ = 1} of concepts whose corresponding propositional variable is true. We construct a propositional formula πΉ such that πΉ β = 1 if and only if π (β) satisfies (5) and (6) for the given sets π of explanations and π of minimal hitting sets. Thus, to find a subset π satisfying (5) and (6), it is sufficient to find a model β of πΉ and compute π (β). We define πΉ as follows: βοΈ βοΈ βοΈ βοΈ πΉ = πΉ (π, π ) = Β¬ππ΄ β§ ππ΄ . (7) πΈβπ π΄βπΈ π βπ π΄βπ Example 3. Let π¦ be the ontology from Example 1. We assign the propositional variables ππ΄ , ππ΅ , ππΆ , ππΆ β² to the concepts π΄, π΅, πΆ, πΆ β² , respectively. Let π = {{π΄, πΆ β² }, {πΆ, πΆ β² }} be the set of explanations πΈ1 and πΈ2 from Example 1 and π a set containing {π΄, πΆ}, i.e., assume we have constructed the left-most branch on the HST on the left-hand side of Figure 1. Then according to (7) we have: πΉ = πΉ (π, π ) = (Β¬ππ΄ β¨ Β¬ππΆ β² ) β§ (Β¬ππΆ β¨ Β¬ππΆ β² ) β§ (ππ΄ β¨ ππΆ ). πΉ has a model β with πβπ΄ = 1 and πβπ΅ = πβπΆ = πβπΆ β² = 0, which gives π (β) = {π΄}. Once the set π determined by a model β of πΉ is found, we can extract either a new explanation πΈ or a new minimal hitting set π from π by minimizing the subsumee using Algorithm 1 or maximizing non-subsumption using Algorithm 3. After that, we can update πΉ according to (7) and compute a new model of πΉ , if there exist any. Once πΉ is unsatisfiable, π contains all explanations for π« and π contains all minimal hitting sets. Algorithm 4: Computing all explanations using a SAT solver ComputeExplanationsSAT(π¦, β, π): compute all explanations for the concept abduction problem π« = β¨π¦, β, πͺβ© input : a knowledge base π¦, a set of hypotheses β, and an observation π such that π¦ |= β β π output : the set of all minimal subsets πΈ β β such that π¦ |= πΈ β π 1 π β β ; 2 πΉ β β€; 3 while β β : πΉ β = 1 do 4 β β choose β : πΉ β = 1; 5 π β {π΄ | πβπ΄ = 1}; 6 if π¦ |= π β π then 7 πΈ β Minimize(π¦, π, π); 8 π β π βͺ {πΈ}; βοΈ 9 πΉ β πΉ β§ {Β¬ππ΄ | π΄ β πΈ}; 10 else 11 π β Maximize(β, βοΈ β, π, π); 12 πΉ β πΉ β§ {ππ΄ | π΄ β β β π }; 13 return π; Algorithm 4 summarizes the described procedure for computing all explanations for a concept abduction problem π« = β¨π¦, β, πͺβ© using a SAT solver. We start by creating an empty set π of explanations (Line 1) and a formula πΉ that is always true (Line 2). Then, in a loop (Lines 3β12), as long as πΉ is satisfiable (which is checked using a SAT solver), we take any model β of πΉ (Line 4), extract the corresponding set π = π (β) that it defines (Line 5), and check the subsumption π¦ |= π β π. If the entailment holds, using Algorithm 1 we compute an explanation for π¦ |= π β π (Line 7), which, by Lemma 1, is an explanation for π«. This explanation is then added to π (Line 8) and πΉ is extended with a new conjunct for this explanation according to (7) (Line 9). If the entailment does not hold, we compute a maximal superset π of π such that π¦ ΜΈ|= π β π using Algorithm 3 (Line 11) and extend πΉ with the corresponding conjunct for the new minimal hitting set π = β β π according to (7) (Line 12). As soon as πΉ becomes unsatisfiable, we return the set π of computed explanations (Line 13). Example 4. Consider the concept abduction problem π« = β¨π¦, β, πͺβ© from Example 1 and propositional encoding of concepts in β from Example 3. The following table shows a run of Algorithm 4 for the inputs π¦, β, and π. Every row in this table corresponds to one iteration of the while-loop (Lines 3β12). The first column gives the value of the interpretation β for πΉ computed in this iteration. The second column shows the value of π computed for this interpretation and whether the entailment π¦ |= π β π holds. The third column shows the result of minimizing the subsumee or maximizing the non-subsumption using Algorithms 1 and 3. The last column shows the conjunct that is added to πΉ for the corresponding explanation or minimal hitting set. π¦ |= min(π ) β π/ πβπ΄ πβπ΅ πβπΆ πβπΆ β² π¦ |=? π β π π¦ ΜΈ|= max(π ) β π πΆ 0 0 0 0 π¦ ΜΈ|= β€ β π π¦ ΜΈ|= {π΄, πΆ} β π ππ΅ β¨ ππΆ β² 0 1 0 0 π¦ ΜΈ|= {π΅} β π π¦ ΜΈ|= {π΅, πΆ} β π ππ΄ β¨ ππΆ β² 1 1 0 0 π¦ |= {π΄, π΅} β π π¦ |= {π΄, π΅} β π Β¬ππ΄ β¨ Β¬ππ΅ 1 0 0 1 π¦ |= {π΄, πΆ β² } β π π¦ |= {π΄, πΆ β² } β π Β¬ππ΄ β¨ Β¬ππΆ β² 0 0 1 1 π¦ |= {πΆ, πΆ β² } β π π¦ |= {πΆ, πΆ β² } β π Β¬ππΆ β¨ Β¬ππΆ β² 0 1 0 1 π¦ ΜΈ|= {π΅, πΆ β² } β π π¦ ΜΈ|= {π΅, πΆ β² } β π ππ΄ β¨ ππΆ We briefly discuss similarities and differences between Algorithm 2 and Algorithm 4: Both algorithms systematically explore subsets of β and minimize the subsumee from such subset to compute explanations. Algorithm 2 constructs such subsets (β β π») manually by removing one axiom appearing in the previously computed explanation (if there is any) in all possible ways. Algorithm 4 enumerates such subsets π with the help of a SAT solver. The main difference is that Algorithm 2 may encounter the same subsets many times (on different branches), whereas the propositional encoding in Algorithm 4 ensures that such subsets never repeat. Of course, an iteration of Algorithm 2 cannot be directly compared to an iteration of Algorithm 4. Both iterations use at most one call to Algorithm 1, but Algorithm 4 may also require a call to Algorithm 3, as well as checking satisfiability of πΉ . The latter requires solving an NP-complete problem, for which no polynomial algorithm is known so far. In order to check satisfiability of πΉ , a SAT solver usually tries several (in worst-case exponentially many) propositional interpretations until a model of πΉ is found. As each such interpretation β corresponds to a subset π (β) β β, this process can be compared to the enumeration of subsets in Algorithm 2. However, a SAT solver usually implements a number of sophisticated optimizations, which make the search for models very efficient in practice, whereas the subset enumeration strategy used by Algorithm 2 is rather simplistic. Hence Algorithm 4 is likely to win in speed. On the other hand, Algorithm 4 requires saving all explanations (and minimal hitting sets) in the propositional formula πΉ , which might result in a formula of exponential size, if the number of such explanations or hitting sets is exponential. In this regard, Algorithm 2 could be more memory efficient since saving (all) explanations is optional (see the discussion at the end of Section 3.2). Hence both algorithms have their own advantages and disadvantages. 4. Conclusions We have presented two alternative algorithms for computing (all or some) solutions to the concept abduction problem: one algorithm is based on Reiterβs hitting set tree algorithm, whereas the other one relies on a SAT encoding. It remains to be analyzed how these algorithms behave in practice and how these algorithms differ on different real-world ontologies, where an important aspect is also finding efficient incremental SAT solvers. In contrast to previous work, the algorithms do not rely on a refutation-based calculus and, hence, can be used also with efficient reasoners for tractable DLs such as β°β and its extensions. Another direction for future work is extending the approach to other forms of (logic-based) abduction, e.g., to ABox abduction and a comparison with other existing approaches, which are mostly focusing on ABox abduction. References [1] C. S. Peirce, Deduction, Induction, and Hypothesis, Popular Science Monthly 13 (1878) 470β482. [2] T. Eiter, G. Gottlob, The complexity of logic-based abduction, J. ACM 42 (1995) 3β42. URL: https://doi.org/10.1145/200836.200838. [3] M. Denecker, A. C. Kakas, Abduction in logic programming, in: A. C. Kakas, F. Sadri (Eds.), Computational Logic: Logic Programming and Beyond, Essays in Honour of Robert A. Kowalski, Part I, volume 2407 of Lecture Notes in Computer Science, Springer, 2002, pp. 402β436. URL: https://doi.org/10.1007/3-540-45628-7_16. [4] C. Elsenbroich, O. Kutz, U. Sattler, A case for abductive reasoning over ontologies, in: B. C. Grau, P. Hitzler, C. Shankey, E. Wallace (Eds.), Proceedings of the OWLED*06 Workshop on OWL: Experiences and Directions, Athens, Georgia, USA, November 10- 11, 2006, volume 216 of CEUR Workshop Proceedings, CEUR-WS.org, 2006. URL: http: //ceur-ws.org/Vol-216/submission_25.pdf. [5] S. Klarman, U. Endriss, S. Schlobach, Abox abduction in the description logic ALC, J. Autom. Reason. 46 (2011) 43β80. URL: https://doi.org/10.1007/s10817-010-9168-z. [6] K. Halland, A. Britz, S. Klarman, Tbox abduction in ALC using a DL tableau, in: M. Bienvenu, M. Ortiz, R. Rosati, M. Simkus (Eds.), Informal Proceedings of the 27th International Workshop on Description Logics, Vienna, Austria, July 17-20, 2014, vol- ume 1193 of CEUR Workshop Proceedings, CEUR-WS.org, 2014, pp. 556β566. URL: http: //ceur-ws.org/Vol-1193/paper_42.pdf. [7] J. PukancovΓ‘, M. Homola, Abox abduction for description logics: The case of multiple observations, in: M. Ortiz, T. Schneider (Eds.), Proceedings of the 31st International Work- shop on Description Logics co-located with 16th International Conference on Principles of Knowledge Representation and Reasoning (KR 2018), Tempe, Arizona, US, October 27th - to - 29th, 2018, volume 2211 of CEUR Workshop Proceedings, CEUR-WS.org, 2018. URL: http://ceur-ws.org/Vol-2211/paper-31.pdf. [8] P. Koopmann, W. Del-Pinto, S. Tourret, R. A. Schmidt, Signature-based abduction for expressive description logics, in: D. Calvanese, E. Erdem, M. Thielscher (Eds.), Proceedings of the 17th International Conference on Principles of Knowledge Representation and Reasoning, KR 2020, Rhodes, Greece, September 12-18, 2020, 2020, pp. 592β602. URL: https://doi.org/10.24963/kr.2020/59. [9] W. Del-Pinto, R. A. Schmidt, Abox abduction via forgetting in ALC, in: The Thirty- Third AAAI Conference on Artificial Intelligence, AAAI 2019, The Thirty-First Innovative Applications of Artificial Intelligence Conference, IAAI 2019, The Ninth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2019, Honolulu, Hawaii, USA, January 27 - February 1, 2019, AAAI Press, 2019, pp. 2768β2775. URL: https://doi.org/10. 1609/aaai.v33i01.33012768. [10] B. Glimm, Y. Kazakov, Classical algorithms for reasoning and explanation in description logics, in: M. KrΓΆtzsch, D. Stepanova (Eds.), Reasoning Web. Explainable Artificial Intel- ligence - 15th International Summer School 2019, Bolzano, Italy, September 20-24, 2019, Tutorial Lectures, volume 11810 of Lecture Notes in Computer Science, Springer, 2019, pp. 1β64. URL: https://doi.org/10.1007/978-3-030-31423-1_1. [11] R. Reiter, A theory of diagnosis from first principles, Artificial Intelligence 32 (1987) 57β 95. URL: https://www.sciencedirect.com/science/article/pii/0004370287900622. doi:https: //doi.org/10.1016/0004-3702(87)90062-2. [12] M. Bienvenu, Complexity of abduction in the EL family of lightweight description logics, in: G. Brewka, J. Lang (Eds.), Principles of Knowledge Representation and Reasoning: Pro- ceedings of the Eleventh International Conference, KR 2008, Sydney, Australia, September 16-19, 2008, AAAI Press, 2008, pp. 220β230. URL: http://www.aaai.org/Library/KR/2008/ kr08-022.php. [13] M. W. Birte Glimm, Yevgeny Kazakov, Concept Abduction for Description Log- ics β Technical Report, Technical Report, Ulm University, Institute of Artifi- cial Intelligence, 2022. https://www.uni-ulm.de/fileadmin/website_uni_ulm/iui.inst.090/ Publikationen/2022/GlKW22a_report.pdf. [14] M. KrΓΆtzsch, F. SimanΔΓk, I. Horrocks, A description logic primer, CoRR abs/1201.4089 (2012). Available at http://arxiv.org/pdf/1201.4089.pdf. [15] T. Eiter, G. Gottlob, The complexity of logic-based abduction, J. ACM 42 (1995) 3β42. URL: https://doi.org/10.1145/200836.200838. doi:10.1145/200836.200838. [16] F. Baader, S. Brandt, C. Lutz, Pushing the β°β envelope further, in: Proc. 5th Workshop on OWL: Experiences and Directions (OWLEDβ08), volume 496, CEUR, 2008. [17] K. Schild, A correspondence theory for terminological logics: Preliminary report, in: J. Mylopoulos, R. Reiter (Eds.), Proc. 12th Int. Joint Conf. on Artificial Intelligence (IJCAIβ91), Morgan Kaufmann, 1991, pp. 466β471. [18] F. M. Donini, F. Massacci, EXPTIME tableaux for πβπ, J. of Artificial Intelligence 124 (2000) 87β138. [19] R. Greiner, B. A. Smith, R. W. Wilkerson, Readings in model-based diagnosis, Morgan Kaufmann Publishers Inc., 1992, pp. 49β53.