<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Journal of the ACM (JACM) 36 (1989) 929-965. doi:10.1145/76359.76371.
[22] M. Funk</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="doi">10.1016/J.FUTURE</article-id>
      <title-group>
        <article-title>SAT-Based Bounded Fitting for the Description Logic ℒ (Extended Abstract)</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Maurice Funk</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jean Christoph Jung</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Tom Voellmer</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Center for Scalable Data Analytics and Artificial Intelligence (ScaDS.AI)</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Leipzig University</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>TU Dortmund University</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <volume>13870</volume>
      <fpage>3</fpage>
      <lpage>6</lpage>
      <abstract>
        <p>Bounded fitting is a general paradigm for learning logical formulas from positive and negative data examples, that has received considerable interest recently. We investigate bounded fitting for concepts formulated in the description logic ℒ and its syntactic fragments. We show that the underlying size-restricted fitting problem is NP-complete for all studied fragments, even in the special case of a single positive and a single negative example. By design, bounded fitting is an Occam algorithm and thus is a sample-eficient PAC learning algorithm, regardless of the studied fragment. We complement this by showing that eficient PAC learning is impossible under standard complexity theoretic assumptions, and that other natural learning algorithms are typically not sample-eficient PAC learning algorithms. Finally, we present an implementation of bounded fitting in ℒ and its fragments based on a SAT solver. We discuss optimizations and compare our implementation to other concept learning tools.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Description Logic</kwd>
        <kwd>Bounded Fitting</kwd>
        <kwd>PAC Learning</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Learning description logic (DL) concepts from given data examples is an important task when working
with large knowledge bases [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ]. For the purpose of this paper, an example is a pair (ℐ, ) where ℐ
is a finite interpretation (describing, e.g., a database or a knowledge graph) and  is some individual
in ℐ. Moreover, a DL concept  fits a set  of positive examples and a set  of negative examples if
ℐ |= () for all (ℐ, ) ∈  and ℐ ̸|= () for all (ℐ, ) ∈  . We mention three applications. First,
the fitting concept may be used as an explanation of the separation between good and bad “scenarios”,
described by  and  , respectively. For example,  and  could be data describing users who visited
(resp., did not visit) a certain page, and a fitting  would explain the users’ behavior from their data.
Second, under the classical query-by-example paradigm [
        <xref ref-type="bibr" rid="ref3 ref4">3, 4</xref>
        ], a human user may reverse-engineer a
DL concept to be used as query by manually selecting elements they want to have returned ( ) or not
returned ( ), and the system comes up with an expression satisfying the demands. Finally, an ontology
engineer may seek a definition of some symbol  satisfied in the interpretation, so they may ask for a
concept separating the instances of  from the non-instances.
      </p>
      <p>
        In this paper, we study the problem of learning concepts formulated in the description logic ℒ,
which is the basic logic underlying the web ontology language OWL 2 DL [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], and its syntactic
fragments. The importance of finding fitting description logic concepts has resulted in both foundational
work [
        <xref ref-type="bibr" rid="ref6 ref7 ref8">6, 7, 8</xref>
        ] and systems. While most systems are based on heuristic search and refinement operators [
        <xref ref-type="bibr" rid="ref10 ref9">9,
10, 11, 12, 13</xref>
        ] or, more recently, also on neural techniques [14, 15], we approach the problem via bounded
iftting. Bounded fitting is a general paradigm for fitting logical formulas to positive and negative examples
that has been investigated recently for the description logic ℰℒ [16] and a range of other logics like
linear temporal logic LTL [17, 18] and computation tree logic CTL [19]. Algorithm 1 provides an
      </p>
      <p>Input: Positive examples  , negative examples 
1 for  := 1, 2, . . . do
2 if there is a concept  ∈ ℒ of size  that fits ,  then
3 return</p>
      <p>Algorithm 1: Bounded Fitting for description logic ℒ.
abstract description of bounded fitting for a given description logic ℒ. It should be clear that, if any
iftting concept exists, bounded fitting always returns a fitting concept of minimal size, which is often
a desirable property. From the practical perspective, human users typically prefer shorter, that is,
simpler concepts in the applications sketched above. From a theoretical perspective, this property
makes bounded fitting an Occam algorithm which implies that it comes with probabilistic generalization
guarantees in Valiant’s probably approximately correct (PAC) learning framework [20, 21]. Intuitively,
this means that bounded fitting needs only few examples to be able to generalize to unseen examples.</p>
      <p>The basic DL ℒ provides the logical constructors conjunction ⊓, disjunction ⊔, negation ¬,
existential restriction ∃, and universal restriction ∀ to build complex concepts from concept names
and ⊥ and ⊤. Motivated by the fact that, depending on the application, one may not need all concept
constructors, fragments ℒ() of ℒ have been studied which allow only a subset  ⊆ {⊓ , ⊔, ¬, ∃, ∀}
of the available constructors. For instance, the mentioned DL ℰℒ is defined by {⊓, ∃}, another popular
logic is ℱℒ0, which is defined by {⊓, ∀}. Bounded fitting has been studied recently for ℰℒ, and we
extend this study here to all other syntactical fragments.</p>
      <p>The paper corresponding to this extended abstract has been accepted for publication at ISWC 2025 [22].
The full paper with all proof details is available on arXiv [23].</p>
    </sec>
    <sec id="sec-2">
      <title>2. Contributions</title>
      <p>Our main contributions are as follows. First, we study the size-restricted fitting problem : given positive
examples  , negative examples  , and a size bound  in unary encoding, determine whether there
is a concept of size at most  that fits  and  . Clearly, this is precisely the problem to be solved in
Line 2 of bounded fitting. Then, motivated by the ability of bounded fitting to generalize well from
few examples, we investigate the generalization abilities of fitting algorithms for DLs ℒ() in Valiant’s
PAC learning framework. Finally, we provide an implementation of bounded fitting for ℒ and its
fragments, that relies on a SAT solver to solve size-restricted fitting in Line 2 of bounded fitting. We
give now a more detailed overview.</p>
      <p>Complexity of size-restricted fitting. We show that size-restricted fitting is NP-complete for ℒ
and all its syntactic fragments ℒ() such that  contains at least ∃ or ∀. This was known for the
fragment ℰℒ [24]. Containment in NP can be shown by a simple guess and check argument. The
lower bound is more technical and rather strong: it applies already in the case of only one positive
and one negative example and over a signature consisting of two role names and one concept name.
It thus strengthens the mentioned result for ℰℒ which requires a non-constant number of positive
examples. The proof is by reduction from the hitting set problem. The examples constructed in the
reduction admit a fitting ℒ concept if and only if there is a fitting ℒ(∃) concept, which means that it
shows NP-hardness for all ℒ() with  containing ∃. NP-hardness for the other fragments follows by
applying a duality principle.</p>
      <p>Theorem 1. Size-restricted fitting for ℒ() is NP-complete for every  ⊆ {⊓ , ⊔, ¬, ∃, ∀} with {∃, ∀} ∩
 ̸= ∅. This already holds if only a single positive and a single negative example are allowed, and over a
signature consisting of two role names and one concept name.</p>
      <p>Generalization. We investigate the learnability of ℒ concepts in Valiant’s PAC learning
framework [20]. A PAC learning algorithm is a fitting algorithm that, given suficiently many labeled examples
drawn from an unknown distribution, returns a concept that generalizes well (that is, has a small error
when evaluated over the entire distribution) with high probability. We call such an algorithm eficient if
it runs in polynomial time and sample-eficient if a polynomial number of examples sufices to ensure
the described probabilistic generalization guarantees. For a precise definition, see the full paper but
also [25]. We start with observing that under reasonable complexity theoretic assumptions, no ℒ()
admits an eficient PAC learning algorithm, that is, an algorithm that runs in polynomial time and
produces a concept that satisfies the definition of PAC learning. This is stated in the following theorem.
Theorem 2. Let  ⊆ {⊓ , ⊔, ¬, ∃, ∀}. If there is an eficient PAC learning algorithm for
ℒ(), then:
1. NP = RP, if  contains at least one of ∃/∀ and {⊓, ⊔} ̸⊆ ;
2. RSA encryption is polynomial time invertible, if {⊓, ⊔} ⊆ .</p>
      <p>We then analyze the generalization ability of fitting algorithms that have favorable properties from
a logical perspective in that they return fitting concepts that are most specific, most general, or of
minimal quantifier depth among all fitting concepts. More precisely, we investigate whether there can
be PAC learning algorithms with such properties that are sample-eficient. We show that, with one
exception, all such algorithms are not sample-eficient, and hence do not generalize well. This was
already known for the fragment ℰℒ of ℒ [16], and some of our proofs rely on similar techniques.
Our results are summarized by the following theorem.</p>
      <p>Theorem 3. Let  ⊆ {⊓ , ⊔, ¬, ∃, ∀} be any set containing at least one of ∃/∀ and at least one of ⊓/⊔,
and let  be a fitting algorithm for ℒ(). Then  is not a sample-eficient PAC learning algorithm, if:
1.  ̸= {∃, ⊔} and  always returns a most specific fitting if one exists;
2.  ̸= {∀, ⊓} and  always returns a most general fitting if one exists;
3.  always returns a fitting of minimal quantifier depth if some fitting exists.</p>
      <p>The exceptions in the theorem are the cases of  = {∃, ⊔} and  = {∀, ⊓}. For these fragments,
bounded fitting is a sample-eficient PAC learning algorithm that returns a most specific or most general,
respectively, fitting concept if it exists.</p>
      <p>Implementation. We implemented bounded fitting for ℒ and its fragments using a SAT solver
to decide the NP-complete size-restricted fitting problems by encoding size-restricted fitting into a
propositional formula.1 We present two optimizations of the basic encoding, one where the structure of
concepts is precomputed and then supplied to the SAT solver and another, where types of elements
are used instead of individual concept names. Additionally, our implementation supports approximate
iftting, the optimization variant of the fitting problem, where one searches for a concept that fits as
many positive and negative examples as possible.</p>
      <p>
        We compare our implementation ALC-SAT+ to other systems that support learning of ℒ concepts,
namely CELOE [12], SParCEL [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], and EvoLearner [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], considering both exact fitting and approximate
1Our implementation is available at https://github.com/SAT-based-Concept-Learning/ALCSAT.
iftting. For evaluating exact fitting we generated sets of positive and negative examples from a fragment
of the YAGO knowledge graph [26], see the full paper for details. For approximate fitting, we compared
the systems on the SML benchmarks [27]. We measured the accuracy and length of the returned
concepts using 10-fold cross validation. Our results on the SML benchmarks are shown in Table 1
where the first line in each cell is the accuracy and the second line is the length of the returned concept;
in both cases, the ± -term denotes the standard deviation. Our tool achieves competitive values for both
accuracy and concept length. In some instances, ALC-SAT+ may return a larger concept compared
to the other tools, however, this means that the accuracy reported cannot be achieved with a smaller
concept.
      </p>
    </sec>
    <sec id="sec-3">
      <title>Acknowledgments</title>
      <sec id="sec-3-1">
        <title>Jean Christoph Jung was supported by DFG project JU 3197/1-1.</title>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Declaration on Generative AI</title>
      <sec id="sec-4-1">
        <title>The authors have not employed any Generative AI tools.</title>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J.</given-names>
            <surname>Lehmann</surname>
          </string-name>
          ,
          <article-title>Learning OWL Class Expressions, volume 6 of Studies on the Semantic Web</article-title>
          , IOS Press,
          <year>2010</year>
          . doi:
          <volume>10</volume>
          .3233/978-1-
          <fpage>61499</fpage>
          -340-7-i.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>J.</given-names>
            <surname>Lehmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Fanizzi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Bühmann</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>d'Amato, Concept learning</article-title>
          ,
          <source>in: Perspectives on Ontology Learning</source>
          , AKA / IOS Press,
          <year>2014</year>
          , pp.
          <fpage>71</fpage>
          -
          <lpage>91</lpage>
          . URL: https://jens-lehmann.org/files/2014/pol_concept_ learning.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>M. M. Zloof</surname>
          </string-name>
          ,
          <article-title>Query by example</article-title>
          ,
          <source>in: American Federation of Information Processing Societies: 1975 National Computer Conference</source>
          ,
          <volume>19</volume>
          -
          <fpage>22</fpage>
          May
          <year>1975</year>
          , Anaheim, CA, USA, volume
          <volume>44</volume>
          <source>of AFIPS Conference Proceedings</source>
          , AFIPS Press,
          <year>1975</year>
          , pp.
          <fpage>431</fpage>
          -
          <lpage>438</lpage>
          . doi:
          <volume>10</volume>
          .1145/1499949.1500034.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>D. M. L. Martins</surname>
          </string-name>
          ,
          <article-title>Reverse engineering database queries from examples: State-of-the-art, challenges</article-title>
          , and research opportunities,
          <source>Information Systems</source>
          <volume>83</volume>
          (
          <year>2019</year>
          )
          <fpage>89</fpage>
          -
          <lpage>100</lpage>
          . doi:
          <volume>10</volume>
          .1016/J.IS.
          <year>2019</year>
          .
          <volume>03</volume>
          .002.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>I.</given-names>
            <surname>Horrocks</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. F.</given-names>
            <surname>Patel-Schneider</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. van Harmelen</given-names>
            ,
            <surname>From</surname>
          </string-name>
          <string-name>
            <surname>SHIQ</surname>
          </string-name>
          and
          <article-title>RDF to OWL: the making of a web ontology language</article-title>
          ,
          <source>Journal of Web Semantics</source>
          <volume>1</volume>
          (
          <year>2003</year>
          )
          <fpage>7</fpage>
          -
          <lpage>26</lpage>
          . doi:
          <volume>10</volume>
          .1016/J.WEBSEM.
          <year>2003</year>
          .
          <volume>07</volume>
          .001.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>M.</given-names>
            <surname>Funk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. C.</given-names>
            <surname>Jung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Lutz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Pulcini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Wolter</surname>
          </string-name>
          ,
          <article-title>Learning description logic concepts: When can positive and negative examples be separated?</article-title>
          ,
          <source>in: Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, IJCAI-19</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>1682</fpage>
          -
          <lpage>1688</lpage>
          . doi:
          <volume>10</volume>
          .24963/ijcai.
          <year>2019</year>
          /233.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J.</given-names>
            <surname>Lehmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Hitzler</surname>
          </string-name>
          ,
          <article-title>Concept learning in description logics using refinement operators</article-title>
          ,
          <source>Machine Learning</source>
          <volume>78</volume>
          (
          <year>2010</year>
          )
          <fpage>203</fpage>
          -
          <lpage>250</lpage>
          . doi:
          <volume>10</volume>
          .1007/s10994-009-5146-2.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>M.</given-names>
            <surname>Funk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. C.</given-names>
            <surname>Jung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Lutz</surname>
          </string-name>
          ,
          <article-title>Actively learning concept and conjunctive queries under ℰ ℒ- ontologies</article-title>
          , in
          <source>: Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, IJCAI-21</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>1887</fpage>
          -
          <lpage>1893</lpage>
          . doi:
          <volume>10</volume>
          .24963/ijcai.
          <year>2021</year>
          /260.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>S.</given-names>
            <surname>Heindorf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Blübaum</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Düsterhus</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Werner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. N.</given-names>
            <surname>Golani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Demir</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ngonga</surname>
          </string-name>
          <string-name>
            <surname>Ngomo</surname>
          </string-name>
          ,
          <article-title>EvoLearner: Learning description logics with evolutionary algorithms</article-title>
          ,
          <source>in: WWW '22: The ACM Web Conference</source>
          <year>2022</year>
          , ACM,
          <year>2022</year>
          , pp.
          <fpage>818</fpage>
          -
          <lpage>828</lpage>
          . doi:
          <volume>10</volume>
          .1145/3485447.3511925.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>A. C.</given-names>
            <surname>Tran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Dietrich</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. W.</given-names>
            <surname>Guesgen</surname>
          </string-name>
          ,
          <string-name>
            <surname>S.</surname>
          </string-name>
          <article-title>Marsland, Parallel symmetric class expression learning</article-title>
          ,
          <source>Journal of Machine Learning Research</source>
          <volume>18</volume>
          (
          <year>2017</year>
          )
          <volume>64</volume>
          :
          <fpage>1</fpage>
          -
          <lpage>64</lpage>
          :
          <fpage>34</fpage>
          . URL: https://jmlr.org/papers/v18/
          <fpage>14</fpage>
          -
          <lpage>317</lpage>
          .html.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>