<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <article-id pub-id-type="doi">10.24963/IJCAI.2021</article-id>
      <title-group>
        <article-title>PAC Learning of Concept Inclusions for Ontology- Mediated Query Answering (Extended Abstract)</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Sergei Obiedkov</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Barış Sertkaya</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Faculty of Compter Science and Engineering, Frankfurt University of Applied Sciences</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Faculty of Computer Science / cfaed / ScaDS.AI, TU Dresden</institution>
          ,
          <addr-line>Dresden</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2020</year>
      </pub-date>
      <volume>12507</volume>
      <fpage>19</fpage>
      <lpage>27</lpage>
      <abstract>
        <p>This extended abstract summarizes results from [1], where we propose a practical method for learning axioms in a Description Logic (DL) ontology using techniques from probably approximately correct (PAC) learning. The goal is to support ontology-mediated query answering (OMQA) [2, 3] by approximating an unknown TBox  through interaction with a domain expert oracle that can decide whether a concept inclusion (CI)  ⊑  is entailed by  . Such an oracle may be instantiated in diferent ways-for example, as a human domain expert; a large language model (LLM); a dataset representative of the domain; or a large, complex ontology from which a smaller, focused one is to be distilled. Our method learns subsumption relationships among a finite set  of concept descriptions, called the base set. This base set constrains the search space of candidate axioms and can be chosen to suit the application-e.g., all concept names, combinations of concept names with existential restrictions up to a ifxed role depth, or a tailored selection relevant to the user. We do not fix a particular DL or a set of constructors; our results apply to arbitrary DLs that support conjunction. The algorithm also employs a sampling oracle that generates CIs over  according to a fixed but arbitrary distribution . Given ,  ∈ (0, 1) , it runs in time polynomial in the relevant parameters and returns a TBox  ′ such that, with probability at least 1 −  (over the algorithm's random choices), the probability (under ) that a CI over  is entailed by exactly one of  and  ′ is at most . We also show how to direct the learning process toward subsumptions that are particularly relevant to a given ABox , by adapting the distribution . This enables the learned axioms to improve recall in query answering over incomplete datasets. Experimental evaluation on benchmark ontologies confirms the efectiveness of our approach. For related work on ontology learning in DLs, see [4, 5, 6, 7, 8, 9, 10, 11, 12].</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Ontologies</kwd>
        <kwd>Description Logics</kwd>
        <kwd>Active Learning</kwd>
        <kwd>Knowledge Acquisition</kwd>
        <kwd>PAC Learning</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
    </sec>
    <sec id="sec-2">
      <title>2. PAC Learning of Concept Inclusions</title>
      <p>
        Pr (︀  | ( |= ) ⇐⇒ ( ′ ̸|= ))︀ ≤ ,

using subsumption queries. We modify the equivalence oracle so that, instead of returning a model of
exactly one of the two non-equivalent Horn formulas, it returns the GCI corresponding to an Horn
clause that is entailed by exactly one of the two formulas. A probably approximately correct algorithm
is obtained by replacing each call to this equivalence oracle with an appropriate number of calls to a
suitable sampling oracle. Please see [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] for more details.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Varying the Query Distribution</title>
      <p>Our definition of approximation involves a distribution  of subsumption queries. This distribution is
meant to reflect the interests of the user of the ontology we are trying to learn. In a basic scenario, we
may assume that users explicitly pose subsumption queries to the ontology and that  is the distribution
of these queries.</p>
      <p>
        A more practically relevant scenario is given by ontology-mediated query answering [
        <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
        ]: Given a
query (a concept description)  and a knowledge base  = ( , ), find all instances of  in . The
TBox  may be only partially known or not known at all, in which case we may use our PAC algorithm
to learn its approximation through interaction with an expert or from a representative dataset. In
this setting, an approximation may be considered good if it ensures high precision and recall of query
answering.
      </p>
      <p>
        Definition 1. Let  = ( , ) be a knowledge base,  ′ be a TBox, and  be a query. Using certain-answer
semantics [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], we define cert(, ) as the set of individual names  from  satisfying  |= (). The
precision and recall of  ′ for  on  are, respectively,
      </p>
      <p>P( ′, ) = | cert(, ( ′, )) ∩ cert(, )| , R( ′, ) = | cert(, ( ′, )) ∩ cert(, )| .</p>
      <p>| cert(, ( ′, ))| | cert(, )|
If the denominator is 0, then the value of the corresponding measure is defined to be 1.</p>
      <p>There are two standard ways to aggregate precision and recall for several queries: macroaveraging
and microaveraging [14].</p>
      <p>Definition 2. Let  = ( , ) be a knowledge base,  ′ be a TBox, and  be a finite set of queries. The
macro precision and recall of  ′ for  on  are the average values of the precision and recall over all
queries from :</p>
      <p>Pmacro( ′, ) =
∑︀∈ P( ′, )
||
and Rmacro( ′, ) =
∑︀∈ R( ′, )
||
.</p>
      <p>The micro precision Pmicro( ′, ) and micro recall Rmicro( ′, ) are defined, respectively, as
∑︀∈ | cert(, ( , )) ∩ cert(, ( ′, ))| and ∑︀∈ | cert(, ( , )) ∩ cert(, ( ′, ))|
.
∑︀∈ | cert(, ( ′, ))|
∑︀∈ | cert(, ( , ))|</p>
      <p>The goal in our OMQA scenario is to learn an approximation  ′ of  with high values of the
macro/micro precision and recall for some set  of queries. If  ′ is a lower approximation of  , then
the precision for every query is 1, and so are the macro and micro precision. In this case, we aim to
maximize the recall. Next we describe a heuristic approach to choosing the distribution of subsumption
queries in the learning algorithm so as to increase the micro recall on a given ABox .</p>
      <p>Consider a subsumption query d  ⊑ . If we care about micro recall, it seems particularly important
to ask this query whenever d  has a lot of instances in 0 = (∅, ), since a positive answer to the
query would then allow us to correctly assert () for many individuals . Therefore, a reasonable
approach seems to be to generate the left-hand sides d  of subsumption queries proportionally to
| cert(d  , 0)|. Regarding the right-hand sides, if () rarely occurs in , this may be due to two
reasons:  is a rare concept, or  is a generalization of other concepts and () can be inferred
from the target TBox  together with what is explicitly asserted in  about . We cannot tell which
of the two it is; so we may want to assume the second case to be on the safe side. Then, we may
want to generate the right-hand sides  of subsumption queries with probabilities proportional to
| Ind() ∖ cert(, 0)|, i.e., to the number of individuals that are not (yet) known to be instances of .</p>
      <p>A problem with this approach is that  ⊑  cannot be learned if  has no instances in 0. To
address this, we need to change the distribution on the fly, so as to take into account what has already
been learned. Thus, having learned  ⊑ , we update 0 by replacing 0 = ∅ with 1 = { ⊑ }
and recalculate the probabilities involved in sampling premises with respect to 1 = (1, ). Now,
| cert(, 1)| &gt; 0, which makes it possible to learn  ⊑ .</p>
      <p>
        This was the method used in the experiments we presented in [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. However, it prioritizes concepts
d  with a large number of instances in  even when these instances are the same for many diferent
 . This may sometimes negatively afect precision or recall for certain concepts in
. Instead, when
sampling left-hand sides of CIs, we should try to maximize the coverage of individuals in . Therefore, in
the experiments presented here, we adopt the following two-stage approach: first sample an individual
 from  uniformly at random and then sample a subset of { ∈  |  ∈ cert(, )} also uniformly
      </p>
    </sec>
    <sec id="sec-4">
      <title>4. Experimental Evaluation</title>
      <p>learned.
We implemented our approach in a prototype tool, paclo1, and evaluated it in the OMQA context. The
expert was simulated using the target TBox  ; i.e., the response to a subsumption query  is positive if
and only if  |= . Subsumption queries were answered with the ELK reasoner [15]. We set  = 0.001
to ensure a high probability of obtaining the desired approximation and averaged results over five runs.
Each setting is defined by a signature, an approximation type (  - or lower  -approximation), and a query
distribution  (uniform or -induced, as described in the previous section).</p>
      <p>
        We tested on six KBs: four generated with OWL2Bench [16] and two from the ORE 2015
repository [17]. Due to space constraints, we report results here only for the KB ore_ont_5596; results for the
other KBs can be found in [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. The ore_ont_5596 KB contains 58 class names, 33 role names, 322 GCIs,
112,320 individuals, 32,990 class assertions and 190,149 role assertions. Two base sets were considered:
the macro and micro recall on the query set  = . The results are shown in Table 1.
1 with all class names and 2 that adds ∃.⊤ for each role , yielding 93 concepts in total. We measure
      </p>
      <p>For each of 1 and 2, the first line represents the precision and recall of the empty  ′, i.e., the quality
of query answering based only on the ABox. This serves our baseline. We omit  -approximations with
the uniform distribution, since it provides hardly any improvement over the baseline. The best macro
and micro recall values for each  are shown in bold. The column | ′| contains the number of axioms</p>
      <p>Overall, -induced distributions yield substantially higher recall, particularly for micro recall.Lower
approximations typically have slightly reduced recall, but may be preferable when perfect precision
is required. Lower approximations for the uniform distribution do show some improvement over the
baseline, but usually smaller than those for the -induced distribution and with larger sets of GCIs.
Note that the perfect recall may not be achievable with a fixed base set , since  may contain axioms
mixing concepts from and outside .</p>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgments</title>
      <p>This work is partly supported by DFG in project 389792660 (TRR 248, Center for Perspicuous Systems),
by BMBF in ScaDS.AI, and by BMBF and DAAD in project 57616814 (SECAI).</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the authors used ChatGPT in order to: Improve writing style.
After using this tool, the authors reviewed and edited the content as needed and take full responsibility
for the publication’s content.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>S.</given-names>
            <surname>Obiedkov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Sertkaya</surname>
          </string-name>
          ,
          <article-title>PAC learning of concept inclusions for ontology-mediated query answering</article-title>
          ,
          <source>International Journal of Approximate Reasoning</source>
          <volume>186</volume>
          (
          <year>2025</year>
          )
          <article-title>109523</article-title>
          . URL: https://www. sciencedirect.com/science/article/pii/S0888613X25001641. doi:https://doi.org/10.1016/j. ijar.
          <year>2025</year>
          .
          <volume>109523</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>M.</given-names>
            <surname>Bienvenu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ortiz</surname>
          </string-name>
          ,
          <article-title>Ontology-mediated query answering with data-tractable description logics</article-title>
          , in: W. Faber,
          <string-name>
            <surname>A</surname>
          </string-name>
          . Paschke (Eds.),
          <source>Reasoning Web. Web Logic Rules - 11th International Summer School</source>
          <year>2015</year>
          , Berlin, Germany,
          <source>July 31 - August 4</source>
          ,
          <year>2015</year>
          , Tutorial Lectures, volume
          <volume>9203</volume>
          of Lecture Notes in Computer Science, Springer,
          <year>2015</year>
          , pp.
          <fpage>218</fpage>
          -
          <lpage>307</lpage>
          . URL: https://doi.org/10.1007/ 978-3-
          <fpage>319</fpage>
          -21768-
          <issue>0</issue>
          _9. doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>319</fpage>
          -21768-
          <issue>0</issue>
          _
          <fpage>9</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>M.</given-names>
            <surname>Bienvenu</surname>
          </string-name>
          ,
          <article-title>Ontology-mediated query answering: Harnessing knowledge to get more from data</article-title>
          , in: S. Kambhampati (Ed.),
          <source>Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence, IJCAI</source>
          <year>2016</year>
          , New York, NY, USA,
          <fpage>9</fpage>
          -
          <issue>15</issue>
          <year>July 2016</year>
          , IJCAI/AAAI Press,
          <year>2016</year>
          , pp.
          <fpage>4058</fpage>
          -
          <lpage>4061</lpage>
          . URL: http://www.ijcai.org/Abstract/16/600.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>F.</given-names>
            <surname>Baader</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Ganter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Sertkaya</surname>
          </string-name>
          , U. Sattler,
          <article-title>Completing description logic knowledge bases using formal concept analysis</article-title>
          , in: M. M.
          <string-name>
            <surname>Veloso</surname>
          </string-name>
          (Ed.),
          <source>Proceedings of the Twentieth International Joint Conference on Artificial Intelligence (IJCAI'07)</source>
          , AAAI Press,
          <year>2007</year>
          , pp.
          <fpage>230</fpage>
          -
          <lpage>235</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>D.</given-names>
            <surname>Borchmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Distel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Kriegel</surname>
          </string-name>
          ,
          <article-title>Axiomatisation of general concept inclusions from finite interpretations</article-title>
          ,
          <source>Journal of Applied Non-Classical Logics</source>
          <volume>26</volume>
          (
          <year>2016</year>
          )
          <fpage>1</fpage>
          -
          <lpage>46</lpage>
          . URL: https://doi.org/10.1080/11663081.
          <year>2016</year>
          .
          <volume>1168230</volume>
          . doi:
          <volume>10</volume>
          .1080/11663081.
          <year>2016</year>
          .
          <volume>1168230</volume>
          . arXiv:https://doi.org/10.1080/11663081.
          <year>2016</year>
          .
          <volume>1168230</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>B.</given-names>
            <surname>Konev</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Lutz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ozaki</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Wolter</surname>
          </string-name>
          ,
          <article-title>Exact learning of lightweight description logic ontologies</article-title>
          ,
          <source>Journal of Machine Learning Research</source>
          <volume>18</volume>
          (
          <year>2018</year>
          )
          <fpage>1</fpage>
          -
          <lpage>63</lpage>
          . URL: http://jmlr.org/papers/v18/
          <fpage>16</fpage>
          -
          <lpage>256</lpage>
          .html.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>A.</given-names>
            <surname>Ozaki</surname>
          </string-name>
          ,
          <article-title>On the complexity of learning description logic ontologies</article-title>
          , in: M.
          <string-name>
            <surname>Manna</surname>
            ,
            <given-names>A</given-names>
          </string-name>
          . Pieris (Eds.),
          <source>Reasoning Web. Declarative Artificial Intelligence: 16th International Summer School</source>
          <year>2020</year>
          , Oslo, Norway, June 24-26,
          <year>2020</year>
          , Tutorial Lectures, Springer International Publishing, Cham,
          <year>2020</year>
          , pp.
          <fpage>36</fpage>
          -
          <lpage>52</lpage>
          . URL: https://doi.org/10.1007/978-3-
          <fpage>030</fpage>
          -60067-
          <issue>9</issue>
          _2. doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>030</fpage>
          -60067-
          <issue>9</issue>
          _
          <fpage>2</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>A.</given-names>
            <surname>Ozaki</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Persia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mazzullo</surname>
          </string-name>
          ,
          <article-title>Learning query inseparable ℰ ℒℋ ontologies</article-title>
          ,
          <source>Proceedings of the AAAI Conference on Artificial Intelligence</source>
          <volume>34</volume>
          (
          <year>2020</year>
          )
          <fpage>2959</fpage>
          -
          <lpage>2966</lpage>
          . URL: https://ojs.aaai.org/index. php/AAAI/article/view/5688. doi:
          <volume>10</volume>
          .1609/aaai.v34i03.
          <fpage>5688</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>M.</given-names>
            <surname>Funk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. C.</given-names>
            <surname>Jung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Lutz</surname>
          </string-name>
          ,
          <article-title>Actively learning concepts and conjunctive queries under elrontologies</article-title>
          , in: Z.
          <string-name>
            <surname>Zhou</surname>
          </string-name>
          (Ed.),
          <source>Proceedings of the Thirtieth International Joint Conference on</source>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>