<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Learning Multiple Description Logics Concepts</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Raphael Melo</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Kate Revoredo</string-name>
          <email>katerevoredog@uniriotec.br</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Aline Paes</string-name>
          <email>alinepaes@ic.uff.br</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science Institute of Computing</institution>
          ,
          <addr-line>UFF Rio de Janeiro</addr-line>
          ,
          <country country="BR">Brazil</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Postgraduate Information Systems Program</institution>
          ,
          <addr-line>UNIRIO Rio de Janeiro</addr-line>
          ,
          <country country="BR">Brazil</country>
        </aff>
      </contrib-group>
      <fpage>17</fpage>
      <lpage>22</lpage>
      <abstract>
        <p>Description logics based languages have became the standard representation scheme for ontologies. They formalize the domain knowledge using interrelated concepts, contained in terminologies. The manual de nition of terminologies is an expensive and error prone task, therefore automatic learning methods are a necessity. In this paper we lay the foundations of a multiple concept learning method that uses virtual concepts to aid the learning process, yielding more compact and readable terminologies. In this paper, we de ne virtual concepts and how they can be implemented in the current concept learning methods. We show through experiments how the method stacks up against other multiple concept learning methods.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Description logics (DLs) [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] form a family of knowledge representation languages,
with di erent expressive power, that are typically decidable fragments of rst
order logic (FOL). With DL it is possible to represent domain concepts and their
relations. Moreover, due to their computational power and expressiveness, they
have been widely used in Semantic Web [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] for ontology representation.
      </p>
      <p>
        The task of de ning the domain knowledge through a DL is usually done
manually, which is time consuming and error prone, even more because the
domain experts themselves do not always agree about the de nitions of concepts
and their relationships [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Therefore, to consider applying machine learning
techniques [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] for automatically learning in DLs is relevant and sometimes even
required. A number of DL learning approaches have been proposed in the
literature [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. In these approaches each concept is learned independently from each
other, thus, none of the concepts being learned are considered in the de nition
of the others. If this assumption is removed, the nal ontology could be clearer
and closer to the way that the concepts are related in the underlying domain.
In this sense, another question arises: "What is the best order to learn a set of
1 The rst author would like to thank CAPES for the nancial support through a
master scholarship.
related concepts?" In [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], we address this problem by de ning an algorithm that
discovers a taxonomy of the concepts being learned and then uses this taxonomy
to de ne the order for learning the concepts. However, since the order is de ned
by a heuristic aimed at nding relations between concepts and subconcepts, it
may be the case that the ordering found does not yield the best solution for
the learning task. Moreover, during the learning process only previously learned
concepts are considered.
      </p>
      <p>
        In this paper, we lay the foundations of multiple concept learning using the
ideas discussed in [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] as motivation. Thus, we propose a learning strategy that
allows concepts not yet learned to appear in the de nition of other concepts,
thus making possible to learn more compact and readable terminologies.
      </p>
      <p>The paper is organized as follows. In Section 2, Description Logic and concept
learning are reviewed. In Section 3, we present the proposed approach to learn
multiple concepts. Section 4 presents some preliminary experimental results.
Section 5 concludes the paper and presents the next steps of the research.
2</p>
    </sec>
    <sec id="sec-2">
      <title>DLs and concept learning in DLs</title>
      <p>DLs knowledge bases (KB) have two components: a TBox and an ABox. The TBox
contains intensional knowledge in the form of a terminology. Knowledge is
expressed in terms of individuals, concepts, and roles. Thus, the terminology
consists of concepts, which denote a set of individuals and roles which denote binary
relationships between individuals. In this paper, we assume the common
assumption made about DL terminologies: (i) only one de nition for a concept name and
(ii) concept de nitions are acyclic. The ABox is composed of assertions about
the individuals of the domain. An assertion states that an individual belongs
to a concept or that a pair of individuals satis es a role. Attached to a DL's
KB there must be a reasoning mechanism, responsible for inferring information
about individuals from the KB.</p>
      <p>
        There are a number of existing approaches to automatically learn DLs
concepts, most of them [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] are inspired by Inductive Logic Programming(ILP) [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]
techniques. The goal is to induce concept descriptions from existing evidences.
When learning a concept, one has the purpose of nding a generalized and
correct de nition of such a concept from a set of examples, as de ned below:
      </p>
      <sec id="sec-2-1">
        <title>De nition 1 (Concept Learning)</title>
        <p>Given:
{ a knowledge base KB,
{ a target concept T arget such as T arget 2= KB,
{ a set of target examples E, divided into positive (Ep) and negative (En)
examples, such that E = Ep [ En</p>
        <sec id="sec-2-1-1">
          <title>Find: { A de nition of the concept C(T arget</title>
          <p>KB \ C 6j= En.</p>
          <p>C) such that KB [ C j= Ep and</p>
          <p>Although De nition 1 requires that the learned concept covers all the positive
examples and none of the negative examples, these hard criteria are usually
relaxed to enable the induction.</p>
          <p>The concept learning task can be expanded to multiple concept learning, as
de ned below:</p>
        </sec>
      </sec>
      <sec id="sec-2-2">
        <title>De nition 2 (Multiple Concept Learning)</title>
        <p>Given:
{ knowledge base KB,
{ n concept learning tasks fAC1 ; AC2 ; : : : ; ACn g, where ACi = fEpi ; Eni g</p>
        <sec id="sec-2-2-1">
          <title>Find:</title>
          <p>{ KB0 = KB [ T , where T is a taxonomy which holds the concepts de nitions
found in the n concept learning tasks.</p>
          <p>In this paper we focus on multiple concept learning methods that return a
T that have compact de nitions.</p>
          <p>
            In [
            <xref ref-type="bibr" rid="ref7">7</xref>
            ] we presented a pre-processing method for multiple concept learning
called terminology learning. It yields compact terminologies by de ning an order
to execute each concept learning task. This order is de nied by nding, before the
learning process, all the subsumee and subsumer relationships among the concept
learning tasks using the shared individuals in the example sets as evidence.
          </p>
          <p>However, although the subsumee and subsumer relation is important when
devising an order, it is not the only relationship among concepts that can impact
the later learned de nition. Thus, in this paper we follow a di erent approach
to nd out the concepts that should be used to de ne another concept. Instead
of directly nding a taxonomy from the set of examples, we let the learning
algorithm decide what is the best way of de ning each concept.
3</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Multiple DL Concept Learning</title>
      <p>
        The concept learning task can be viewed as a search problem over the space
of concepts, created using three basic elements [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]: (i) a re nement operator
to build the search tree of concepts; (ii) a search algorithm to control how this
search tree is traversed; (iii) a scoring function to evaluate the nodes of the tree
and to point out the best current concept candidate.
      </p>
      <p>The re nement operator is responsible for de ning a number of rules, from
which a valid candidate de nition for a concept is yielded. The candidate de
nitions are created from combinations of known concepts, roles and constructors.
The constructors are di erent according to the DL language chosen. We propose
to take into account an additional type of concept when the re nement operator
is generating a concept de nition, henceforth called virtual concept.</p>
      <p>Virtual concepts are concepts that do not have an explicit de nition yet. As
usual, these concepts have a set of examples related to it, divided into positives
and negative examples. Once a de nition for a concept is learned, it should have
considered the individuals marked as positive examples as belonging to it. On
the other hand, the negative examples are the individuals that do not belong
to that concept and its de nition should be able to indicate that. Since each
concept has these sets of examples associated to it, it is possible to take into
account the set of examples instead of the concept de nition when referring to
a concept. In this way, we can say that, while a concept is not selected to be the
target one, it also behaves as a virtual concept.</p>
      <p>In order to consider virtual concepts in the concept learning task, it is
necessary to make the scoring function cope with them. Usually, the scoring function
is either the cover relationship or a variation of it. The cover relationship is a
function that evaluates how many positive and negative examples are inferred
by a candidate de nition. So this can be achieved by transforming the example
sets of a virtual concept into assertion inside the ABox, e.g., Albert 2 Ep of C,
then the assertion C(Albert) will be added to the ABoxes of all concept learning
tasks that could use the virtual concept C.</p>
      <p>
        We argue that adding the assertions of a virtual concept can have the same
result as that of the regular inference if the following assumptions holds: all the
virtual concepts in the multiple concept learning task should share the same
ABox and all the relevant individuals for a particular virtual concept should be
covered in one of its example sets. If that is indeed the case, it is possible to use
virtual concepts in the same way that other non-target instantiated concepts are
used inside the concept learning task. Moreover, the addition of virtual concepts
in the learning process will allow it to nd concept de nitions more compact
then the ones found with the terminology learning method [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. The terminology
learning method only deals with one type of usage relationship between virtual
concepts, the subsumee and subsumer relationship. This kind of relationship
is only related to the concept and subconcept relationship, e.g. in the kinship
domain we have Grandfather v Father. However, it can not reliably work with
relationships that di ers from subsumer and subsumee, for example, the
disjunction between Grandfather and Grandmother to de ne Grandparent. The method
proposed in this paper is capable of dealing with it because we turn the learning
task into the responsible component for nding relationships between concepts.
      </p>
      <p>The proposed method has the local goal to nd the most compact de nition
for a single virtual concept. The major concern that arises from this is the
formation of cycles. This could be avoided by two di erent approaches: (i) when
constructing the ABox' for a target concept, the addition of facts associated to
virtual concepts that uses the current target concept in their de nition can be
avoided. (ii) with a post-processing procedure. The rst one is likely to be more
e cient in returning a solution, but does not guarantee that this is the optimal
solution, while the second may have a better chance to return the optimal
solution, but some concepts may be relearned several times, i.e., the learning process
can become less e cient.</p>
    </sec>
    <sec id="sec-4">
      <title>Preliminary Experiments</title>
      <p>To evaluate the proposed method some experiments were devised considering
the concepts GRANDPARENT (GP), GRANDFATHER (GF) and
GRANDMOTHER (GM) from the Family ontology.</p>
      <p>A knowledge base of the family domain was used to run the experiments 3.
Each individual cited in the knowledge base is either a positive example or a
negative example to a concept. The description learning component chosen to
learn the concepts is the DL-Learner system with default settings, since it is a
largely used environment to learn DLs.</p>
      <p>The rst analysis we conducted showed that our proposal, Multiple Concept
Learning (MCL), is able to learn a terminology more compact and readable than
the Concept Learning (CL) approach, which learns the concepts individually and
independently. Table 1 shows the de nition for the three concepts. Notice that
GRANDPARENT is de ned as a disjunction of the two other concepts.</p>
      <p>Another evaluation concerns whether MCL is able to learn a more compact
terminology when compared with the Terminology Learning (TL) task. For this
comparison, we considered two concept orders for MCL:
(i) &lt; GRANDPARENT; GRANDFATHER; GRANDMOTHER &gt; and
(ii) &lt; GRANDMOTHER; GRANDFATHER; GRANDPARENT &gt;.</p>
      <p>
        The rst one, was found by the approach proposed in [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] and then is the same
one used by TL. Moreover, we avoid cycles by changing the correspondent ABox
as described in Section 3. The results in Table 1 shows that di erent learning
orders yield di erent results, and that the proposed method can achieve the same
result as the terminology learning by reversing the order. This demonstrates that
the proposed method is more versatile than the terminology learning, because it
isn't bound to a learning order, while still maintaining the accuracy.
      </p>
      <p>To sum up, these results point out that the method has the potential for
nding compact solutions when avoiding cycles with pre-processing (MCL +
Order 1 or 2), and its ability to nd optimal solutions, if a post processing
method to avoid cycle is used (MCL).
5</p>
    </sec>
    <sec id="sec-5">
      <title>Conclusion and future remarks</title>
      <p>
        In this paper we laid the foundations over which a multiple concept learning
method could be built. We de ned a new type of concept, the virtual concept,
analogous with the de nition of the concept learning task, but with a twist to
make it usable inside the existing learning process of another concept. The
proposed method opens the possibility of nding all the possible usage relationships
among virtual concepts. Because of this, we argued that the use of virtual
concepts could yield better results than the ones found with the method presented
in [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] and in regular concept learning methods. We also presented directions to
deal with the cycle problem that may appear with the use of this method. An
3 ftp://ftp.cs.utexas.edu/pub/mooney/forte
Experiment Concept De nition Length
      </p>
      <p>GP EXISTS parent.EXISTS parent.TOP. 5
CL GF (male AND EXISTS parent.EXISTS parent.TOP). 7
GM (female AND EXISTS parent.EXISTS parent.TOP). 7</p>
      <p>GP (grandfather OR grandmother). 3
MCL GF (grandparent AND male). 3
GM (grandparent AND female). 3</p>
      <p>GP EXISTS parent.EXISTS parent.TOP. 5
TL=CL+Order GF (grandparent AND male). 3
GM (grandparent AND female). 3</p>
      <p>GP (grandfather OR grandmother). 3
MCL+Order 1 GF (male AND EXISTS married.grandmother). 5
GM (female AND EXISTS parent.EXISTS parent.TOP). 7</p>
      <p>GP EXISTS parent.EXISTS parent.TOP. 5
MCL+Order 2 GF (grandparent AND male). 3</p>
      <p>GM (grandparent AND female). 3
experiment concerning the Kinship domain showed that it is possible to learn a
clearer and more compact terminology without requiring a good ordering of the
concepts.</p>
      <p>In the future we would like to analyze the behavior of the method on di erent
data sets and di erent DL languages, since we believe the proposed method is
capable to work with all possible constructors sets.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <given-names>F.</given-names>
            <surname>Baader</surname>
          </string-name>
          and
          <string-name>
            <given-names>W.</given-names>
            <surname>Nutt</surname>
          </string-name>
          , \
          <article-title>Basic description logics," in The description logic handbook (F.</article-title>
          <string-name>
            <surname>Baader</surname>
            ,
            <given-names>D. L.</given-names>
          </string-name>
          <string-name>
            <surname>McGuinness</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Nardi</surname>
            , and
            <given-names>P. F.</given-names>
          </string-name>
          <string-name>
            <surname>Patel-Schneider</surname>
          </string-name>
          , eds.), pp.
          <volume>47</volume>
          {
          <issue>100</issue>
          , Cambridge University Press, 2 ed., may
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <given-names>T.</given-names>
            <surname>Berners-Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Hendler</surname>
          </string-name>
          , and
          <string-name>
            <given-names>O.</given-names>
            <surname>Lassila</surname>
          </string-name>
          , \
          <article-title>The semantic web,"</article-title>
          <source>Scienti c American</source>
          , vol.
          <volume>5</volume>
          , no.
          <issue>284</issue>
          , p.
          <fpage>34</fpage>
          ,
          <year>2001</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <given-names>A.</given-names>
            <surname>Maedche</surname>
          </string-name>
          and
          <string-name>
            <given-names>S.</given-names>
            <surname>Staab</surname>
          </string-name>
          , \
          <article-title>Ontology learning for the semantic web,"</article-title>
          <source>IEEE Intelligent Systems and Their Applications</source>
          , vol.
          <volume>16</volume>
          , no.
          <issue>2</issue>
          , pp.
          <volume>72</volume>
          {
          <issue>79</issue>
          ,
          <year>2001</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4. T. Mitchell, Machine Learning. New York:
          <string-name>
            <surname>McGraw-Hill</surname>
          </string-name>
          ,
          <year>1997</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <given-names>J.</given-names>
            <surname>Lehmann</surname>
          </string-name>
          and
          <string-name>
            <given-names>P.</given-names>
            <surname>Hitzler</surname>
          </string-name>
          , \
          <article-title>Concept learning in description logics using re nement operators,"</article-title>
          <source>Machine Learning</source>
          , vol.
          <volume>78</volume>
          , no.
          <issue>1-2</issue>
          , pp.
          <volume>203</volume>
          {
          <issue>250</issue>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <given-names>N.</given-names>
            <surname>Fanizzi</surname>
          </string-name>
          , C. d'Amato, and
          <string-name>
            <given-names>F.</given-names>
            <surname>Esposito</surname>
          </string-name>
          , \
          <article-title>Dl-foil concept learning in description logics,"</article-title>
          <source>in Proceedings of the 18th International Conference on Inductive Logic Programming (ILP-2008)</source>
          , vol.
          <source>5194 LNAI of Lecture Notes in Computer Science</source>
          , pp.
          <volume>107</volume>
          {
          <issue>121</issue>
          , Springer,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <given-names>R.</given-names>
            <surname>Melo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Revoredo</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Paes</surname>
          </string-name>
          , \
          <article-title>Terminology learning through taxonomy discovery," BRACIS "to appear"</article-title>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <given-names>L.</given-names>
            <surname>Iannone</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Palmisano</surname>
          </string-name>
          , and
          <string-name>
            <given-names>N.</given-names>
            <surname>Fanizzi</surname>
          </string-name>
          , \
          <article-title>An algorithm based on counterfactuals for concept learning in the semantic web,"</article-title>
          <source>Applied Intelligence</source>
          , vol.
          <volume>26</volume>
          , no.
          <issue>2</issue>
          , pp.
          <volume>139</volume>
          {
          <issue>159</issue>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9. L. De Raedt,
          <source>Logical and Relational Learning</source>
          . Springer,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>