<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Reasoning by Analogy in Description Logics through Instance-based Learning</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Claudia d'Amato</string-name>
          <email>claudia.damato@di.uniba.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Nicola Fanizzi</string-name>
          <email>fanizzi@di.uniba.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Floriana Esposito</string-name>
          <email>esposito@di.uniba.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Dipartimento di Informatica, Universita` degli Studi di Bari Campus Universitario</institution>
          ,
          <addr-line>Via Orabona 4, 70125 Bari</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>- This work presents a method founded in instancebased learning for inductive (memory-based) reasoning on ABoxes. The method, which exploits a semantic dissimilarity measure between concepts and instances, can be employed both to answer class membership queries and to predict new assertions that may be not logically entailed by the knowledge base. In a preliminary experimentation, we show that the method is sound and it is actually able to induce new assertions that might be acquired in the knowledge base.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>I. INTRODUCTION</title>
      <p>
        Most of the research on ontology reasoning has been
focusing on methods based on deductive reasoning. However,
important tasks that are likely to be provided by new
generation knowledge-based systems, such as construction, revision,
population and evolution are likely to be supported also by
inductive methods. This has brought to an increasing interest
in machine learning and knowledge discovery methods applied
to ontological representations (see [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] and, more recently,
[
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]).
      </p>
      <p>
        We propose an algorithm which is based on a notion of
concept similarity for performing a form of lazy learning on
typical ontological representations. Namely, by combining an
instance-based (analogical) approach with a notion of semantic
dissimilarity, this paper intends to demonstrate the
applicability of inductive reasoning in this field which can be considered
another form of approximate reasoning (see discussion in
[
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]). In particular, we have adapted the general
instancebased learning approach like the k- Nearest Neighbor [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] to
the specific multi-relational setting for ontology languages.
A couple of technical problems had to be solved for this
adaptation to ontology representations: 1) the Open World
Assumption (OWA) that is made in this context; 2) in this
multi-class problem setting disjunction cannot be assumed by
default.
      </p>
      <p>
        The standard ontology languages are founded in Description
Logics (henceforth DLs) as they borrow the language
constructors for expressing complex concept definitions.
Instancebased methods depend on a similarity measure defined on
the instance space. In this perspective, a variety of measures
for concept representations have been proposed (see [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] for
a survey). As pointed out in [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], most of these measures
focus on the similarity of atomic concepts within hierarchies or
simple ontologies, based on a few relations. Thus, it becomes
necessary to investigate similarity in more complex languages.
      </p>
      <p>
        It has been observed that, adopting richer representations,
the structural properties have less and less impact in assessing
semantic similarity. Hence, the vision of similarity based only
on a structural (graph-based) approach, such as in [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]
may fall short. We have proposed some dissimilarity measures
for non trivial DL languages, based on the semantics conveyed
by the ABox assertions, which are suitable for being used
in instance-based methods [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. These measures elicit
the underlying semantics by querying the knowledge base
for assessing the concept extensions, estimated through their
retrieval [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ], as also hinted in [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. Besides, the overall
similarity is also (partially) influenced by the concepts which
are related through role restrictions. Moreover, in many other
typical tasks (e.g. conceptual clustering or definition), it is
necessary to assess the similarity between concepts (resp.
individuals). By recurring to the notion of most specific
concept of an individual with respect to an ABox [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ], as
representatives of the individuals at the concept level, the
measures for concepts can be extended to such cases.
      </p>
      <p>This analogical reasoning procedure like this may be
employed to answering class membership queries through
analogical rather than deductive reasoning which may be more robust
with respect to noise and is likely to suggest new knowledge
(which was not logically derivable). Specifically we developed
the method also for an application of semantic web service
discovery where services are annotated in DLs.</p>
      <p>Another application might regard supporting various tasks
for the knowledge engineer, such as the acquisition of
candidate assertions for enriching ontologies with partially
populated ABoxes: the outcomes given by the procedure can be
utilized as recommendations. Indeed, as we show in the
experimentation, the newly induced assertions are quite accurate
(commission errors, i.e. predicting a concept erroneously, were
rarely observed). In turn, the outcomes of the procedure may
also trigger other related tasks such as induction (revision) of
(faulty) knowledge bases.</p>
      <p>The paper is organized as follows. In the next section,
the representation language is briefly presented. Two concept
dissimilarity measures are recalled and exploited in a modified
version of the k-NN classification procedure. The results of a
preliminary experimental evaluation of the method using these
two measures are shown and, finally, possible developments
of the method are examined.</p>
      <p>II. ALC KNOWLEDGE BASES</p>
      <p>
        We recall the basics of ALC [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], a logic which adopts
constructors supported by the standard ontology languages
(see the DL handbook [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] for a thorough reference).
      </p>
      <p>In DLs, descriptions are inductively defined starting with a
set NC of primitive concept names and a set NR of primitive
roles. Complex descriptions are built using primitive concepts
and roles and the constructors showed in the following. The
semantics of the descriptions is defined by an interpretation
I = (ΔI , ·I ), where ΔI is a non-empty set, the domain of
the interpretation, and ·I is the interpretation function that
maps each A ∈ NC to a set AI ⊆ ΔI and each R ∈ NR to
RI ⊆ ΔI × ΔI .</p>
      <p>The top concept &gt; is interpreted as the whole domain of
objects ΔI , while the bottom concept ⊥ corresponds to ∅.
Complex descriptions can be built in ALC using the following
constructors. The language supports full negation: given any
description C, denoted ¬C, it amounts to ΔI \ CI . Concept
conjunction, denoted C1 u C2, yields the extension C1I ∩ C2I ;
dually, concept disjunction, denoted C1 t C2, yields the union
C1I ∪ C2I . Finally, the existential restriction, denoted ∃R.C, is
interpreted as {x ∈ ΔI | ∃y ∈ ΔI ((x, y) ∈ RI ∧ y ∈ CI )}
and the value restriction ∀R.C has as its extension the set
{x ∈ ΔI | ∀y ∈ ΔI ((x, y) ∈ RI → y ∈ CI )}.</p>
      <p>The main inference is subsumption between concepts based
on their semantics: given two descriptions C and D, C
subsumes D, denoted by C w D, iff for every interpretation
I it holds that CI ⊇ DI . When C w D and D w C then
they are equivalent, denoted with C ≡ D.</p>
      <p>A knowledge base K = hT , Ai contains a TBox T and an
ABox A:
• T is the set of definitions C ≡ D, meaning that, for every
interpretation I, CI = DI , where C is the concept name
and D is its description constructed as above;
• A contains concept and role assertions about the
worldstate, e.g. C(a) and R(a, b), meaning that, for every I,
aI ∈ CI and (aI , bI ) ∈ RI .</p>
      <p>
        A related inference used in the following is instance
checking, that is deciding whether an individual is an instance of a
concept [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ], [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. Conversely, it may be necessary to find the
concepts which an individual belongs to (realization problem),
especially the most specific one (see Def. 5).
      </p>
      <p>
        Semantically equivalent (yet syntactically different)
descriptions can be given for the same concept. Nevertheless,
equivalent concepts can be reduced to a normal form by means of
rewriting rules that preserve their equivalence [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ]:
      </p>
      <p>A description D is in ALC normal form iff D ≡ ⊥ (then
D := ⊥) or if D ≡ &gt; (then D := &gt;) or if D = D1 t · · · t Dn
(∀i = 1, . . . , n, Di 6≡ ⊥) with
l</p>
      <p>A u l
A∈prim(Di)
• exR(Di) is the set of concepts in the existential
restrictions of the role R.</p>
      <p>For any R, every sub-description in exR(Di) and valR(Di) is
in normal form.</p>
      <p>In the following, let L = ALC/≡ be the quotient set of
ALC descriptions in normal form.</p>
      <p>III. DISSIMILARITY MEASURES IN DESCRIPTION LOGICS</p>
      <p>
        We recall two definition of dissimilarity measures for ALC
descriptions expressed in normal form [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. These
measures are based on both the structure and the semantics of the
concept descriptions.
      </p>
      <sec id="sec-1-1">
        <title>A. Overlap Dissimilarity Measure</title>
        <p>The first measure, is derived from a measure of the overlap
between concepts, that can be defined as follows:</p>
        <p>Definition 1 (overlap function): Let I be the canonical
interpretation of the ABox A. The overlap function f is defined1:
fC :=LF×n L 7→ R+ defined for descriptions C, D ∈ L, with
i=1 Ci and D = Fm</p>
        <p>j=1 Dj
disjunctive level:
f (C, D) := ft(C, D) =  0∞
 maxi,j fu(Ci, Dj )</p>
        <p>C ≡ D
C u D ≡ ⊥
otherwise</p>
        <p>X ft(valR(Ci), valR(Dj ))</p>
        <p>R∈NR
existential restrictions:</p>
        <p>N
f∃(Ci, Dj ) := RX∈NR kX=1 p=m1,a..x.,M ft(Cik, Djp)
where Cik ∈ exR(Ci) and Djp ∈ exR(Dj ) and we suppose
w.l.o.g. that N = |exR(Ci)| ≥ |exR(Dj )| = M , otherwise the
indices N and M as well as Ci and Dj are to be exchanged
in the formula above.</p>
        <p>1The name A of the ABox is omitted for simplicity.</p>
        <p>The function f represents a measure of the overlap between
two descriptions expressed in ALC normal form. It measures
the similarity of the input concepts based on the similarity
between their extensions (approximated with the retrieval) and
also on the similarity of the concepts reached by the role
restrictions. Namely, It is defined recursively beginning from
the top level of the descriptions (a disjunctive level) up to
the bottom level represented by (conjunctions of) primitive
concepts.</p>
        <p>
          Overlap at the disjunctive level is treated as the maximum
overlap between the disjunctive forms of the input concepts. At
conjunctive levels, instead of simply considering the similarity
like in Tversky’s measure [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ] (as we do for prim’s), that
could turn out to be null in case of disjoint retrieval sets for
the input concepts, we distinguish the overlaps between the
prims and those between the concepts in the scope of the
various role restrictions2; the contribution of the overlap of
concepts reached through role restrictions can be penalized by
tweaking the parameter λ. The measure for primitive concepts
resembles Tversky’s measure and it represents a semantic
baseline since it depends on the semantics of the knowledge
base, as conveyed by the ABox assertions. This is in line with
to the ideas in [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ], [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ], where semantics is elicited as a
probability distribution over the domain of the interpretation.
        </p>
        <p>As for the role restrictions overlap, for the universal
restrictions we simply add the overlap measures varying the role,
while for the existential restrictions the measure is trickier:
borrowing the idea of the existential mappings we consider, per
role, all possible matches between the concepts in the scope
of existential restrictions, then we consider the maximal sum
of overlaps resulting from all the matches.</p>
        <p>Now, it is possible to derive a dissimilarity measure based
on f as follows:</p>
        <p>
          Definition 2 (overlap dissimilarity measure): The overlap
dissimilarity measure is a function d : L × L 7→ [
          <xref ref-type="bibr" rid="ref1">0, 1</xref>
          ] such
that, given two concept descriptions C, D ∈ L, C = Fin=1 Ci
and D = Fm
j=1 Dj :
        </p>
        <p> 1 if f (C, D) = 0
d(C, D) :=  0 if f (C, D) = ∞</p>
        <p> f (C1,D) otherwise</p>
        <p>Function d simply measures the level of dissimilarity
between two concepts as the inverse of the overlap function
f . Particularly, if f (C, D) = 0, i.e. there is no overlap
between the considered concepts, then d must indicate that
the two concepts are totally different, indeed d(C, D) = 1,
the maximum value of its range. If f (C, D) = ∞ this means
that the two concepts are totally overlapped and consequently
d(C, D) = 0 that means that the two concept are
indistinguishable, indeed d assumes the minimum value of its range.
If the considered concepts have a partial overlap then their
2We tried also other solutions, such as considering minima or products of
the three overlap measures, yet experimentally this did not yield a better
performance whereas the computation time was increased. For practical
reasons, the maximal measure ∞ has been replaced with large numbers
depending on the cardinality of the set of individuals in the ABox: |Ind(A)|.
dissimilarity is inversely proportional to their overlap, since
in this case f (C, D) &gt; 1 and consequently 0 &lt; d(C, D) &lt; 1.</p>
      </sec>
      <sec id="sec-1-2">
        <title>B. A Dissimilarity Measure Based on Information Content</title>
        <p>
          As discussed in [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ], a measure of concept (dis)similarity
can be derived from the notion of Information Content (IC)
that, in turn, depends on the probability of an individual to
belong to a certain concept. Now, differently from other works,
which assume that a probability distribution for the concepts
in an ontology is known, we derive it from the knowledge
base, from the distribution that can be estimated therein [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ],
[
          <xref ref-type="bibr" rid="ref10">10</xref>
          ].
        </p>
        <p>
          In order to approximate this probability for a certain concept
C, we recur to its extension w.r.t. the considered ABox in a
fixed interpretation. Namely, we chose the canonical
interpretation IA, which is the one adopting the set of individuals
mentioned in the ABox as its domain and the identity as
its interpretation function [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ]. Now, given a concept C its
probability is estimated by:
        </p>
        <p>pr(C) = |CIA |/|ΔIA |
Finally, we can compute the information content of a concept,
employing this probability:</p>
        <p>I C(C) = − log pr(C)</p>
        <p>
          A measure of the concept dissimilarity is now formally
defined [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ]:
        </p>
        <sec id="sec-1-2-1">
          <title>Definition 3 (IC dissimilarity): Let A be an ABox with</title>
          <p>canonical interpretation I. The information content
dissimilarity measure is a function g : L × L 7→ R+ defined recursively
for any two normal form concept descriptions C, D ∈ L, with
C = Fn j=1 Dj</p>
          <p>i=1 Ci and D = Fm
disjunctive level:



g(C, D) := gt(C, D) = 
0
∞
min gu(Ci, Dj )
11 ≤≤ ij ≤≤ nm</p>
          <p>if C ≡ D
if C u D = ⊥
otherwise
gu(Ci, Dj ) := gP (prim(Ci), prim(Dj ))+</p>
          <p>+λ(g∀(Ci, Dj ) + g∃(Ci, Dj ))
conjunctive level:
where Cik ∈ exR(Ci) and Djp ∈ exR(Dj ) and we suppose
w.l.o.g. that N = |exR(Ci)| ≥ |exR(Dj )| = M , otherwise the
indices N and M are to be exchanged in the formula above.</p>
          <p>The function g represents a measure of the dissimilarity
between two descriptions expressed in ALC normal form.
It is defined recursively beginning from the top level of
the descriptions (a disjunctive level) up to the bottom level
represented by (conjunctions of) primitive concepts.</p>
          <p>Now g has values in [0, ∞]. It may be useful to derive a
normalized dissimilarity measure as shown in the following.</p>
        </sec>
        <sec id="sec-1-2-2">
          <title>Definition 4 (normalized IC dissimilarity): Let A be an</title>
          <p>
            ABox with canonical interpretation I. The normalized
information content dissimilarity measure is the function d :
L × L 7→ [
            <xref ref-type="bibr" rid="ref1">0, 1</xref>
            ], such that given the concept descriptions in
ALC normal form C = Fin=1 Ci and D = Fm
j=1 Dj , let
 0
d(C, D) :=  1
          </p>
          <p>1
 1 − g(C,D)
if g(C, D) = 0
if g(C, D) = ∞
otherwise</p>
        </sec>
      </sec>
      <sec id="sec-1-3">
        <title>C. Measuring the Dissimilarity between Individuals</title>
        <p>The notion of Most Specific Concept is commonly exploited
for lifting individuals to the concept level.</p>
        <sec id="sec-1-3-1">
          <title>Definition 5 (most specific concept): Given an ABox A</title>
          <p>and an individual a, the most specific concept of a w.r.t. A is
the concept C, denoted MSCA(a), such that A |= C(a) and
for any other concept D such that A |= D(a), it holds that
C v D.</p>
          <p>
            In case of cyclic ABoxes expressed in a DL with
existential restrictions the MSC may not be expressed by a finite
description [
            <xref ref-type="bibr" rid="ref15">15</xref>
            ], yet it may be often approximated.
          </p>
          <p>
            On performing experiments related to another similarity
measure exclusively based on concept extensions [
            <xref ref-type="bibr" rid="ref21">21</xref>
            ], we
noticed that, recurring to the MSC for lifting individuals to
the concept level, just falls short: indeed the MSCs may be
too specific and unable to include other (similar) individuals
in their extensions. By comparing descriptions reduced to the
normal form we have given a more structural definition of
dissimilarity. However, since MSCs are computed from the
same ABox assertions, reflecting the current knowledge state,
this guarantees that structurally similar representations will be
obtained for semantically similar concepts.
          </p>
          <p>Let us recall that, given the ABox, it is possible to calculate
the most specific concept of an individual a w.r.t. the ABox,
MSC(a) or at least its approximation MSCk(a) up to a
certain description depth k. In some cases these are equivalent
concepts but in general we have that MSCk(a) w MSC(a).</p>
          <p>Given two individuals a and b in the ABox, we consider
MSCk(a) and MSCk(b) (supposed in normal form). Now, in
order to assess the dissimilarity between the individuals, d
measure can be applied to these descriptions:</p>
          <p>d(a, b) := d(MSCk(a), MSCk(b))</p>
          <p>This may turn out to be handy in several tasks, namely both
in inductive reasoning (construction, repairing of knowledge
bases) and in information retrieval.</p>
        </sec>
      </sec>
      <sec id="sec-1-4">
        <title>D. Discussion</title>
        <p>
          We proved in [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ], [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ] that these measures are really
dissimilarity measures according to the formal definition [
          <xref ref-type="bibr" rid="ref22">22</xref>
          ],
considering that their input (what is actually compared) is
equivalence classes in L, i.e. ALC concept descriptions with
the same normal form.
        </p>
        <p>As previously mentioned, we have also attempted slightly
different measure definitions (e.g. considering minima or
products at the conjunctive level that sound more intuitive) which
experimentally did not prove more effective than the simple
(additive) definition given above.</p>
        <p>
          The computational complexity of the measures depends
on the complexity of the required reasoning services for
computing the concepts retrieval. Namely both subsumption
and instance-checking are P-space for the ALC logic[
          <xref ref-type="bibr" rid="ref18">18</xref>
          ], yet
such inference can be computed once and preliminarily, before
the measures are computed for the method we will present in
the following.
        </p>
        <p>Obviously when computing the dissimilarity measures for
cases involving individuals, the extra cost of computing MSCs
(or their approximations) has to be added.</p>
        <p>Nevertheless, in practical applications, these computations
may be efficiently carried out exploiting the statistics that
are maintained by the DBMSs query optimizers. Besides, the
counts that are necessary for computing the concept extensions
could be estimated by means of the probability distribution
over the domain.</p>
        <p>IV. A NEAREST NEIGHBOR CLASSIFICATION PROCEDURE</p>
        <p>IN DESCRIPTION LOGICS</p>
        <p>We briefly review the basics of the k-Nearest Neighbor
method (k-NN ) and propose how to exploit the classification
procedure for inductive reasoning. In this lazy approach to
learning, a notion of distance (or dissimilarity) measure for
the instance space is employed to classify a new instance.</p>
        <p>Let xq be the instance that requires a classification. Using
a dissimilarity measure, the set of k nearest pre-classified
instances is selected. The objective is to learn a discrete-valued
target function h : IS 7→ V from a space of instances IS to
a set of values V = {v1, . . . , vs}. In its simplest setting, the
k-NN algorithm approximates h for new instances xq on the
ground of the value that h assumes in the neighborhood of
xq, i.e. the k closest instances to the new instance in terms
of a dissimilarity measure. Precisely, it is assigned according
to the value which is voted by the majority of instances in
the neighborhood. The the classification function hˆ can be
formally defined as follows:
k
hˆ(xq) := argmax X δ(v, h(xi))</p>
        <p>v∈V i=1
where δ is a function that returns 1 in case of matching
arguments and 0 otherwise. Note that the hypothesized function hˆ
is defined only extensionally, that is the k-NN method does not
return an intensional classification model (e.g. a function or a
concept definition), it merely gives an answer for new query
instances to be classified, employing the procedure above.</p>
        <p>This simple formulation does not take into account the
similarity among instances, except when selecting the instances to
be included in the neighborhood. Therefore a modified setting
is generally adopted, weighting the vote on the grounds of the
similarity of the instance:
k
hˆ(xq) := argmax X wiδ(v, h(xi))
v∈V i=1
(1)
where, usually, wi = 1/d(xi, xq) or wi = 1/d(xi, xq)2.</p>
        <p>Usually this method is employed to classify vectors of
features in some n-dimensional instance space (e.g. often
IS = Rn). Let us now turn to adapt the method to the more
complex context of DLs descriptions. Preliminarily, it should
be observed that a strong assumption of this setting is that it
can be employed to assign a value (e.g. a class) to a query
instance among a set of values which can be regarded as a set
of pairwise disjoint concepts/classes. This is an assumption
that cannot be always valid. In this case, indeed, an individual
could be an instance of more than one concept.</p>
        <p>Let us consider a new value set V = {C1, . . . , Cs} of
concepts Cj that may be assigned to a query instance xq.
If they were to be considered as disjoint (like in the standard
machine learning setting), the decision procedure would adopt
the hypothesis function defined in Eq. (1), with the query
instance assigned the single concept voted by the weighted
majority of instances in the neighborhood.</p>
        <p>In the general case considered in this paper, when the
disjointness of the classes cannot be assumed (unless explicitly
stated in the TBox), one can adopt another answering
procedure, decomposing the multi-class problem into smaller binary
classification problems (one per concept). Therefore, a simple
binary value set (V = {−1, +1}) is to be employed. Then,
for each single concept (say Cj ), a hypothesis hˆj is computed
on the fly during the classification phase:
v∈V</p>
        <p>k
hˆj (xq) := argmax X δ(v, hj (xi))</p>
        <p>d(xq, xi)2
i=1
∀j ∈ {1, . . . , s}
(2)
where each function hj , simply indicates the occurrence (+1)
or absence (−1) of the corresponding assertion in the ABox for
the k training instances xi: Cj (xi) ∈ A. Alternately3, hj may
return +1 when Cj (xi) is logically entailed by the knowledge
base K, and −1 otherwise.</p>
        <p>The problem with non-explicitly disjoint concepts is also
related to the Closed World Assumption usually made in the
context of Information Retrieval and Machine Learning. That
is the reason for adapting the standard setting to cope both with
the case of non-disjoint concepts and with the OWA which is
commonly made in the Semantic Web context. To deal with
the OWA, the absence of information on whether a certain
instance x belongs to the extension of concept Cj should not
3For the sake of simplicity and efficiency, this case will not be considered
in the following.
be interpreted negatively; rather, it should count as neutral
information. Thus, one can still adopt the decision procedure
in Eq. (2), however another value set has to be adopted for
the hj ’s, namely V = {−1, 0, +1}, where the three values
denote, respectively, positive occurrence, absence and negative
occurrence (positive for the concept negation). Formally:
hj (x) =
 +1
 0
 −1</p>
        <p>Cj (x) ∈ A
Cj (x) 6∈ A ∧ ¬Cj (x) 6∈ A
¬Cj (x) ∈ A</p>
        <p>Again, a more complex procedure may be devised by simply
substituting the notion of occurrence (absence) of assertions in
(from) the ABox with one based on logic entailment (denoted
with `) from the knowledge base, i.e. K ` Cj (x), K 6` Cj (x)
nor K 6` ¬Cj (x) and K ` ¬Cj (x), respectively. Although this
may help reaching the precision of deductive reasoning, it is
also much more computationally expensive, since the simple
lookup in the ABox must be replaced with instance checking.</p>
        <p>
          From a computational viewpoint this procedure could be
implemented to provide an answer even more efficiently than
with a standard deductive reasoner. Indeed, once the
retrieval of the primitive concepts is computed, the dissimilarity
measures can be easily computed by means of a dynamic
programming algorithm. Besides, the various measures could
be maintained in an ad hoc data structure which would allow
for an efficient retrieval of the nearest neighbors, such as the
kD-trees or ball trees [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ]
        </p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>V. EXPERIMENTS</title>
      <sec id="sec-2-1">
        <title>A. Experimental Setting</title>
        <p>In order to assess the validity of the presented method
with the dissimilarity measures proposed in Sect. III, we have
applied it to the instance classification problem, with four
different ontologies represented in OWL: FSM,
SURFACEWATER-M ODEL from the Prote´ge´ library4, the FINANCIAL
ontology5 employed as a testbed for the PELLET reasoner and
a small FAMILY ontology written in our lab. Although they
are represented in languages that are different from ALC, we
simply discarded these details when constructing the MSC
approximations to be able to apply the presented measures.</p>
        <p>FAMILY is an ALCF ontology describing kinship
relationships. It is made up of 14 concepts (both primitive
and defined), some of them are declared to be disjoint, 5
object properties, 39 distinct individual names. Most of the
individuals are asserted to be instances of more than one
concept, and are involved in more than one role assertions.
This ontology has been written to have a small yet more
complex case with respect to the following ones. Indeed,
while the other ontologies are more regular, i.e. only some
concepts are employed in the assertions (the others are defined
only intensionally), in the FAMILY ontology every concept
4See the webpage:
http://protege.stanford.edu/plugins/owl/owl-library
5See the webpage: http://www.cs.put.poznan.pl/
alawrynowicz/financial.owl
has at least one instance asserted. The same happens for the
assertions on roles; particularly, there are some cases where
role assertions constitute a chain from an individual to another
one, by means of other intermediate assertions.</p>
        <p>The FSM ontology describes the domain of finite state
machines using the SOF (D) language. It is made up of
20 (both primitive and defined) concepts (some of them are
explicitly declared to be disjoint), 10 object properties, 7
datatype properties, 37 distinct individual names. About half
of the individuals are asserted as instances of a single concept
and are not involved in any role (object property).</p>
        <p>SURFACE-W ATER-M ODEL is an ALCOF (D) ontology
describing the domain of the surface water and the water quality
models. It is based on the Surface-water Models Information
Clearinghouse (SMIC) of the USGS. Namely, it is an ontology
of numerical models for surface water flow and water quality
simulation. The application domain of these models comprises
lakes, oceans, estuaries etc.. These models are classified based
on their availability, application domain, dimensions, partial
differential equation solver, and characteristics types. It is
made up of 19 concepts (both primitive and defined) without
any specification about disjointness, 9 object properties, 115
distinct individual names; each of them is an instance of a
single class and only some of them are involved in object
properties.</p>
        <p>FINANCIAL is an ALCIF ontology that describes the
domain of eBanking. It is made up of 60 (both primitive and
defined) concepts (some of them are declared to be disjoint),
17 object properties, and no datatype property. It contains
17941 distinct individual names. From the original ABox, we
randomly extracted assertions for 652 individuals.</p>
        <p>The classification method was applied to all the individuals
in each ontology; namely, the individuals were checked to
assess if they were instances of the concepts in the ontology
through the analogical method. The performance was
evaluated comparing its responses to those returned by a standard
reasoner6 as a baseline.</p>
        <p>Specifically, for each individual in the ontology the MSC
is computed and enlisted in the set of training (or test)
examples. Each example is classified applying the adapted
kNN method presented in the previous section. As a value of k
we chose p|Ind(A)|, as advised in the instance-based learning
literature.</p>
        <p>The experiment has been repeated twice adopting a
leaveone-out cross validation procedure with both the dissimilarity
measures defined in Section III.</p>
        <p>For each concept in the ontology, we measured the
following parameters for the evaluation:
• match rate: number of cases of individuals that got
exactly the same classification by both classifiers with
respect to the overall number of individuals;
• omission error rate: amount of unlabeled individuals (our
method could not determine whether it was an instance
6We employed PELLET: http://pellet.owldl.com
or not) while it was to be classified as an instance of that
concept;
• commission error rate: amount of individuals
(analogically) labeled as instances of a concept, while they
(logically) belong to that concept or vice-versa
• induction rate: amount of individuals that were found to
belong to a concept or its negation, while this information
is not logically derivable from the knowledge base
We report the average rates obtained over all the concepts in
each ontology and also their standard deviation.</p>
      </sec>
      <sec id="sec-2-2">
        <title>B. Experiments Employing the Overlap Measure</title>
        <p>By looking at Tab. I reporting the experimental outcomes
with the dissimilarity measure based on the overlap (see Def.
2), preliminarily it is important to note that, for every ontology,
the commission error was quite low. This means that the
classifier did not make critical mistakes i.e. cases when an
individual is deemed as an instance of a concept while it really
is an instance of another disjoint concept.</p>
        <p>In particular, by looking at the outcomes related to the
FAMILY ontology, it can be observed that the match rate
is the lowest while the highest rate of omission errors was
reported. This may be due to two facts: 1) very few individuals
were available w.r.t. the number of concepts7; 2) sparse data
situation: instances are irregularly spread over the concepts,
that is they might be instances of different concepts, which
are sometimes disjoint. Hence the MSC approximations that
were computed also resulted very different one from another,
which reduces the possibility of significantly matching similar
MSCs. This is a known drawback of the Nearest-Neighbor
methods. Nevertheless, as mentioned above, it is important to
note that the algorithm did not make any commission error
and it is able to infer new knowledge (11%).</p>
        <p>As regards the FSM ontology, we have observed the
maximum match rate with respect to the classification given by the
logic reasoner. Moreover, differently from the other ontologies,
both the omission error rate and induction rate were null. A
very limited percentage of incorrect classification cases was
observed. These outcomes were probably due to the fact that
individuals in this ontology are quite regularly divided by the
assertions on concepts and roles, so computing their MSCs,
these are all very similar to each other and so the amount
7Instance-based methods make an intensive use of the information about
the individuals and improve their performance with the increase of the number
of instances considered.
it is well known that the NN procedure becomes more and
more precise as more instances can be considered. The price
to be paid was a longer computation time.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>VI. CONCLUSIONS AND FUTURE WORK</title>
      <p>Match Commission Omission Induction</p>
      <p>Rate Rate Rate Rate
FAMILY .608±.230 .000±.000 .330±.216 .062±.217</p>
      <p>FSM .899±.178 .096±.179 .000±.000 .005±.024</p>
      <p>S.-W.-M. .820 ±.241 .000±.000 .064±.111 .116±.246
FINANCIAL .807±.091 .024±.076 .000±.001 .169±.046</p>
      <p>In this work we have coupled with a method for inductive
inference on ABoxes with two composite concept
similarity measures. The method is based on the classification in
analogy with the majority of neighbor training individuals.</p>
      <p>It can be naturally exploited for predicting/suggesting missing
of information they convey is very low. A choice of a lower information about individuals thus enabling a sort of inductive
number k of neighbors could probably help committing those retrieval. For example, it may be employed a query to the
residual errors. KB may be issued. Specifically we are targeting also the task</p>
      <p>For the same reasons, also for the SURFACE-W ATER- of service discovery when for semantically annotated web
MODEL ontology quite a high rate of matching classifications service. Even more so the outcomes of our method may be
was reported (yet less then with the previous ontology); decisive for enabling a series of other inductive tasks such as
moreover, some cases of omission error (6%) were observed. clustering, case-based reasoning, , etc..
The induction rate was about 12% which means that for The proposed method is able to induce new assertions, in
this ontology our classifier always assigned individuals to addition to the knowledge already derivable by means of a
the correct concepts but, in some cases, it could also induce reasoner. Then it seems to be able to enhance the standard
new assertions. Since this rate represents assertions that were instance-based classification. Particularly, an increase in
accunot logically deducible from the ontology and yet they were racy was observed when the instances increase in number and
inferred inductively by the analogical classifier, these figures are homogeneously spread.
would be a positive outcome (provided this knowledge were The presented measure can be refined tweaking the
weightdeemed as correct by an expert). Particularly, in this case the ing factor λ for decreasing the impact of the dissimilarity
increase of the induction rate has been due to the presence of between nested sub-concepts in the descriptions on the
deassertions of mutual disjointness for some of the concepts. termination of the overall value.</p>
      <p>
        Results are no different also for the case of the experiments Lately, we have defined another measure for ALN [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ].
with FINANCIAL ontology that largely exceeds the others Hence a natural extension may concern the definition of
in terms of number of concepts and individuals. Namely, dissimilarity measures in more expressive DLs languages. For
the observed match rate is again above the 80% and the example, a normal form for ALCN can be obtained based on
rest of the cases are comprised in the induction rate (17%), those for ALN and ALCN . Then, by exploiting the notion of
leaving a limited margin to residual errors. This corroborates existential mappings [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ], the measure presented in this paper
a fact about the NN learners, that is their reaching better may be extended to more expressive DLs. We are currently
and better performance in the limit, as long as new training developing other kind of semantic similarities based on the
instances become available. Actually, performing a 10-fold idea of the hypothesis-driven distance.
cross validation we obtained almost the same results. Besides, as mentioned, this method could be extended
with different (yet still computationally tractable) answering
C. Experiments with the Information Content Measure procedures grounded on statistical inference (non-parametric
      </p>
      <p>The average results obtained by adopting the procedure with tests based on ranked distances), in order to accept answers
the measure based on information content (see Def. 4) are as correct with a high degree of confidence. Furthermore,
reported in Table II. the k-NN method in its classical form is particularly suitable</p>
      <p>
        By analyzing this table it is possible to note that no sensible for the automated induction of missing values for (scalar or
variation was observed in the classifications performed using numeric) datatype properties of an individual as an estimate
the first dissimilarity measure. Particularly, with both measures derived from the values of the datatypes for the surrounding
the method correctly classified all the individuals, without individuals.
commission errors. The reason is that, in most of the cases, Kernels are another means to express a notion of similarity
the individuals of these ontologies are instances of one concept is some unknown feature space. We are working at the
only and they are involved in a few roles (object properties). definition of kernel functions on DLs representations [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ], thus
Some of the figures are slightly lower than those observed in allowing the exploitation of kernel methods efficiency (e.g. the
the other experiment: this is confirmed by a higher variability. support vector machines) in a multi-relational setting.
      </p>
      <p>Surprisingly, the results on the larger ontologies (S.-W.-M.
and FINANCIAL) perfectly match those obtained with the other ACKNOWLEDGMENTS
measure which is probably due to the fact that we used a leave- This work was partially supported by the regional interest
one-out cross-validation mode which yielded a high value for projects DIPIS (DIstributed Production as Innovative System)
the number k of training instances for the neighborhood and and DDTA (Distretto Digitale Tessile Abbigliamento) in the
context of the tasks related to the semantic web services
discovery.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>W.</given-names>
            <surname>Cohen</surname>
          </string-name>
          and H. Hirsh, “
          <article-title>Learning the CLASSIC description logic</article-title>
          ,”
          <source>in Proceedings of the 4th International Conference on the Principles of Knowledge Representation and Reasoning</source>
          , P. Torasso,
          <string-name>
            <given-names>J.</given-names>
            <surname>Doyle</surname>
          </string-name>
          , and E. Sandewall, Eds. Morgan Kaufmann,
          <year>1994</year>
          , pp.
          <fpage>121</fpage>
          -
          <lpage>133</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>J.-U.</given-names>
            <surname>Kietz</surname>
          </string-name>
          and
          <string-name>
            <given-names>K.</given-names>
            <surname>Morik</surname>
          </string-name>
          , “
          <article-title>A polynomial approach to the constructive induction of structural knowledge,” Machine Learning</article-title>
          , vol.
          <volume>14</volume>
          , no.
          <issue>2</issue>
          , pp.
          <fpage>193</fpage>
          -
          <lpage>218</lpage>
          ,
          <year>1994</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>S.</given-names>
            <surname>Staab</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Maedche</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Nedellec</surname>
          </string-name>
          , and
          <string-name>
            <given-names>P.</given-names>
            <surname>Wiemer-Hastings</surname>
          </string-name>
          , Eds., ECAI2000 Workshop on Ontology Learning, ser.
          <source>CEUR WS Proceedings</source>
          ,
          <year>2000</year>
          , vol.
          <volume>31</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>S.</given-names>
            <surname>Staab</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Maedche</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Nedellec</surname>
          </string-name>
          , and E. Hovy, Eds.,
          <source>IJCAI2001 Workshop on Ontology Learning</source>
          ,
          <year>2001</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>F.</given-names>
            <surname>Esposito</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Fanizzi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Iannone</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Palmisano</surname>
          </string-name>
          , and
          <string-name>
            <surname>G. Semeraro.</surname>
          </string-name>
          , “
          <article-title>Knowledge-intensive induction of terminologies from metadata,” in ISWC2004</article-title>
          ,
          <source>Proceedings of the 3rd International Semantic Web Conference</source>
          , ser. LNCS, F. van
          <string-name>
            <surname>Harmelen</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>McIlraith</surname>
            ,
            <given-names>and D.</given-names>
          </string-name>
          <string-name>
            <surname>Plexousakis</surname>
          </string-name>
          , Eds., vol.
          <volume>3298</volume>
          . Springer,
          <year>2004</year>
          , pp.
          <fpage>441</fpage>
          -
          <lpage>455</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>M. d'Aquin</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Lieber</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Napoli</surname>
          </string-name>
          , “
          <article-title>Decentralized case-based reasoning for the Semantic Web,” in Proceedings of the 4th International Semantic Web Conference, ISWC2005, ser</article-title>
          . LNCS,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Gil</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Motta</surname>
          </string-name>
          , E. Benjamins, and
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Musen</surname>
          </string-name>
          , Eds.,
          <source>no. 3279</source>
          . Springer,
          <year>2005</year>
          , pp.
          <fpage>142</fpage>
          -
          <lpage>155</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>P.</given-names>
            <surname>Hitzler</surname>
          </string-name>
          and D. Vrandec˘ic´, “
          <article-title>Resolution-based approximate reasoning for OWL DL,” in Proceedings of the 4th International Semantic Web Conference, ISWC2005, ser</article-title>
          . LNCS,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Gil</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Motta</surname>
          </string-name>
          , E. Benjamins, and
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Musen</surname>
          </string-name>
          , Eds.,
          <source>no. 3279</source>
          . Springer,
          <year>2005</year>
          , pp.
          <fpage>383</fpage>
          -
          <lpage>397</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>T.</given-names>
            <surname>Mitchell</surname>
          </string-name>
          ,
          <article-title>Machine Learning</article-title>
          .
          <source>McGraw-Hill</source>
          ,
          <year>1997</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>M.</given-names>
            <surname>Rodr</surname>
          </string-name>
          <article-title>´ıguez, “Assessing semantic similarity between spatial entity classes</article-title>
          ,
          <source>” Ph.D. dissertation</source>
          , University of Maine,
          <year>1997</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>A.</given-names>
            <surname>Borgida</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Walsh</surname>
          </string-name>
          , and H. Hirsh, “
          <article-title>Towards measuring similarity in description logics</article-title>
          ,” in Working Notes of the International Description Logics Workshop, ser.
          <source>CEUR Workshop Proceedings</source>
          , Edinburgh, UK,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>M. W.</given-names>
            <surname>Bright</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. R.</given-names>
            <surname>Hurson</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S. H.</given-names>
            <surname>Pakzad</surname>
          </string-name>
          , “
          <source>Automated resolution of semantic heterogeneity in multidatabases.” ACM Transaction on Database Systems</source>
          , vol.
          <volume>19</volume>
          , no.
          <issue>2</issue>
          , pp.
          <fpage>212</fpage>
          -
          <lpage>253</lpage>
          ,
          <year>1994</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>P.</given-names>
            <surname>Resnik</surname>
          </string-name>
          , “
          <article-title>Semantic similarity in a taxonomy: An information-based measure and its application to problems of ambiguity in natural language</article-title>
          ,
          <source>” Journal of Artificial Intelligence Research</source>
          , vol.
          <volume>11</volume>
          , pp.
          <fpage>95</fpage>
          -
          <lpage>130</lpage>
          ,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>C. d'Amato</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Fanizzi</surname>
            , and
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Esposito</surname>
          </string-name>
          , “
          <article-title>A dissimilarity measure for the ALC description logic,” in Semantic Web Applications</article-title>
          and Perspectives, 2nd Italian Semantic Web Workshop SWAP2005, P. Bouquet and G. Tummarello, Eds., vol.
          <volume>166</volume>
          .
          <string-name>
            <surname>Trento</surname>
          </string-name>
          , Italy: CEUR,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14] --, “
          <article-title>A dissimilarity measure for ALC concept descriptions</article-title>
          ,”
          <source>in Proceedings of the 21st Annual ACM Symposium of Applied Computing</source>
          , SAC2006, H. Haddad, Ed. Dijon, France: ACM,
          <year>2006</year>
          , pp.
          <fpage>1695</fpage>
          -
          <lpage>1699</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>F.</given-names>
            <surname>Baader</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Calvanese</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>McGuinness</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Nardi</surname>
          </string-name>
          , and P. PatelSchneider, Eds.,
          <source>The Description Logic Handbook</source>
          . Cambridge University Press,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>F.</given-names>
            <surname>Bacchus</surname>
          </string-name>
          , “
          <article-title>Lp, a logic for representing and reasoning with statistical knowledge,”</article-title>
          <source>Computational Intelligence</source>
          , vol.
          <volume>6</volume>
          , pp.
          <fpage>209</fpage>
          -
          <lpage>231</lpage>
          ,
          <year>1990</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>M.</given-names>
            <surname>Schmidt-Schauß</surname>
          </string-name>
          and G. Smolka, “
          <article-title>Attributive concept descriptions with complements</article-title>
          ,
          <source>” Artificial Intelligence</source>
          , vol.
          <volume>48</volume>
          , no.
          <issue>1</issue>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>26</lpage>
          ,
          <year>1991</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>F. M.</given-names>
            <surname>Donini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lenzerini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Nardi</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Schaerf</surname>
          </string-name>
          , “
          <article-title>Deduction in concept languages: From subsumption to instance checking</article-title>
          ,
          <source>” Journal of Logic and Computation</source>
          , vol.
          <volume>4</volume>
          , no.
          <issue>4</issue>
          , pp.
          <fpage>423</fpage>
          -
          <lpage>452</lpage>
          ,
          <year>1994</year>
          . [Online]. Available: citeseer.ist.psu.edu/donini94deduction.html
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>S.</given-names>
            <surname>Brandt</surname>
          </string-name>
          ,
          <string-name>
            <surname>R.</surname>
          </string-name>
          <article-title>K u¨sters, and</article-title>
          <string-name>
            <surname>A</surname>
          </string-name>
          .-Y. Turhan, “
          <article-title>Approximation and difference in description logics</article-title>
          ,”
          <source>in Proceedings of the International Conference on Knowledge Representation</source>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Fensel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Giunchiglia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>McGuinness</surname>
          </string-name>
          ,
          <string-name>
            <given-names>and M.-A.</given-names>
            <surname>Williams</surname>
          </string-name>
          , Eds. Morgan Kaufmann,
          <year>2002</year>
          , pp.
          <fpage>203</fpage>
          -
          <lpage>214</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>A.</given-names>
            <surname>Tversky</surname>
          </string-name>
          , “
          <article-title>Features od similarity,” Psycological Review</article-title>
          , vol.
          <volume>84</volume>
          , no.
          <issue>4</issue>
          , pp.
          <fpage>327</fpage>
          -
          <lpage>352</lpage>
          ,
          <year>1997</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <surname>C. d'Amato</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Fanizzi</surname>
            , and
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Esposito</surname>
          </string-name>
          , “
          <article-title>A semantic similarity measure for expressive description logics,” in Proceedings of Convegno Italiano di Logica Computazionale, CILC05, A</article-title>
          . Pettorossi, Ed., Rome, Italy,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>H.</given-names>
            <surname>Bock</surname>
          </string-name>
          .,
          <source>Analysis of Symbolic Data: Exploratory Methods for Extracting Statistical Information from Complex Data</source>
          . Springer-Verlag,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>I. H.</given-names>
            <surname>Witten</surname>
          </string-name>
          and
          <string-name>
            <given-names>E.</given-names>
            <surname>Frank</surname>
          </string-name>
          , Data Mining, 2nd ed. Morgan Kaufmann,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>N.</given-names>
            <surname>Fanizzi</surname>
          </string-name>
          and
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>d'Amato, “A similarity measure for the ALN description logic</article-title>
          ,”
          <source>in Proceedings of Convegno Italiano di Logica Computazionale, CILC06</source>
          , Bari, Italy,
          <year>2006</year>
          , http://cilc2006.di.uniba.it/ download/camera/15 Fanizzi CILC06.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>R.</given-names>
            <surname>Kusters</surname>
          </string-name>
          and
          <string-name>
            <given-names>R.</given-names>
            <surname>Molitor</surname>
          </string-name>
          , “
          <article-title>Computing least common subsumers in ALE N ,”</article-title>
          <source>in Proceedings of the International Joint Conference on Artificial Intelligence</source>
          , IJCAI2001, B. Nebel, Ed.,
          <year>2001</year>
          , pp.
          <fpage>219</fpage>
          -
          <lpage>224</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>N.</given-names>
            <surname>Fanizzi</surname>
          </string-name>
          and
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>d'Amato, “A declarative kernel for ALC concept descriptions</article-title>
          .”
          <source>in Proceedings of the 16th International Symposium on Methodologies for Intelligent Systems, ISMIS</source>
          <year>2006</year>
          ,
          <article-title>ser</article-title>
          . LNAI,
          <string-name>
            <given-names>F.</given-names>
            <surname>Esposito</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Ras</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Malerba</surname>
          </string-name>
          , and G. Semeraro, Eds., vol.
          <volume>4203</volume>
          . Springer,
          <year>2006</year>
          , pp.
          <fpage>322</fpage>
          -
          <lpage>331</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>