<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Approximate Measures of Semantic Dissimilarity under Uncertainty</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Nicola Fanizzi</string-name>
          <email>fanizzi@di.uniba.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Claudia d'Amato</string-name>
          <email>claudia.damato@di.uniba.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Floriana Esposito</string-name>
          <email>esposito@di.uniba.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Dipartimento di Informatica Universita` degli studi di Bari Campus Universitario Via Orabona 4</institution>
          ,
          <addr-line>70125 Bari</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>We propose semantic distance measures based on the criterion of approximate discernibility and on evidence combination. In the presence of incomplete knowledge, the distance measures the degree of belief in the discernibility of two individuals by combining estimates of basic probability masses related to a set of discriminating features. We also suggest ways to extend this distance for comparing individuals to concepts and concepts to other concepts.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        In the context of reasoning in the Semantic Web, a growing interest is being
committed to alternative inductive procedures extending the scope of the methods
that can be applied to concept representations. Among them, many are based on
a notion of similarity such as case-based reasoning [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], retrieval [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], conceptual
clustering [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] or ontology matching [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. However this notion is not easily captured
in a definition, especially in the presence of uncertainty.
      </p>
      <p>
        As pointed out in the seminal paper [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] concerning similarity in Description
Logics (DL), most of the existing measures focus on the similarity of atomic
concepts within simple hierarchies. Besides, alternative approaches are based on
related notions of feature similarity or information content. All these approaches
have been specifically aimed at assessing concept similarity.
      </p>
      <p>
        In the perspective of crafting inductive methods for the aforementioned tasks,
the need for a definition of a semantic similarity measure for individuals arises,
that is a problem that so far received less attention in the literature. Some
dissimilarity measures for individuals in specific DL representations have recently
been proposed which turned out to be practically effective for the targeted
inductive tasks [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], however they are still partly based on structural criteria which
determine also their main weakness: they can hardly scale to complex languages.
      </p>
      <p>
        We devised a new family of dissimilarity measures for semantically annotated
resources, which can overcome the aforementioned limitations [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. Our measures
are mainly based on Minkowski’s measures for Euclidean spaces [18] induced by
means of a method developed in the context of relational machine learning [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ].
We extend the idea a notion of discernibility borrowed from rough sets theory [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]
which aims at the formal definition of vague sets (concepts) by means of their
approximations. In this paper, we propose (semi-)distance measures based on
semantic discernibility and on evidence combination [
        <xref ref-type="bibr" rid="ref5">16, 5, 15</xref>
        ].
      </p>
      <p>
        Namely, the measures are based on the degree of discernibility of the input
individuals with respect to a committee of features, which are represented by
concept descriptions expressed in the concept language of choice. One of the
advantages of these measures is that they do not rely on a particular language for
semantic annotations. However, these new measures are not to be regarded as
absolute, since they depend both on the choice (and cardinality) of the features
committee and on the knowledge base they are applied to. These measures can
easily be computed based on statistics on individuals that are likely to be
maintained by knowledge base management systems designed for storing instances
(e.g. [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]), which can determine a potential speed-up in the measure computation
during knowledge-intensive tasks.
      </p>
      <p>
        Furthermore, we also propose a way to extend the presented measures to
the case of assessing concept similarity by means of the notion of medoid [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ],
i.e., in a categorical context, the most centrally located individual in a concept
extension w.r.t. a given metric.
      </p>
      <p>The remainder of the paper is organized as follows. In the next section, we
recall the basics of approximate semantic distance measures for individuals in
a DL knowledge base. Hence, we extend the measures with a more principled
treatment of uncertainty based on evidence combination. Conclusions discuss
the applicability of these measures in further works,
2</p>
    </sec>
    <sec id="sec-2">
      <title>Semantic Distance Measures</title>
      <p>
        Since our method is not intended for a particular representation, in the following
we assume that resources, concepts and their relationships may be defined in
terms of a generic representation that may be mapped to some DL language
with the standard model-theoretic semantics (see the handbook [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] for a thorough
reference).
      </p>
      <p>In this context, a knowledge base K = hT , Ai contains a TBox T and an
ABox A. T is a set of concept definitions. A contains assertions concerning the
world state. The set of the individuals (resources) occurring in A will be denoted
with Ind(A). Each individual can be assumed to be identified by its own URI.
Sometimes, it could be useful to make the unique names assumption on such
individuals.</p>
      <p>As regards the inference services, our procedure requires performing
instancechecking and the related service of retrieval, which will be used for the
approximations.
2.1</p>
      <sec id="sec-2-1">
        <title>A Simple Semantic Metric for Individuals</title>
        <p>
          Aiming at distance-based tasks, such as clustering or similarity search, we have
developed a new measure with a definition that totally depends on semantic
aspects of the individuals in the knowledge base [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ], following ideas borrowed
from metric learning in clausal spaces [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ].
        </p>
        <p>Indeed, for our purposes, we needed functions to measure the (dis)similarity
of individuals. However individuals do not have a syntactic (or algebraic)
structure that can be compared. Then the underlying idea is that. on a semantic level,
similar individuals should behave similarly with respect to the same concepts. A
way for assessing the similarity of individuals in a knowledge base can be based
on the comparison of their semantics along a number of dimensions represented
by a set of concept descriptions (henceforth referred to as the committee).
Particularly, the rationale of the measure is to compare them on the grounds of their
behavior w.r.t. a given set of concept descriptions, say F = {F1, F2, . . . , Fk},
which stands as a group of discriminating features expressed in the language
taken into account.</p>
        <p>We begin with defining the behavior of an individual w.r.t. a certain concept
in terms of projecting it in this dimension:
Definition 2.1 (projection function). Given a concept Fi ∈ F, the related
projection function</p>
        <p>πi : Ind(A) 7→ {0, 1/2, 1}
is defined:
∀a ∈ Ind(A)
πi(a) :=
 1
 0
 1/2</p>
        <p>K |= Fi(a)
K |= ¬Fi(a)
otherwise
The case of πi(a) = 1/2 corresponds to the case when a reasoner cannot give
the truth value for a certain membership query. This is due to the Open World
Assumption normally made in this context. Hence, as in the classic probabilistic
models uncertainty is coped with by considering a uniform distribution over the
possible cases.</p>
        <p>Now the discernibility functions related to the committee concepts which
compare the two input individuals w.r.t. these concepts through their
projections:</p>
        <sec id="sec-2-1-1">
          <title>Definition 2.2 (discernibility function). Given a feature concept Fi ∈ F,</title>
          <p>the related discernibility function
is defined as follows:
∀(a, b) ∈ Ind(A) × Ind(A)</p>
          <p>
            Finally, a whole family of distance functions for individuals inspired to Minkowski’s
distances Lp [18] can be defined as follows [
            <xref ref-type="bibr" rid="ref8">8</xref>
            ]:
Definition 2.3 (dissimilarity measures). Let K = hT , Ai be a knowledge
base. Given a set of concept descriptions F = {F1, F2, . . . , Fk}, a family of
dissimilarity measures {dpF}p∈IN, contains functions
          </p>
          <p>
            dpF : Ind(A) × Ind(A) 7→ [
            <xref ref-type="bibr" rid="ref1">0, 1</xref>
            ]
defined
∀(a, b) ∈ Ind(A) × Ind(A):
dpF(a, b) :=
          </p>
          <p>Lp(πi(a), πi(b))
|F|
= k1 uuvtp Xk δi(a, b)p
i=1</p>
          <p>Note that k depends on F and the effect of the factor 1/k is just to normalize
the norms w.r.t. the number of features that are involved. It is worthwhile to
recall that these measures are not absolute, then they should be also be considered
w.r.t. the committee of choice, hence comparisons across different committees
may not be meaningful. Larger committees are likely to decrease the measures
because of the normalizing factor yet these values is affected also by the degree
of redundancy of the features employed.
2.2</p>
        </sec>
      </sec>
      <sec id="sec-2-2">
        <title>Example</title>
        <p>Let us consider a knowledge base in a DL language made up of a TBox:
T = { Female ≡ ¬Male,</p>
        <p>Parent ≡ ∀child.Being u ∃child.Being,
Father ≡ Male u Parent,</p>
        <p>
          FatherWithoutSons ≡ Father u ∀child.Female
}
and of an ABox:
A = { Being(ZEUS), Being(APOLLO), Being(HERCULES), Being(HERA),
Male(ZEUS), Male(APOLLO), Male(HERCULES),
Parent(ZEUS), Parent(APOLLO), ¬Father(HERA),
God(ZEUS), God(APOLLO), God(HERA), ¬God(HERCULES),
hasChild(ZEUS, APOLLO), hasChild(HERA, APOLLO),
hasChild(ZEUS, HERCULES),
}
Suppose F = {F1, F2, F3, F4} = {Male, God, Parent, FatherWithoutSons}. Let us
compute the distances (with p = 1):
d1F(ZEUS, HERA) = (|1 − 0| + |1 − 1| + |1 − 1| + |0 − 0|) /4 = 1/4
d1F(HERA, APOLLO) = (|0 − 1| + |1 − 1| + |1 − 1| + |0 − 1/2|) /4 = 3/8
d1F(APOLLO, HERCULES) = (|1 − 1| + |1 − 0| + |1 − 1/2| + |1/2 − 1/2|) /4 = 3/8
d1F(HERCULES, ZEUS) = (|1 − 1| + |0 − 1| + |1/2 − 1| + |1/2 − 0|) /4 = 1/2
d1F(HERA, HERCULES) = (|0 − 1| + |1 − 0| + |1 − 1/2| + |0 − 1/2|) /4 = 3/4
d1F(APOLLO, ZEUS) = (|1 − 1| + |1 − 1| + |1 − 1| + |1/2 − 0|) /4 = 1/8
2.3
It is easy to prove that these dissimilarity functions have the standard properties
for semi-distances [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ]:
        </p>
      </sec>
      <sec id="sec-2-3">
        <title>Proposition 2.1 (semi-distance). For a fixed feature set F and p &gt; 0, given</title>
        <p>any three instances a, b, c ∈ Ind(A). it holds that:
1. dpF(a, b) ≥ 0 and dpF(a, b) = 0 if a = b
2. dpF(a, b) = dpF(b, a)
3. dpF(a, c) ≤ dpF(a, b) + dpF(b, c)</p>
        <p>This measure is not a distance since it does not hold that a = b if dpF(a, b) = 0.
This is the case of indiscernible individuals with respect to the given committee
of features F. However, if the unique names assumption were made then one may
define a supplementary dimension for the committee (a sort of meta-feature F0)
based on equality, such that:
∀(a, b) ∈ Ind(A) × Ind(A)
and then
δ0(a, b) :=
dpF(a, b) := k +1 1 tuvup Xk δi(a, b)p
i=0
The resulting measures are distance measures.</p>
        <p>
          Compared to other proposed dissimilarity measures [
          <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
          ], the presented
functions do not depend on the constructors of a specific language, rather they
require only (retrieval or) instance-checking for computing the projections through
class-membership queries to the knowledge base.
        </p>
        <p>
          The complexity of measuring he dissimilarity of two individuals depends on
the complexity of such inferences (see [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ], Ch. 3). Note also that the
projections that determine the measure can be computed (or derived from statistics
maintained on the knowledge base) before the actual distance application, thus
determining a speed-up in the computation of the measure. This is very
important for algorithms that massively use this distance, such as all instance-based
methods.
        </p>
        <p>
          So far we made the assumption that F may represent a sufficient number
of (possibly redundant) features that are able to discriminate really different
individuals. The choice of the concepts to be included – feature selection – may
be crucial. Therefore, we have devised specific optimization algorithms founded
in randomized search which are able to find optimal choices of discriminating
concept committees [
          <xref ref-type="bibr" rid="ref7 ref8">8, 7</xref>
          ].
        </p>
        <p>
          The fitness function to be optimized is based on the discernibility factor [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]
of the committee. Given the whole set of individuals Ind(A) (or just a
holdout sample to be used to induce an optimal measure) HS ⊆ Ind(A) the fitness
function to be maximized is:
discernibility(F, HS ) :=
        </p>
        <p>X
k</p>
        <p>X δi(a, b)
(a,b)∈HS2 i=1</p>
        <p>
          However, the results obtained so far with knowledge bases drawn from
ontology libraries [
          <xref ref-type="bibr" rid="ref7 ref9">7, 9</xref>
          ] show that (a selection) of the (primitive and defined) concepts
is often sufficient to induce satisfactory dissimilarity measures.
3
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Dissimilarity Measures Based on Uncertainty</title>
      <p>The measure defined in the previous section deals with uncertainty in a uniform
way: in particular, the degree of discernibility of two individuals is null when
they have the same behavior w.r.t. the same feature, even in the presence of
total uncertainty of class-membership for both. When uncertainty regards only
one projection, then they are considered partially (possibly) similar.</p>
      <p>
        We would like to make this uncertainty more explicit1. One way to deal with
uncertainty would be considering intervals rather than numbers in [
        <xref ref-type="bibr" rid="ref1">0,1</xref>
        ] as a
measure of dissimilarity. This is similar to the case of imprecise probabilities [17].
      </p>
      <p>
        In order to extend the measure, we propose an epistemic definition based on
rules for combining evidence [
        <xref ref-type="bibr" rid="ref5">5, 15</xref>
        ]. The ultimate aim is to assess the distance
between two individuals as a combination of the evidence that they differ based
on some selected features (as in the previous section).
      </p>
      <p>The distance measure that is to be defined is again based on the degree of
belief of discernibility of individuals w.r.t. such features. To this purpose the
probability masses of the basic events (class-membership) have to be assessed.
However, in this case we will not treat uncertainty in the classic probabilistic
way (uniform probability). Rather, we intend to take into account uncertainty
in the computation.</p>
      <p>The new dissimilarity measure will be derived as a combination of the degree
of belief in the discernibility of the individuals w.r.t. each single feature. Before
introducing the combination rule (that will have the measure as a specialization),
the basic probability assignments have to be considered, especially for the cases
when instance-checking is not able to provide a certain answer.</p>
      <p>
        As in previous works [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], we may estimate the concept extensions recurring to
their retrieval [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], i.e. the individuals of the ABox that can be proved to belong
to a concept. Thus, in case of uncertainty, the basic probabilities masses for each
feature concept, can be approximated2 in the following way:
1 We are referring to a notion of epistemic (rather than aleatory) probability [15],
which seems more suitable for our purposes. See Shafer’s introductory chapter in [16]
on this distinction.
2 In case of a certain answer received from the reasoner, the probability mass amounts
to 0 or 1.
∀i ∈ {1, . . . , k}
      </p>
      <p>
        mi(K |= Fi(a)) ≈ |retrieval(Fi, K)|/|Ind(A)|
mi(K |= ¬Fi(a)) ≈ |retrieval(¬Fi, K)|/|Ind(A)|
mi(K |= Fi(a) ∨ K |= ¬Fi(a)) ≈ 1 − mi(K |= Fi(a)) − mi(K |= ¬Fi(a))
where the retrieval(·, ·) operator returns the individuals which can be proven to
be members of the argument concept in the context of the current knowledge
base [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. The rationale is that the larger the (estimated) extension the more
likely is for individuals to belong to the concept. These approximated probability
masses become more precise as more information is acquired. Alternatively, these
masses could come with the ontology as a supplementary for of prior knowledge.
      </p>
      <p>As in the previous section, we define a discernibility function related to a
fixed concept which measures the amount of evidence that two input individuals
may be separated by that concept:
Definition 3.1 (discernibility function). Given a feature concept Fi ∈ F,
the related discernibility function
The extreme values {0, 1} are returned when the answers from the
instancechecking service are certain for both individuals. If the first individual is an
instance of the i-th feature (resp., its complement) then the discernibility
depends on the belief of class-membership to the complement concept of the other
individual. Otherwise, if there is uncertainty for the former individual but not
for the latter, the function changes its perspective, swapping the roles of the
two individuals. Finally, in case there were uncertainty for both individuals, the
discernibility is computed as the chance that they may belong one to the feature
concept and one to its complement,</p>
      <p>
        The combined degree of belief in the case of discernible individuals, assessed
using the mixing combination rule [
        <xref ref-type="bibr" rid="ref12">12, 15</xref>
        ], can give a measure of the semantic
distance between them.
      </p>
      <sec id="sec-3-1">
        <title>Definition 3.2 (weighted average measure). Given an ABox A, a dissim</title>
        <p>ilarity measure for the individuals in A</p>
        <p>
          F
davg : Ind(A) × Ind(A) 7→ [
          <xref ref-type="bibr" rid="ref1">0, 1</xref>
          ]
dFavg(a, b) :=
k
X wiδi(a, b)
i=1
wi =
ui
u
The choices for the weights are various. The most straightforward one is, of
course, considering uniform weights: wi = 1/k. Another one is
where
ui =
        </p>
        <p>1
|Ind(A) \ retrieval(Fk, K)|
and u =
k
X ui
i=1</p>
        <p>It is easy to see that this can be considered as a generalization of the measure
defined in the previous section (for p = 1).
3.1</p>
        <sec id="sec-3-1-1">
          <title>Discussion</title>
          <p>It can be proved that function has the standard properties for semi-distances:
Proposition 3.1 (semi-distance). For a fixed choice of weights {wi}ik=1,
func</p>
          <p>F
tion davg is a semi-distance.</p>
          <p>The underlying idea for the measure is to combine the belief of the
dissimilarity of the two input individuals provided by several sources, that are related
to the feature concepts. In the original framework for evidence composition the
various sources are supposed to be independent, which is generally unlikely to
hold. Yet, from a practical viewpoint, overlooking this property for the sake of
simplicity may still lead to effective methods, as the Na¨ıve Bayes approach in
Machine Learning demonstrates.</p>
          <p>It could also be criticized that the subsumption hierarchy has not been
explicitly involved. However, this may be actually yielded as a side-effect of the
possible partial redundancy of the various concepts, which has an impact on
their extensions and thus on the related projection function. A tradeoff is to be
made between the number of features employed and the computational effort
required for computing the related projection functions.</p>
          <p>The discriminating power of each feature concept can be weighted in terms
of information and entropy measures. Namely, the degree of information yielded
by each of these features can be estimated as follows:</p>
          <p>Hi(a, b) = − X mi(A) log(mi(A))</p>
          <p>A⊆Θ
where 2Θ, w.r.t. the frame of discernment 3 [16, 15] Θ = {D, D}. then, the sum</p>
          <p>X
(a,b)∈HS</p>
          <p>Hi(a, b)
provides a measure of the utility of the discernibility function related to each
feature which can be used in randomized optimization algorithms.
3.2</p>
        </sec>
        <sec id="sec-3-1-2">
          <title>Extensions</title>
          <p>
            Following the rationale of the average link criterion used in agglomerative
clustering [
            <xref ref-type="bibr" rid="ref11">11</xref>
            ], the measures can be extended to the case of concepts, by recurring
to the notion of medoids.
          </p>
          <p>The medoid of a group of individuals is the individual that has the highest
similarity w.r.t. the others. Formally. given a group G = {a1, a2, . . . , an}, the
medoid is defined:
n
medoid(G) = argmin X d(a, aj )</p>
          <p>a∈G j=1</p>
          <p>Now, given two concepts C1, C2, we can consider the two corresponding
groups of individuals obtained by retrieval Ri = {a ∈ Ind(A) | K |= Ci(a)},
and their resp. medoids mi = medoid(Ri) for i = 1, 2 w.r.t. a given measure
dpF (for some p &gt; 0 and committee F). Then the function for concepts can be
defined as follows:</p>
          <p>dpF(C1, C2) := dpF(m1, m2)</p>
          <p>Similarly, if the distance of an individual a to a concept C has to be assessed,
one could consider the nearest (resp. farthest) individual in the concept extension
or its medoid. Let m = medoid(retrieval(C)) w.r.t. a given measure dpF. Then the
measure for this case can be defined as follows:</p>
          <p>dpF(a, C) := dpF(a, m)</p>
          <p>Of course these approximate measures become more and more precise as the
knowledge base is populated with an increasing number of individuals.
4</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Concluding Remarks</title>
      <p>
        We have proposed the definition of dissimilarity measures over spaces of
individuals in a knowledge base. The measures are not language-dependent, differently
from other previous proposals [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], yet they are parameterized on a committee of
concepts. Optimal committees can be found via randomized search methods [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
3 Here D stands for the case of discernible individuals w.r.t. Fi, D for the opposite case,
and some probability mass may be assigned also to the uncertain case represented
by {D, D}.
      </p>
      <p>Besides, we have extended the measures to cope with cases of uncertainty by
means of a simple evidence combination method.</p>
      <p>
        One of the advantages of the measures is that their computation can be
very efficient in cases when statistics (on class-membership) are maintained by
the KBMS [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. As previously mentioned, the subsumption relationships among
concepts in the committee is not explicitly exploited in the measure for making
the relative distances more accurate. The extension to the case of concept
distance may also be improved. Hence, scalability should be guaranteed as far as a
good committee has been found and does not change also because of the
locality properties observed for instances in several domains (e.g. social or biological
networks).
      </p>
      <p>A refinement of the committee may become necessary only when a
degradation of the discernibility factor is detected due to the availability of somewhat
new individuals. This may involve further tasks such as novelty or concept drift
detection.
4.1</p>
      <sec id="sec-4-1">
        <title>Applications</title>
        <p>
          The measures have been integrated in an instance-based learning system
implementing a nearest-neighbor learning algorithm: an experimentation on
performing semantic-based retrieval proved the effectiveness of the new measures,
compared to the outcomes obtained adopting other measures [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ]. It is worthwhile
to mention that results where not particularly affected by feature selection: often
using the very concepts defined in the knowledge base provides good committees
which are able to discern among the different individuals [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ].
        </p>
        <p>
          We are also exploiting the implementation of these measures for performing
conceptual clustering [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ], where (a hierarchy of) clusters is created by grouping
instances on the grounds of their similarity, possibly triggering the induction of
new emerging concepts [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ].
4.2
        </p>
      </sec>
      <sec id="sec-4-2">
        <title>Extensions</title>
        <p>
          The measure may have a wide range of application of distance-based methods to
knowledge bases. For example, logic approaches to ontology matching [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] may be
backed up by the usage of our measures, especially when concepts to be matched
across different terminologies are known to share a common set of individuals.
Ontology matching could be a phase in a larger process aimed at data integration.
Moreover metrics could also support a process of (semi-)automated classification
of new data also as a first step towards ontology evolution.
        </p>
        <p>Another problem that could be tackled by means of dissimilarity measures
could be the ranking of the answers provided by a matchmaking algorithm based
on the similarity between the concept representing the query and the retrieved
individuals.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgments</title>
      <p>The authors would like to thank the anonymous reviewers for their suggestions.
[15] K. Sentz and S. Ferson. Combination of evidence in Dempster-Shafer theory.</p>
      <p>Technical Report SAND2002-0835, SANDIATech, April 2002.
[16] G. Shafer. A Mathematical Theory of Evidence. Princeton University Press, 1976.
[17] P. Walley. Statistical reasoning with imprecise probabilities. Chapman and Hall,</p>
      <p>London, 1991.
[18] P. Zezula, G. Amato, V. Dohnal, and M. Batko. Similarity Search – The Metric
Space Approach. Advances in Database Systems. Springer, 2007.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>F.</given-names>
            <surname>Baader</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Calvanese</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>McGuinness</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Nardi</surname>
          </string-name>
          , and P. Patel-Schneider, editors.
          <source>The Description Logic Handbook</source>
          . Cambridge University Press,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A.</given-names>
            <surname>Borgida</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.J.</given-names>
            <surname>Walsh</surname>
          </string-name>
          , and
          <string-name>
            <given-names>H.</given-names>
            <surname>Hirsh</surname>
          </string-name>
          .
          <article-title>Towards measuring similarity in description logics</article-title>
          . In I. Horrocks,
          <string-name>
            <given-names>U.</given-names>
            <surname>Sattler</surname>
          </string-name>
          , and F. Wolter, editors,
          <source>Working Notes of the International Description Logics Workshop</source>
          , volume
          <volume>147</volume>
          <source>of CEUR Workshop Proceedings</source>
          , Edinburgh, UK,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>C. d'Amato</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Fanizzi</surname>
            , and
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Esposito</surname>
          </string-name>
          .
          <article-title>Reasoning by analogy in description logics through instance-based learning</article-title>
          . In G. Tummarello,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bouquet</surname>
          </string-name>
          , and O. Signore, editors,
          <source>Proceedings of Semantic Web Applications and Perspectives, 3rd Italian Semantic Web Workshop, SWAP2006</source>
          , volume
          <volume>201</volume>
          <source>of CEUR Workshop Proceedings</source>
          , Pisa, Italy,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>M. d'Aquin</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Lieber</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Napoli</surname>
          </string-name>
          .
          <article-title>Decentralized case-based reasoning for the Semantic Web</article-title>
          . In Y. Gil,
          <string-name>
            <given-names>E.</given-names>
            <surname>Motta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Benjamins</surname>
          </string-name>
          , and M. A. Musen, editors,
          <source>Proceedings of the 4th International Semantic Web Conference, ISWC2005, number 3279 in LNCS</source>
          , pages
          <fpage>142</fpage>
          -
          <lpage>155</lpage>
          . Springer,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>D.</given-names>
            <surname>Dubois</surname>
          </string-name>
          and
          <string-name>
            <given-names>H.</given-names>
            <surname>Prade</surname>
          </string-name>
          .
          <article-title>On the combination of evidence in various mathematical frameworks</article-title>
          . In J. Flamm and T. Luisi, editors,
          <source>Reliability Data Collection and Analysis</source>
          , pages
          <fpage>213</fpage>
          -
          <lpage>241</lpage>
          .
          <year>1992</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>J.</given-names>
            <surname>Euzenat</surname>
          </string-name>
          and
          <string-name>
            <given-names>P.</given-names>
            <surname>Shvaiko</surname>
          </string-name>
          . Ontology matching. Springer,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>N.</given-names>
            <surname>Fanizzi</surname>
          </string-name>
          , C. d'Amato, and
          <string-name>
            <given-names>F.</given-names>
            <surname>Esposito</surname>
          </string-name>
          .
          <article-title>Evolutionary conceptual clustering of semantically annotated resources</article-title>
          .
          <source>In Proceedings of the 1st International Conference on Semantic Computing, IEEE-ICSC2007</source>
          , pages
          <fpage>783</fpage>
          -
          <lpage>790</lpage>
          , Irvine, CA,
          <year>2007</year>
          . IEEE Computer Society Press.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>N.</given-names>
            <surname>Fanizzi</surname>
          </string-name>
          , C. d'Amato, and
          <string-name>
            <given-names>F.</given-names>
            <surname>Esposito</surname>
          </string-name>
          .
          <article-title>Induction of optimal semi-distances for individuals based on feature sets</article-title>
          .
          <source>In Working Notes of the International Description Logics Workshop, DL2007</source>
          , volume
          <volume>250</volume>
          <source>of CEUR Workshop Proceedings</source>
          , Bressanone, Italy,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>N.</given-names>
            <surname>Fanizzi</surname>
          </string-name>
          , C. d'Amato, and
          <string-name>
            <given-names>F.</given-names>
            <surname>Esposito</surname>
          </string-name>
          .
          <article-title>Instance-based query answering with semantic knowledge bases</article-title>
          . In R. Basili and M.T. Pazienza, editors,
          <source>Proceedings of the 10th Congress of the Italian Association for Artificial Intelligence</source>
          ,
          <source>AI*IA2007</source>
          , volume
          <volume>4733</volume>
          <source>of LNAI</source>
          , pages
          <fpage>254</fpage>
          -
          <lpage>265</lpage>
          , Rome, Italy,
          <year>2007</year>
          . Springer.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>I. R.</given-names>
            <surname>Horrocks</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Turi</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S. K.</given-names>
            <surname>Bechhofer</surname>
          </string-name>
          .
          <article-title>The Instance Store: DL reasoning with large numbers of individuals</article-title>
          . In V. Haarslev and
          <string-name>
            <surname>R.</surname>
          </string-name>
          <article-title>M¨oller</article-title>
          , editors,
          <source>Proceedings of the 2004 Description Logic Workshop</source>
          ,
          <string-name>
            <surname>DL</surname>
          </string-name>
          <year>2004</year>
          , volume
          <volume>104</volume>
          <source>of CEUR Workshop Proceedings</source>
          , pages
          <fpage>31</fpage>
          -
          <lpage>40</lpage>
          . CEUR,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>A.K. Jain</surname>
            ,
            <given-names>M.N.</given-names>
          </string-name>
          <string-name>
            <surname>Murty</surname>
            , and
            <given-names>P.J.</given-names>
          </string-name>
          <string-name>
            <surname>Flynn</surname>
          </string-name>
          .
          <article-title>Data clustering: A review</article-title>
          .
          <source>ACM Computing Surveys</source>
          ,
          <volume>31</volume>
          (
          <issue>3</issue>
          ):
          <fpage>264</fpage>
          -
          <lpage>323</lpage>
          ,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>C. K.</given-names>
            <surname>Murphy</surname>
          </string-name>
          .
          <article-title>Combining belief functions when evidence conflicts</article-title>
          .
          <source>Decision Support Systems</source>
          ,
          <volume>29</volume>
          (
          <issue>1-9</issue>
          ),
          <year>2000</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Pawlak</surname>
          </string-name>
          .
          <source>Rough Sets: Theoretical Aspects of Reasoning About Data</source>
          . Kluwer Academic Publishers,
          <year>1991</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>M.</given-names>
            <surname>Sebag</surname>
          </string-name>
          .
          <article-title>Distance induction in first order logic</article-title>
          . In S. Dˇzeroski and N. Lavraˇc, editors,
          <source>Proceedings of the 7th International Workshop on Inductive Logic Programming, ILP97</source>
          , volume
          <volume>1297</volume>
          <source>of LNAI</source>
          , pages
          <fpage>264</fpage>
          -
          <lpage>272</lpage>
          . Springer,
          <year>1997</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>