<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Similarity Measurement about Ontology-based Semantic Web Services</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Yi Zhao Chair of Computer Engineering</string-name>
          <email>yi.zhao@fernuni-hagen.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Fernuniversita ̈t Hagen</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Xia Wang Digital Enterprise Research Institute IDA Business Park</institution>
          ,
          <addr-line>Lower Dangan Galway</addr-line>
          ,
          <country country="IE">Ireland</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Yihong Ding Dept. of Computer Science Brigham Young University Provo</institution>
          ,
          <addr-line>Utah</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Measurement of semantic similarity between Web services is an important factor for Web service discovery, composition, and even execution. Semantic Web services (SWS) are usually specified based on ontologies. The measurement of semantic similarity between Web services thus can be reduced to computing semantic distances between ontologies. In this paper, we briefly surveyed three major existing ontology-distance-computation algorithms and enhanced them to measure the single and multiple ontolgies similarity in SWS context. Based on this survey, we summarized a new hybrid ontology-similarity-measurement methodology that measures similarity between Semantic Web Services.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Service similarity is crucial to service discovery,
selection, composition, and even execution. Especially
semantic service discovery aims to locate the best matched
service, it mostly depends on the measurement of the similarity
between an user’s service requirements and the profiles of
published services. Currently, the semantic Ontology
languages for services, such as the OWL-S1 and the WSMO2,
are required to semantically represent service capabilities,
including non-function information (including Qos of
service), functional information (IOPE of services operations,
denoting input, output, precondition and effect). The
ser1OWL-S, http : ==www:w3:org=Submission=OW L ¡ S=
2WSMO, http : ==www:wsmo:org
vice discovery, therefore, focuses on the matchmaking of
service capability [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] and QoS, while less work is done on
ontology-based services selection.
      </p>
      <p>
        Moreover, Ontology receives great attention in the
progressively emerging Semantic Web [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] and Semantic Web
Services [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], by formally defining the concepts and
relationships in a machine understandable way and enabling
knowledge sharing and reuse. As the elements of the
representation of semantic services, the similarity of the ontologies
used is crucial to service similarity, especially when
considering the discovery and execution.
      </p>
      <p>
        Ontology similarity which is related to Ontology
mapping is a well known topic in information retrieval, database
integration systems, and artificial intelligence fields. Also,
there is a wealth of work on similarity measures of
Ontology concepts and concept-related notions [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ]. However,
the measurement of similarity of ontologies and concepts
itself is not easy, additionally many specific features (see
section 2.2) in Semantic Web Service description
environment.
      </p>
      <p>After surveying the previous Ontology similarity
measures and their application situations, in our SWS context
two methods are combined to adapt to calculate the
semantic distance of single formal Ontology concepts, e.g.
code and zip in the examples of figure 1. The approaches
are a fuzzy-weighted associative network (edge-based
measure) and an information-theoretical approach
(contentbased measure).</p>
      <p>
        Also regarding the compound ontology concepts in
semantic service context, for example, in figure 1 the concepts
findZipCodeDistance and CalcDistTwoZipsKm, are not the
formal single terms as the ones in WordNet4, in this case
the traditional method (e.g. edit distance of strings) is not
useful to measure their semantic similarity. Therefore, we
refine the hierarchical clustering algorithm to calculate the
distance of two compound concept terms, similar to [
        <xref ref-type="bibr" rid="ref4 ref7">7, 4</xref>
        ].
      </p>
      <p>In this paper we aim to solve the ontology similarity in
a semantic service environment. First, we differentiate two
cases of service ontology concepts: single and compound
ontology concept to measure their similarity in service
context. Then, a hybrid Ontology-similarity measurement is
proposed by combining and refining three existing methods.
Finally, we define our ontology similarity-based model as
simS = §simO 2 [0; 1] to improve the service selection;
This model fuzzily and quantitatively measures the service
similarity basing on service ontology similarity.</p>
      <p>This paper is structured as follows. In Section 2 we state
the occurring problems of Ontology in Semantic Web
Service description context, and investigate the specific name
features of ontology concept. A Ontology concept distance
definition and three refined ontology similarity algorithms
are discussed with examples of single formal term and
compound term in Section 3. The service similarity
measurement is defined in Section 4. In Section 5 and 6 we
respectively discuss related works, and give conclusion and
indications for future work.
2
2.1</p>
    </sec>
    <sec id="sec-2">
      <title>Ontology in SWS Description Context</title>
      <sec id="sec-2-1">
        <title>Problem Statements</title>
        <p>In order to illustrate the challenge of measuring the
similarity of semantic Web services, we extract a set of zip code
related services from the dataset of OWL-S annotated Web
Services of the University College Dublin3. In Figure 1.,
there are snatch description of four services, which are used
for looking up a zip code or calculating the distance
between two places according to the given zip codes. The
information shown is retrieved from the wsdl documents of
the respective service.</p>
        <p>Current service matchmaking algorithms normally focus
on measuring the syntactic (as service name, service text
description and so on) and semantic (as service capabilities)
of service. Taking sws4 and sws5 of Figure 1 as examples,
if we assume that zip and code have the similar meaning,
intuitively, by comparing service name and operation name,
service sws4 and sws5 are regarded as similar from the
signature level; and by matching their operation, as both
operation2, they also have similar inputs and outputs, so that sws4
and sws5 are concluded as similar services. This means that
4WordNet, an online lexical reference system, http :
==wordnet:princeton:edu=</p>
        <p>3The Semantic Web Services Repository at the Smart Media Institute
in University College Dublin, http : ==moguntia:ucd:ie=repository=.
both can provide detailed information of a city according to
the given zip code.</p>
        <p>Further, if we assume that a machine can understand
some similarity between fzip; ZipCode; Zip Code 1;
code; code1g and fCaleDisT woZipsKm; f indZipCode
Distanceg, then intuitively and naively from the above
example services of Figure 1, we know that sws1:operation1
is similar to sws5:operation1; sws2:operation1 is similar to
sws3:operation2 and sws4:operatio2; and sws3:operation1
is similar to sws4:operation1.</p>
        <p>Obviously, similarity, whether syntactic or semantic, the
matching of ontology concepts used in service description
is a critical challenge. If a machine can not understand the
meaning of service concepts, it also cannot infer the imply
relationships, then the automatic matching and discovery of
services is impossible.
2.2</p>
      </sec>
      <sec id="sec-2-2">
        <title>Naming Conventions for Ontologies</title>
        <p>Intuitively and in ontology-related work, when
ontology terminology is mentioned, it mostly means the terms
in thesauri, e.g., Wordnet. On the other hand, the ontology
terms defined in applications is very different from the
formal words. Generally, the ontology concept used in service
semantic descriptions are most compound terms, which are
named depending on service developers by their ontology
knowledge, experience and wonted. The situation is made
worse by the following practices (parts of examples from
Fig.1.):
Abbreviations Names are not given in their correct forms,
but shortened, e.g. CalcDistTwoZipsKm;</p>
        <sec id="sec-2-2-1">
          <title>Associated words with capitalization or delimiters</title>
          <p>Words have the form of associations of several words
parts (full word or abbreviation) with delimiters,
normally a part’s first letter capitalized, and sometimes
also using underscore, dash or space, e.g., LogIn,
AcctName, ArrivalAirport In.</p>
        </sec>
        <sec id="sec-2-2-2">
          <title>Words with suffix and prefix Examples are hasFlavour,</title>
          <p>locatedIn.</p>
        </sec>
        <sec id="sec-2-2-3">
          <title>Variations or misspelling Names may be variations of</title>
          <p>word often due to grammatical flexion, e.g.,
Booking, madeFromGrape; And defined words are in
misspelling format for machine.</p>
          <p>Free inventions Any other cases the traditional similarity
measures (based, e.g., on WordNet) are prevented to
work.</p>
          <p>
            Considering the above compound concept terms, the
existing ontology measure algorithms can not work.
Moreover, the data clustering algorithm from data mining field
can be borrowed to apply to this case. This paper
enhances the clustering algorithm in [
            <xref ref-type="bibr" rid="ref4">4</xref>
            ] to measure the
semantic closeness of composed terms.
3
3.1
          </p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Ontology Similarity</title>
      <sec id="sec-3-1">
        <title>Ontology Concept Distance</title>
        <p>To semantically measure Ontology concept distance, we
should consider both concept structure and concept content.
Fortunately, both of these information are prolifically
provided by service description. Here, we define the semantic
distance dis of the assumed concepts C and D (which could
be single formal term or compound term) as:
3
dis = w1 ¤ Diss + w2 ¤ Disi + w3 ¤ Disc; X wi = 1 (1)
i=1
where Diss is the distance basing on the structure of
concept in service Ontology, the Disi basing on the
common contents shared by concepts, and the Disc is only used
to measure the compound concept terms by clustering
concepts, basing on the concept elements co-occurrence.
Formulae 1 not only considers the different concept naming
features, but also make up the loss of any single approach,
because the service description context is just a structure
and a short piece of text, not a corpus or thesaurus.</p>
        <p>In the following sections, we will present the detail
explanation of every distance measurement.
3.2</p>
      </sec>
      <sec id="sec-3-2">
        <title>Fuzzy-weighted Associative Network</title>
        <p>
          Concepts in a hierarchical taxonomy are all related by
certain relationships, based on which concepts can be
represented in an associative network consisting of nodes and
edges, where nodes denote concepts, edge denotes the
binary relationship of the two linked concepts. Also for
service description Ontology, such associative network with
fuzzy-weighted value on each link can be constructed, in
which the similarity of concepts can be measured by the
shortest distance as [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] and [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ], which is defined as Diss
in our context.
        </p>
        <p>
          As the detailed explanation by [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] and
correspondence to OWL-Lite, we define four concept relations as
generalization (e.g., superclass), specification (e.g.,
subclass), negative association (e.g., disjoined) and positive
association (e.g., equivalent).
        </p>
        <p>
          Therefore, the distances of arbitrary two nodes in the
network can be calculated based on Tables 1–3 [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ]. In Table.1
s; g; p and n represent explicit relationships, that is, each
two notes relationship can be evaluated basing on
triangular norms. ¿ in Table 2. are the triangular norms (t-norms),
which is defined in Table3, where ® or ¯ are fuzzy-weighted
strength values of relations (0 · ®; ¯ · 1), n is the
degree of dependence (¡1 · n · 1) between the
relationships, details please refer to [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ]. In the tables those fields
are marked with X for which there is no definition.
Therefore, the relationship of two arbitrary concepts can easily be
inferred by traveling through the associative network.
g
s
p
n
g
g
p
p
n
s
p
s
p
n
p
p
p
p
n
n
n
n
n
X
g
s
p
n
g
¿3
¿1
¿2
¿2
s
¿1
¿3
¿2
¿2
p
¿2
¿2
¿3
¿3
n
¿2
¿2
¿3
X
sim(C; D) = I(Id(ecsocmrimptoino(nC( C;D;D)))) = lolgogPP(d(ecsocmrimptoino(nC( C;D;D))))
(2)
where common(C; D) is a proposition that states the
commonalities between C and D, I(common(C; D)) is the
amount of information contained in this proposition and,
similarly, I(description(C; D)) is a proposition describing
what C and D are. In our service context, we refine the
similarity expression as follows to calculate the distance Disi:
where C and D are two Ontology concept classes of
OWLLite, jC \ Dj is the number of common elements of C and
D, e.g., the number of shared attributes, instances and
relational classes, ° and ± are weight values defining the relative
importance of their non-common characteristics.
        </p>
        <p>Given two single-form Ontology concept terms from
two differen Web service description, as t1 and t2, which
are respectively described by a set of other class terms
as their properties, instances and relational members (e.g.,
“g,s,n,p”). There are two cases:
² Two terms t1 and t2are organized in one hierarchical
structure, which is transformed to a fuzzy weighted
associative network of Section 3.2.</p>
        <p>Reconsidering the example in Fig. 1, it assumes
that in the Zip service application domain, terms
have the relationships (which are all experimental
data, not the real value) cp.Fig. 2. For example,
the distance of term State and Zip is examined, the
shortest path is path = fState; P lace; Code; Zipg
with State =)g;0:9 P lace, P lace =)p;0:9 Code
and Code =)s;0:9 Zip. So that it hold that
¿2(¿2(0:9; 0:9)0:9) = 0:729), following Table 1-3,
that means State =)p;0:729 Zip, finally we get
Diss(t1; t2) = 0:729.
² Terms, t1 and t2, are concept classes, respectively
consisting of a set of properties and instances as ontology
vocabulary according to Section 3.3.</p>
        <p>Assuming that their cardinality are jt1j = 9, jt2j = 6,
and they share the number of elements jt1 \ jt2j = 5,
5
we obtain Disi(t1; t2) = 5+4+1 = 0:5, where °; ± =
0:5.
3.5</p>
      </sec>
      <sec id="sec-3-3">
        <title>Clustering for</title>
      </sec>
      <sec id="sec-3-4">
        <title>Compound</title>
        <p>Clustering is also a well known approach to group data
on the basis of a certain similarity criteria. We adapt this
clustering mechanism here to group the compound
Ontology concept terms, which are from different service
description Ontologies, e.g. findZipCodeDistance and
CalcDistTwoZipsKm, in order to calculate their similarity by
the distance disc.</p>
        <p>
          In the clustering algorithm, the association rule of two
terms t1 and t2 is defined as follows [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]:
        </p>
        <p>t1 ¡! t2(s; c)
where, the support s is the probability s = P (t1) = kTt1 k
kT k
that t1 occurs in T , kT k is the cardinality of the ontology
terms’ domain, kTt1 k is the cardinality of the set which
contains t1, the confidence c is the occurrence probability of t2
in the case that t1 occurred, i.e., c = P (t2jt1) = kTt1;t2 k ,
kTt1 k
with kTt1;t2 k is the cardinality of the set containing both
t1 and t2. The distance of two terms is weighted by their
conditional probability c. The center of a cluster is the term
which has the highest occurrence probability of the cluster.</p>
        <p>In detail, including the natural language term extraction
the clustering algorithm is used by us as follows:
1. Read service description document .owl, move all
OWL-Lite tags, extract names and parameters, and
delete redundancies in the vocabularies. The result is
a bag of unique words including composted concept
terms, denoting T = ft1; t2; :::g.
2. Preprocess all composted terms in T as follows.</p>
        <p>Suppose that ti 2 T is a composite term, we split
it up on the basis of its delimiters, such as capital
letters, into several parts. Then, we deal with each
part towards extracting the word stem by removing
stop words, suffixes and prefixes, restituting
abbreviations or correcting misspelling, deleting redundant
vocabulary terms and so on, resulting in the set ti =
fti1; ti2:::g. Substituting ti by all tij 2 ti for all i,
ultimately yields T 0.
3. Compute the values s and c for any two terms in T 0,
store them into a table in descending order, cluster
them on the basis of their confidence c ¸ ¿c and
support s ¸ ¿s (¿s and ¿c are thresholds either
assigned or obtained experimentally), resulting in the
set T 00 = (X1; X2; :::; Xk) of k clusters.</p>
        <p>
          Roughly speaking, Xi; 1 · i · k is a cluster
including those terms whose co-occurrence probabilities
exceed the threshold ¿c. In traditional agglomeration
clustering algorithms, T 00 is an intermediate result,
while in our context we should improve it in order to
find an optimal clustering for our computation. This is
the rationale of the algorithm’s further steps.
4. In each Xi µ T 00, remove the frequent and rare
parameters to avoid the query expansion and over-fitting
problems, which are discussed in the field of
information retrieval [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ].
5. Split and merge the clusters in T 00, in order to wipe
off the noise terms and optimize clusters by
agglomerating terms according to concentric circularities with
different radii.
        </p>
        <p>
          The inner circularity consists of those terms, which
are, at least, close to half of the other terms. Similarly,
the terms in the outer circularity are, at least, close to
a quarter of the other ones. They are called them 12
radius [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]. And wiping off the terms, which are not in
any circularity.
        </p>
        <p>For example, to merge two cluster X1 and X2, when
8i 2 X1 [ X2,
1
k jjj 2 X1 [ X2; i 6= j; i ¡!)j(c &gt; ¿c1) k¸ 2 (k X1 k + k X2 k ¡1)
(4)</p>
        <p>Now, when calculating the distance between two random
composite terms, here we used c1 and c2 distinctively, first,
preprocess them using step (2) to obtain c1 = fc11; c12; :::g
and c2 = fc21; c22; :::g, and then measure their similarity
disc by the probability of pairs of two terms to occur in the
same cluster. As measure the maximum, minimum or or
mean may be employed. Here we take the maximum as the
optimistic way, the formula is as follows,</p>
        <p>Disc = n0m;ax(sim(t1i; t2j )j8t1i 2 t1; t2j 2 t2); iofthetr1wi;its2ej: 2 Xk
Obviously, such a formula implies as extreme case, that is,
all of the sub-terms of c1 and c2 have been wiped off as the
noise words, such case have no way to scale the distance
of ontology concepts. This part of work is right what our
experiment will analysis, to evaluate the frequency of its
occurrence.
4</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Service Similarity</title>
      <p>
        In our previous work [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ], a semantic service model for
selection is proposed as s = (N F; F; Q; C ). By this model,
the service selection can happen by filtering single property
as Non-functional (in this model, only the service name and
service category and short service text description defined
as non-function) or combined properties as Non-functional,
(5)
functional (basing on logical subsumption computing)
together with qualities of services. Obviously, either
Nonfunction or function-based based selection, the ontology
concept similarity is critical fact for service selection.
      </p>
      <p>Under this selection model, we define an ontology-based
service similarity algorithm. Especially when the
nonfunction properties are considered during service selection,
because the non-functional related service selection is
ontology based.</p>
      <p>The idea is to measure the service similarity by the
similarity of service name, service operations name, which
are defined as Ontology concepts. We do not compare
the whole piece service Ontologies, for example simSO :
(SOi) £ (SOj ) ! [0::1], where SOi is the service
ontology for service si; We only consider how similar two
single ontology concepts are in service ontology context, as
sim(ci; cj ) = ff (ci; cj ) j ci 2 SOi ^ cj 2 SOj g and
the function f (ci; cj ) = mink=1;:::;j dis(ci; ck). Therefore,
our work is different from Ontology mapping.</p>
      <p>The proposed Ontology-based service selection
basically measure by the service name concepts and
operations similarity, called lexical semantic level. It is
defined as simService = simConcepts + simoP eration, where
simConcept is the sum similarity of all the concepts of
services, and simoP eration is the sum similarity of the
operation parameters with their data types.
5</p>
    </sec>
    <sec id="sec-5">
      <title>Related Work</title>
      <p>
        Similarity of ontologies has widely been researched, e.g.,
in the fields of information retrieval, artificial intelligence,
databases, and especially in data mining and web mining.
Many similarity measures are applied, e.g., Bernstein et al.
in [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] use two ways to measure the semantic similarity of
objects in an ontology, which are organized in a
hierarchical ontology structure, viz., the edge-based [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] (a shorter
path from one node to the other) and the node-based [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]
(the notion of shared information content) approach.
Actually, they present five different distance measures of
ontologies, where ontology distance stands for the shortest
path through a common ancestor in a directed acyclic graph.
However, computational degree and weight of edge are not
considered. The vector space approaches computing
cosine or Euclidean distances of k-dimensional vectors [
        <xref ref-type="bibr" rid="ref1 ref13">1, 13</xref>
        ]
do not easily apply to nominal concepts, as it is difficult to
represent them as vectors. The Full-text Retrieval Method
(TF/IDF) is mostly used in information retrieval [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] to
compare documents, which are considered as bags of words.
However, it is inadequate for structure concepts as semantic
relations between them are ignored.
      </p>
      <p>
        The work most closely related to ours are the studies on
ontologies in the semantic web or in semantic web services,
such as [
        <xref ref-type="bibr" rid="ref4 ref7">7, 4</xref>
        ] and [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. While they consider to cluster the
similar terms, and most recur to TF/IDF to measure
concept similarity, we follow Dong’s notion of name clustering
agglomeration algorithms. Maedche et al. also propose an
approach to cluster ontology-based data, using the
hierarchical clustering algorithm to consider instances of concept
similarity. Hau et al. elaborate a metric to measure the
similarity of semantic services annotated with OWL ontologies.
They mainly depend on the information-theoretic approach
to match similar ontology instances. Doan et al. computes
the common information content of ontologies to scale their
similarity. We combine multiple approaches to adapt to
SWS environments. Based on a study of definitions and
features of ontologies expressed in OWL, and from a
computational point of view, we calculate the distance of two
ontologies.
6
      </p>
    </sec>
    <sec id="sec-6">
      <title>Conclusions</title>
      <p>Ontology similarity is unquestionable important for
Semantic Web Service similarity when we consider the
semantic service discovery, selection, composition, and even
execution. This paper tries to propose a ontology
similaritybased approach to measure service similarity and presents
the primary work on it. The contributions of this paper are
summarized as, 1) analysis the ontology similarity problem
in semantic service context, and classify the ontology
concept name features used by service description; 2) present
a hybrid ontology concept distance method, and further to
measure the service similarity.</p>
      <p>As the complexity of ontology-based service similarity,
under our model, there is still a lot left for our future work,
including the set matching of the ontology-based concept
and its type, also the detailed implementation and
evaluation. However, fortunately the preliminary experiments
show that this new methodology works well.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Baeza-Yates</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ribeiro-Neto</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <source>Modern Information Retrieval</source>
          ,
          <string-name>
            <surname>ACM Press</surname>
          </string-name>
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Berners-Lee</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hendler</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Lassila</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <source>The Semantic Web, Scientific American</source>
          ,
          <volume>284</volume>
          (
          <issue>5</issue>
          ):
          <fpage>34</fpage>
          -
          <lpage>43</lpage>
          ,
          <year>2001</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Bernstein</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kaufmann</surname>
          </string-name>
          , E.,
          <string-name>
            <surname>Buerki</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Klein</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , How Similar Is It? Towards Personalized Similarity Measures in Ontologies, International Tagung Wirtschaftsinformatik,
          <year>February 2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Dong</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Alon</surname>
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Halevy</surname>
            , Madhavan,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nemes</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          , and Zhang, J.,
          <source>Simlarity Search for Web Services</source>
          ,
          <string-name>
            <surname>VLDB</surname>
          </string-name>
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Fankhauser</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kracker</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Neuhold</surname>
            ,
            <given-names>E.J.</given-names>
          </string-name>
          ,
          <article-title>Semantic vs. structural resemblance of classes</article-title>
          ,
          <source>ACM SIGMOD RECORD</source>
          , vol(
          <volume>20</volume>
          ),
          <volume>4</volume>
          :
          <fpage>59</fpage>
          -
          <lpage>63</lpage>
          ,
          <year>1991</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Gomez-Perez</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Corcho</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <article-title>Ontology languages for the Semantic Web, Intelligent Systems</article-title>
          , IEEE, Jan/Feb pp.
          <fpage>54</fpage>
          -
          <lpage>60</lpage>
          ,
          <year>2002</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Hau</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Darlington</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>A Semantic</given-names>
            <surname>Similarity Measure for Semantic Web Services</surname>
          </string-name>
          , Conference WWW2005,
          <year>Japan 2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Jones</surname>
            ,
            <given-names>K.S.</given-names>
          </string-name>
          ,
          <article-title>Automatic keyword classification for information retrieval</article-title>
          ,
          <source>Archon Books</source>
          ,
          <year>1971</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Klir</surname>
            ,
            <given-names>G.L.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Folger</surname>
            ,
            <given-names>T.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fuzzy</surname>
            <given-names>Sets</given-names>
          </string-name>
          , Uncertainty and Information, Prentice Hall,
          <year>1988</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>J.H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kim</surname>
            <given-names>H.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>Y.J.</given-names>
          </string-name>
          ,
          <source>Information Retrieval Based on Conceptual Distance in IS-A Hierarchies. J. of Documentation</source>
          ,
          <volume>49</volume>
          :
          <fpage>188</fpage>
          -
          <lpage>207</lpage>
          ,
          <year>1993</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Lin</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>An</surname>
          </string-name>
          Information-Theoretic Definition of Similarity,
          <source>Fifteenth International Conference on Machine Learning</source>
          ,
          <year>1998</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Maedche</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Zacharias</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <article-title>Clustering Ontologybased Metadata in the Semantic Web</article-title>
          ,
          <source>Pro. of the Joint Conferences ECML'02 and PKDD'02</source>
          ,
          <year>2002</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13] Mitchell,
          <string-name>
            <given-names>T. M.</given-names>
            ,
            <surname>Machine</surname>
          </string-name>
          <string-name>
            <given-names>Learning</given-names>
            ,
            <surname>McGraw-Hill</surname>
          </string-name>
          , New York,
          <year>1997</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Resnik</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <article-title>Semantic Similarity in a Taxonomy: An Information-Based Measure and its Application to Problems of Ambiguity in Natural Language</article-title>
          ,
          <source>J. of Artificial Intelligence Research</source>
          ,
          <volume>11</volume>
          :
          <fpage>95</fpage>
          -
          <lpage>130</lpage>
          ,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Paolucci</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kawmura</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Payne</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Sycara</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <source>Semantic Matching of Web Services Capabilities</source>
          ,
          <string-name>
            <surname>ISWC</surname>
          </string-name>
          , Italy,
          <year>2002</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhao</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Wolfgan</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <source>Selection Model of Semantic Web Service, 7th International FLINS Conference on Applied Artificial Intelligence</source>
          ,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Sycara</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Widoff</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Klusch</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Lu</surname>
            ,
            <given-names>J.G.</given-names>
          </string-name>
          ,
          <source>LARKS: Dynamic Matchmaking Among Heterogeneous Software Agents in Cyberspace. Autonomous Agents and Multi-Agent Systems</source>
          ,
          <volume>5</volume>
          :
          <fpage>173</fpage>
          -
          <lpage>203</lpage>
          ,
          <year>2002</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Witten</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Frank</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Data</surname>
            <given-names>Mining:</given-names>
          </string-name>
          <article-title>Practical Machine Learning Tools and Techniques with Java Implementations</article-title>
          , Morgan Kaufmann,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Weinstein</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Birmingham</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <article-title>Comparing concepts in differentiated ontologies</article-title>
          ,
          <source>In Proc. of KAW-99</source>
          ,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>