<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Predicting the quality of semantic relations by applying Machine Learning Classifiers to the Semantic Web</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Miriam Fernandez</string-name>
          <email>m.fernandez@open.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Marta Sabou</string-name>
          <email>r.m.sabou@open.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Petr Knoth</string-name>
          <email>p.knoth@open.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Enrico Motta</string-name>
          <email>e.motta@open.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>General Terms Algorithms</institution>
          ,
          <addr-line>Measurement, Design, Experimentation</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Knowledge Media Institute (KMi) The Open University Walton Hall, Milton Keynes</institution>
          ,
          <addr-line>Mk7 6AA</addr-line>
          ,
          <country country="UK">United Kingdom</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>In this paper, we propose the application of Machine Learning (ML) methods to the Semantic Web (SW) as a mechanism to predict the correctness of semantic relations. For this purpose, we have acquired a learning dataset from the SW and we have performed an extensive experimental evaluation covering more than 1,800 relations of various types. We have obtained encouraging results, reaching a maximum of 74.2% of correctly classified semantic relations for classifiers able to validate the correctness of multiple types of semantic relations (generic classifiers) and up to 98% for classifiers focused on evaluating the correctness of one particular semantic relation (specialized classifiers).</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Semantic Web</kwd>
        <kwd>Semantic Relations</kwd>
        <kwd>Machine Learning</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. INTRODUCTION</title>
      <p>
        The problem of relation extraction between two terms is a
wellknown research problem traditionally addressed by the Natural
Language Processing (NLP) community. The approaches found in
the literature follow several different trends like: the exploitation
of lexical patters to extract relations from textual corpora [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], the
generation of statistical measures that detect correlations between
words based on their frequency within documents [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] or, the
exploitation of structured knowledge resources like WordNet1 to
detect or refine relations [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
      </p>
      <p>With the evolution of the SW notion of knowledge reuse, from an
ontology-centered view, to a more fine-grained perspective where
individual knowledge statements (i.e., semantic relations) are
reused rather than entire ontologies, a parallel problem arises:
estimating the correctness of a known relation between two terms.
As an illustrative example, imagine the two following relations:
Book – containsChapter –Chapter, Chapter ⊆ Book. While the
relation Book – containsChapter –Chapter can be considered
correct independently of an interpretation context, in the case of
Chapter ⊆ Book, subsumption has been used incorrectly to model
a meronymy relation.</p>
      <p>
        One of the first attempts to address this problem is the work of
Sabou et al. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. In this wok the authors investigate the use of the
Semantic Web (SW) as a source of evidence for predicting the
correctness of a semantic relation. They show that the SW is not
just a motivation to investigate the problem, but a large collection
of knowledge-rich results that can be exploited to address it.
Following this idea, the work presented in this paper makes use of the
SW as a source of evidence for predicting the correctness of
semantic relations. However, as opposed to [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], which introduces
several evaluation measures based on the adaptation of existing
Natural Language methodologies to SW data, this work aims to
approach the problem using Machine Learning (ML) techniques.
For this purpose, we have worked on: a) acquiring a
mediumscale learning dataset from the SW and b) performing an
experimental evaluation covering more than 1,800 relations of various
types. We have obtained encouraging results, reaching a
maximum of 74.2% of correctly classified semantic relations for
classifiers able to validate the correctness of multiple types of
semantic relations (generic classifiers) and up to 98% for classifiers
focused on evaluating the correctness of one particular semantic
relation (specialized classifiers).
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. ACQUIRING A LEARNING DATASET</title>
      <p>
        The problem addressed in this work can be formalized as a
classification task. In this type of Machine Learning problems, the
learning method is presented with a set of classified examples
from which it is expected to learn how to predict the classification
of unseen examples. The collection of classified examples, or the
learning dataset, is obtained in three phases. In the first phase, a
set of manually evaluated semantic relations is acquired. These
relations can be seen as a quadruple &lt;s, R, t, e&gt; where s is the
source term, t is the target term, R is the relation to be evaluated,
and e {T, F} is a manual Boolean evaluation provided by users
where T denotes a true or correct relation, and F denotes a false or
incorrect relation; e.g., &lt;Helicopter, ⊆ , Aircraft, T&gt;. This
experimental data is obtained from the datasets of the Ontology
Alignment Evaluation Initiative2 (OAEI) and includes the
AGROVOC/NALT and the OAEI'08 datasets. These datasets
comprise a total of 1,805 semantic relations of different types: ⊆,
⊇, ⊥ and named. Among them, 1,129 are evaluated as true (T),
1 http://wordnet.princeton.edu/
2 http://oaei.ontologymatching.org/
correct relations, and 676 are evaluated as false (F), incorrect
relations. In the second phase, a set of SW mappings (occurrences
of relations containing the same or equivalent source, s and target,
t terms in the publicly available SW data) is obtained for each
particular semantic relation. These mappings are extracted using
the services of the Watson SW gateway. Specific details about the
SW mapping extraction algorithm can be found in [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. In the
third phase, these mappings are formalized and represented in
terms of the values of their features (or attributes). The selected
attributes to represent each classified example are:
• e, the relation correctness {T, F}. This is the class attribute,
i.e., the one that will be predicted for future examples.
• Type(R), the type of relation to be evaluated: ⊆, ⊇, ⊥ and
named relations.
• | M |, the number of mappings.
• | M ⊆ |, the number of subclass mappings.
• | M ⊇ |, the number of superclass mappings.
• | M ⊥ |, the number of disjoint mappings.
• | M R |, the number of named related mappings.
• | M S |, the number of sibbling mappings.
• For each particular mapping Mi we consider
      </p>
      <p>Type (Ri), the relation type of the mapping: ⊆, ⊇, ⊥, named
and sibling.</p>
      <p>Pl (Mi) the path length of the mapping Mi
Np (Mi) the number of paths that lead to the mapping Mi.
Note that for sibling and named mappings the connection
can be derived from 2 different paths connected by a
common node.
| Mi ⊆ |, the number of subclass relations in Mi
| Mi ⊇ |, the number of superclass relations in Mi
| Mi ⊥|, the number of disjoint relations in Mi
| Mi R |, the number of named relations in Mi</p>
    </sec>
    <sec id="sec-3">
      <title>3. EXPERIMENTS AND RESULTS</title>
      <p>This study addressed four different classification problems:
predicting the correctness of any particular semantic relation (generic
classifiers) and predicting the correctness of a given type of
semantic relation: ⊆, ⊇ or named (specialized classifiers). Note that
the ⊥ relation has been discarded from our experiments due to the
lack of negative examples. To address each of these problems,
three different classifiers: the J48 Decision Tree, the NaiveBayes
classifier, and the LibSVM classifier, all of them provided by
Weka [5] were used. Each classifier was applied using the whole set
of attributes (Section 2) or a filtered set of attributes (af) obtained
using a combination of the cfSubsetEval and the BestFirst
algorithms [5]. To train and test the classifiers, each dataset was
divided in the following way: approximately 70% of the data was
used for training and 30% of the data was used for testing. This
division was done manually to avoid the appearance of mappings
coming from the same semantic relation in the training and the
test sets. Note that the SW mappings coming from the same
semantic relation share in common at least the first eight attributes,
therefore, it is important to maintain them together in the same set
(either the train or the test set) for a fair evaluation. To evaluate
the classifiers and compare them against each other the following
measures were selected: the percentage of correctly classified
instances, the percentage of incorrectly classified instances and,
the weighted average of the values obtained using the following
measures for the positive and negative class: True Positives rate
(TP), False Positives rate (FP), Precision, Recall, F-Measure
(FMea) and ROC area value. More details about these measures can
be found in [5]. The results obtained by the best classifier for each
classification problem can be seen in Table 1.
ROC</p>
      <p>Generic
J48
named</p>
    </sec>
    <sec id="sec-4">
      <title>4. CONCLUSIONS AND FUTURE WORK</title>
      <p>In this paper, we investigate the problem of predicting the
correctness of semantic relations. Our hypothesis is that ML
methods can be adapted to exploit the SW as a source of
knowledge to perform this task. The result of our
experiments are promising, reaching a maximum of 74.2% of
correctly classified semantic relations for classifiers able to
validate the correctness of multiple types of semantic
relations (generic classifiers) and up to 98% for classifiers
focused on evaluating the correctness of one particular
semantic relation (specialized classifiers).</p>
      <p>Despite the success in the prediction process obtained by
the classifiers, it is important to highlight that only 60% of
the relations contained in these datasets were covered by
the SW. This limits our approach to domains where
semantic information is available, which constitutes an open
problem for future research work.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Budanitsky</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Hirst</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <article-title>Evaluating WordNet-based measures of semantic distance</article-title>
          .
          <source>2006. Computational Linguistics</source>
          ,
          <volume>32</volume>
          (
          <issue>1</issue>
          ):
          <fpage>13</fpage>
          -
          <lpage>47</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Calibrasi</surname>
            ,
            <given-names>R.L.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Vitanyi</surname>
          </string-name>
          ,
          <string-name>
            <surname>P.M. The Google Similarity Distance</surname>
          </string-name>
          .
          <year>2007</year>
          .
          <source>IEEE Transactions on Knowledge and Data Engineering</source>
          ,
          <volume>19</volume>
          (
          <issue>3</issue>
          ):
          <fpage>370</fpage>
          -
          <lpage>383</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Cimiano</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <year>2006</year>
          .
          <article-title>Ontology Learning and Population from Text: Algorithms, Evaluation and Applications</article-title>
          . SpringerVerlag New York, Inc.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Sabou</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Gracia</surname>
            ,
            <given-names>J. Spider: Bringing</given-names>
          </string-name>
          <string-name>
            <surname>Non-Equivalence Mappings</surname>
          </string-name>
          to OAEI.
          <year>2008</year>
          .
          <source>In Proc. of the Third International Workshop on Ontology Matching</source>
          [5]
          <string-name>
            <surname>Witten</surname>
            ,
            <given-names>I.H.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Eibe F. Data Mining</surname>
          </string-name>
          .
          <source>Practical Machine Learning Tools and Techniques</source>
          .
          <year>2000</year>
          . The Morgan Kaufmann Series in Data Management Systems
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>