<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Multilingual Ontology Matching Evaluation { A First Report on using MultiFarm</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Christian Meilicke</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Cassia Trojahn</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ondrej Svab-Zamazal</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Dominique Ritze</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>INRIA &amp; LIG</institution>
          ,
          <addr-line>Grenoble</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Proceedings of the Second International Workshop on Evaluation of Semantic Technologies</institution>
          ,
          <addr-line>IWEST 2012</addr-line>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>University of Economics</institution>
          ,
          <addr-line>Prague</addr-line>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>University of Mannheim</institution>
        </aff>
      </contrib-group>
      <abstract>
        <p>This paper reports on the rst usage of the MultiFarm dataset for evaluating ontology matching systems. This dataset has been designed as a comprehensive benchmark for multilingual ontology matching. In this rst set of experiments, we analyze how state-of-the-art matching systems { not particularly designed for the task of multilingual ontology matching { perform on this dataset. Our experiments show the hardness of MultiFarm and result in baselines for any algorithm specifically designed for multilingual ontology matching. Moreover, this rst reporting allows us to draw relevant conclusions for both multilingual ontology matching and ontology matching evaluation in general.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>reference alignments or are not fully open. Thus, they are not suitable for an
extensive evaluation.</p>
      <p>
        For overcoming the lack of a comprehensive benchmark for multilingual
ontology, the MultiFarm7 dataset has been designed. This dataset is based on the
OntoFarm [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] dataset, which has been used successfully in OAEI in the
Conference track. MultiFarm is composed of a set of seven ontologies translated in
eight di erent languages and the complete corresponding alignments between
these ontologies.
      </p>
      <p>
        In this paper, we report on the rst usage of MultiFarm for multilingual
ontology matching evaluation. In [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], we have deeply discussed the design of
MultiFarm, focusing on its multilingual features and the speci cities of the
translation process, with a very preliminary report on its evaluation. Here, we extend
this preliminary evaluation and provide a deep discussion on the performance
of matching systems. Our evaluation is based on a representative subset of
MultiFarm8 and a set of state-of-the-art matching systems participating in OAEI
campaigns. These systems have not particularly been designed for matching
ontologies described in di erent languages. The choice for these settings was caused
by the fact that { to our knowledge { there exists no multilingual ontology
matching system that is executable out of the box. For example, an implementation
of a multilingual matching system is described in [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], however, it is not available
for download.
      </p>
      <p>We expect that the results of these systems can be topped by speci c
methods, however, with our experiments we establish a rst non-trivial baseline
speci c for the MultiFarm dataset. To our knowledge, such a comprehensive
evaluation has not yet been conducted so far in the eld of multilingual ontology
matching.</p>
      <p>The rest of the paper is organised as follows. In Section 2, we rst introduce
the OntoFarm dataset and then we present its multilingual counterpart. We
shortly discuss the hardness of MultiFarm and present the results that have
been gathered in previous OAEI campaigns on the OntoFarm. In Section 3, we
present the evaluation setting used to carry out our experiments and list the tools
we have evaluated. In particular, we discuss why and how we applied speci c
con gurations to some of the tools. In Section 4, we nally describe the results
of our experiments. We mainly focus on highly aggregated results due to the
enormous amount of generated data. In Section 5, we conclude the paper and
discuss directions for future work.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Background on MultiFarm</title>
      <p>
        In the following, we shortly describe the OntoFarm dataset, explain how
MultiFarm has been constructed, and roughly report about evaluation results of the
OAEI Conference track.
7 The dataset has been thoroughly described in [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] and is available at http://web.
      </p>
      <p>informatik.uni-mannheim.de/multifarm/
8 We have been discarding Russian and Chinese languages.
2.1</p>
      <sec id="sec-2-1">
        <title>OntoFarm</title>
        <p>The OntoFarm dataset is based on a set of 16 ontologies from conference
organisation domain. All contained ontologies di er in numbers of classes, properties,
and in their DL expressivity. They are very suitable for ontology matching tasks
since they were independently designed by di erent people who used various
kinds of resources for ontology design:
{ actual conferences and their web pages,
{ actual software tools for conference organisation support, and
{ experience of people with personal participation in organisation of actual
conferences</p>
        <p>Thus, the OntoFarm dataset describes a quite realistic matching scenario and
has been successfully applied in the OAEI within the Conference track since 2006.
In 2008, a rst version of the reference alignments was created and then annually
enriched and updated up to current 21 reference alignments built between seven
(out of 16) ontologies. Each of them has between four to 25 correspondences.
The relatively small number of correspondences in the reference alignments are
based on the fact that the reference alignments contain only simple equivalence
correspondences. Due to di erent modeling styles of the ontologies, for many
concepts and properties thus no equivalent counterparts exist.
2.2</p>
      </sec>
      <sec id="sec-2-2">
        <title>MultiFarm</title>
        <p>For generating the MultiFarm dataset, those seven OntoFarm ontologies, for
which reference alignments are available, were manually translated into eight
di erent languages (Chinese, Czech, Dutch, French, German, Portuguese,
Russian, and Spanish). Since native speakers with a certain knowledge about the
ontologies translated them, we do not expect any serious errors but of course
they can never be excluded at all. Based on these translations, it is possible
to re-create cross-lingual variants of the original test cases from the OntoFarm
dataset as well as to exploit the translations more directly. Thus, the MultiFarm
dataset contains two types of cross-lingual reference alignments.</p>
        <p>We have depicted a small subset of the dataset shown in Figure 1. This
gure indicates the cross-lingual reference alignments between di erent ontologies,
derived from original alignments and translations (type (i)), and cross-lingual
reference alignments between the same ontologies, which are directly based on
the translations or on exploiting transitivity of translations (type (ii)). Reference
alignments of type (i) cover only a small subset of all concepts and properties.
We have explained this above for the original test cases of the OntoFarm dataset.
In contrast, for test cases of type (ii) there are (translated) counterparts for each
concept and property.</p>
        <p>Overall, the MultiFarm dataset has 36 49 test cases. 36 is a number of
pairs of languages { each English ontology has its 8 language variants. 49 is the
number of all reference alignments for each language pair. This is implied from
example for type (i)</p>
        <p>example for type (ii)
es
en
CMT</p>
        <p>CMT
de</p>
        <p>CMT
pt</p>
        <sec id="sec-2-2-1">
          <title>EKAW de</title>
        </sec>
        <sec id="sec-2-2-2">
          <title>EKAW es</title>
        </sec>
        <sec id="sec-2-2-3">
          <title>EKAW</title>
          <p>pt
CMT
original reference alignment</p>
        </sec>
        <sec id="sec-2-2-4">
          <title>EKAW en</title>
          <p>the number of original reference alignments (21) which is doubled (42) due to the
fact that there is a di erence between CM Ten-EKAWde and CM Tde-EKAWen
in comparison with the original test cases where the test cases CM T -EKAW
and EKAW -CM T are not distinguished. Additionally, we can also construct
new reference alignments for matching each ontology on its translation which
gives us seven additional reference alignments for each pair.</p>
          <p>
            The main motivation for creating the MultiFarm dataset has been the ability
to create a comprehensive set of test cases of type (i). We have especially argued
in [
            <xref ref-type="bibr" rid="ref10">10</xref>
            ] that type (ii) test cases are not well suited for evaluating multilingual
ontology matching systems, because they can be solved with very speci c methods
that are not related to the multilingual matching task.
2.3
          </p>
        </sec>
      </sec>
      <sec id="sec-2-3">
        <title>Test Hardness</title>
        <p>The OntoFarm dataset has a very heterogeneous character due to di erent
modeling styles by various people. This leads to a high hardness of the resulting test
cases. For example, the object property writtenBy occurs in several OntoFarm
ontologies. When only considering the labels, one would expect that a
correspondence like writtenBy = writtenBy correctly describes that these object
properties are equivalent. However, in ontology O1 the property indicates that
a paper (domain) is written by an author (range), while in O2 the property
describes that a review (domain) is written by a reviewer (range). Therefore, this
correspondence is not contained in the reference alignment between O1 and O2.
Similarly, comparing the English against the Spanish variant, there are the
object properties writtenBy and escrito por. Pure translation would, similarly
to the monolingual example, not result in detecting a correct correspondence.
For that reason, the MultiFarm type (i) test cases go far beyond being a simple
translation task.</p>
        <p>The cross-lingual test cases of MultiFarm are probably much harder than
the monolingual test cases of the OntoFarm. It is thus important to know how</p>
        <p>F-measure=0.7</p>
        <p>F-measure=0.6
F-measure=0.5
2009-Top
2009-Avg
2010-Top
2010-Avg
2011-Top
2011-Avg
2011:
YAM++
CODI
LogMap
matching systems perform on the OntoFarm dataset. These results can be
understood as an upper bound that will be hard to top by results achieved for
MultiFarm. In Figure 2, we have depicted some results of previous OAEI
campaigns in a precision/recall triangular graph. This graph shows precision, recall,
and F-measure in a single plot. It includes the best (squares) and average
(circles) results of the 2009, 2010 and 2011 Conference track as well as results of
the three best ontology matching systems (triangles) from 2011. Best results are
considered according to the highest F-measure which corresponds to exactly one
ontology matching system for each year. In 2011, YAM++ achieved the highest
F-measure that is why its triangle overlaps with the white square depicting the
best result of 2011.</p>
        <p>rec=1.0
rec=.8
rec=.6
pre=.6
pre=.8
pre=1.0</p>
        <p>On the one hand, Figure 2 shows that there is an improvement every year,
except the average results of the last year. A reason might be the availability
of the complete dataset over several years. Since the MultiFarm dataset has not
been used in the past, we expect that evaluation results also improve over the
years. On the other hand, we can see that recall is not very high (.63 in 2010
and .60 in 2011 for the best matching systems). This indicates that test cases of
the OntoFarm dataset are especially di cult regarding recall measure.
3</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Evaluation Settings</title>
      <p>In the following, we explain how we executed our evaluation experiments and
list the matching systems that have been evaluated.
3.1</p>
      <sec id="sec-3-1">
        <title>Evaluation Work ow</title>
        <p>Following a general de nition, matching is the process that determines an
alignment A for a pair of ontologies O1 and O2. Besides the ontologies, the are other
input parameters that are relevant for the matching process, namely: (i) the
use of an input alignment A0, which is to be extended or completed by the
process; (ii) parameters that a ect the matching process, for instance, weights
and thresholds; and (iii) external resources used by the matching process, for
instance, common knowledge and domain speci c thesauri.</p>
        <p>In this paper, we focus on evaluating a standard matching task. (i) In most
of our experiments, we do not modify the parameters that a ect the matching
process. For two systems, we made an exception from this rule and report very
brie y on the results. (ii) We do not use an additional input alignment at all.
Note that most systems do not support such a functionaility. (iii) We put no
restriction on the external resources that are taken into account by the evaluated
systems. Thus, we use the system standard settings for our evaluation. However,
we obviously focus on the matching process where labels and annotations of O1
and O2 are described in di erent languages.</p>
        <p>Tests
Results</p>
        <p>O1
O2</p>
        <p>Tool</p>
        <p>SEALS Bundle</p>
        <p>SEALS platform</p>
        <p>The most common way to evaluate the quality of a matching process is to
evaluate the correctness (precision) and completeness (recall) of its outcome A by
comparing A against a reference alignment R. Since 2010, in the context of OAEI
campaigns, the process of evaluating matching systems has been automated
thanks to the SEALS platform (Figure 3). For OAEI 2011, participants have
been invited to wrap their tools into a format that can be executed by the
platform, i.e. the matching process is not conducted by the tool developer but
by the organisers of an evaluation using the platform. For the purpose of this
paper, we bene t from the large number of matching tools that become available
for our evaluation. Furthermore, evaluation test cases are available in the SEALS
repositories and can be used by everyone. Thus, all of our experiments can be
completely reproduced.
3.2</p>
      </sec>
      <sec id="sec-3-2">
        <title>Evaluated Matching Systems</title>
        <p>
          As stated before, a large set of matching systems has already been uploaded to
the platform in the context of OAEI 2011. We apply most of these tools to the
MultiFarm dataset. In particular, we evaluated the tools AROMA [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ], CIDER [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ],
CODI [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ], CSA [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ], LogMap and LogMapLt [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ], MaasMatch [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ], MapSSS [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ],
YAM++ [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ] and Lily [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ]. For most of these tools, we used the version
submitted to OAEI 2011. However, some tool developers have already submitted a
new version with some modi cations. This is the case for CODI, LogMap and
MapSSS. Moreover, the developer of LogMap has additionally uploaded a lite
version of their matching systems called LogMapLt.
        </p>
        <p>There have also been some systems participating in OAEI 2011 that are
not listed here. We have not added them to the evaluation for di erent reasons.
Some of these systems cannot nish the MultiFarm matching process in less than
several weeks while others generate empty alignments for nearly all matching
tasks or terminate with an error. We have to emphasise that none of these
systems has originally been designed to solve the multilingual matching task.
4</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Results</title>
      <p>In the following, we discuss the evaluation results on di erent perspectives: rst,
aggregating the results obtained for all pairs of test cases (and languages) per
matcher and then focusing on the di erent pairs of languages.
4.1</p>
      <sec id="sec-4-1">
        <title>Di erences in Test Cases</title>
        <p>As explained in Section 2, the dataset can be divided in (i) those test cases
where the ontologies to be matched are translations of di erent ontologies and
(ii) those test cases where the same original ontology has been translated into two
di erent languages and the translated ontologies have to be matched. We display
the results for test cases of type (i) on the left and those for type (ii) on the right
of Table 1. We have ordered the systems according to the F-measure for the test
cases of type (i). The best results, in terms of F-measure, are achieved by CIDER
(18%) followed by CODI (13%), LogMap (11%) and MapSSS (10%). CIDER
has both better precision and recall scores than any other system. Compared
to the top-results that have been reported for the original Conference dataset
(F-measure &gt; 60%), the test cases of the MultiFarm dataset are obviously much
harder. However, an F-measure of 18% is already a remarkable result given the
fact that we executed CIDER in its default setting.</p>
        <p>The outcomes for test cases of type (ii) di er signi cantly. In particular, the
results of MapSSS (67% F-measure) are surprisingly compared to the results
presented for test cases of type (i). This system can leverage the speci cs of type
matcher
CIDER
CODI
LogMap
MapSSS
LogMapLt
MaasMatch
CSA
YAM++
AromaLily
(ii) test cases to cope with the problem of matching labels expressed in di erent
languages. Similar to MapSSS, we also observe a higher F-measure for CODI,
CSA, and YAM++. We have marked those systems with an asterisk. Note that
all these systems have an F-measure of at least ve times higher than the
Fmeasure for test cases of type (i). For all other systems, we observe a slightly
decreased F-measure comparing test cases of type (i) with type (ii).</p>
        <p>Again, we have to highlight the di erences between both types of test cases.
Reference alignments of type (i) cover only a small fraction of all concepts and
properties described in the ontologies. This is not the case for test cases of
type (ii). Here, we have complete alignments that connect each concept and
property with an equivalent counterpart in the other ontology. There seems
to be a clear distinction between systems that are specialised or con gured to
generate complete alignments in the absence of (easy) usable label description,
and other systems that focus on generating good results for test cases of type
(i).</p>
        <p>Comparing these results with the results for the OAEI 2011 Benchmark track,
it turns out that all systems marked with an asterisk have been among the top
ve systems of this track. All Benchmark test cases have a similar property,
namely, their reference alignments contain for each entity of the smaller ontology
exactly one counterpart in the larger ontology. An explanation for this can be
that these systems have been developed or at least con gured to score well for
the Benchmark track. For that reason, they generate good results for test cases
of type (ii), while their results for test cases of type (i) are less good. MapSSS
and CODI are an exception. These systems generate good results for both test
cases of type (i) and (ii).
4.2</p>
      </sec>
      <sec id="sec-4-2">
        <title>Di erences in Languages</title>
        <p>Besides aggregating the results per matcher, we have analysed the results per
pair of languages (Table 2), for the case where di erent ontologies are matched
(type (i) in Table 1). As stated before, the multilingual aspect of the matching
process can be negligible for matchers that are able to adapt their strategies to
match structurally similar ontologies. We have also compared the matchers with
a simple edit distance algorithm on labels (edna).</p>
        <p>With exception of Aroma and Lily, which are not able to deal with the
complexity of the matching task, for most of the test cases no matcher has lower
F-measure than edna. For some of them, however, LogMap, LogMapLt,
MaasMatch and YAM++, respectively, have not provided any alignment. YAM++
has a speci c behaviour and is not able to match the English ontologies to any
other languages. For the other matchers, it (incidentally) happens mostly for
the pairs of languages that do not share the same root language (e.g. es-nl or
de-es). The exception is LogMapLt, which is not able to identify any
correspondence between fr-pt, even if these languages have the same root language (e.g.
Latin) and thus have a similar vocabulary. It could be expected that
matchers should be able to nd a higher number of correspondences for the pairs of
languages where there is an overlap in their vocabularies because most of the
matcher apply some label similarity strategy. However, it is not exactly the case
in MultiFarm. The dataset contains many complex correspondences that cannot
be found by a single translation process or by string comparison. This can be
partially corroborated by the very low performance of edna in all test cases.</p>
        <p>Looking at the results for each pair of languages, per matcher, the best ve
Fmeasures are obtained for de-en (31%), es-fr/es-pt (29%), de-es/en-es (25%), all
for CIDER, en-es/en-fr (24%), for CODI, and fr-nl (23%) again for CIDER. We
could observe that 3 ahead pairs contain languages with some degree of overlap
in their vocabularies (i.e., de-en, es-fr, es-pt). For each individual matcher, seven
out of eight matchers have their best scores for these pairs (exception is YAM++
that scores better for cz-pt and de-pt), with worst scores in cz-fr, es-nl, which
have very di erent vocabularies.</p>
        <p>When aggregating the results per pair of languages, that order is mostly
preserved (highly a ected by CIDER): de-en (17%), es-pt (16%), en-es (12%),
de-es/en-fr (11%), followed by fr-nl/en-nl (10%). The exception is for the pair
es-fr, where the aggregated F-measure decreases to 7%. Again, the worst scores
are obtained for cz-fr, nl-pt and es-nl. We can observe that, for most of the
cases, the features of the languages (i.e., their overlapping vocabularies) have
an impact in the matching results. However, there is no universal pattern and
we have cases with similar languages where systems score very low (fr-pt, for
instance). This has to be further analysed with a deep analysis of the individual
pairs of ontologies.
5</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Discussion</title>
      <p>Some of the reported results are relevant for multilingual ontology matching in
general, while others help us to understand the characteristics of the MultiFarm
dataset. The latter ones are relevant for any further evaluation that builds on the
dataset. Moreover, we can also draw some conclusions that might be important
for the use of datasets in the general context of ontology matching evaluation.
Exploiting structural information Very good results for test cases of type (ii)
can be achieved by methods non-speci c to multilingual ontology matching.
The result of MapSSS is an interesting example. This was also one of the main
reasons why the MultiFarm dataset has been constructed as a comprehensive
collection for test cases of type (i) and (ii). We suggest to put a stronger focus on
test cases of type (i) in the context of evaluating multilingual ontology matching
techniques. Otherwise, it remains unclear whether the measured results are based
on multilingual techniques or on exploiting that the matched ontologies can
interpreted as versions of the same ontology.</p>
      <p>
        Finding a good con guration The results for test cases of type (i) show that
stateof-the-art matching systems are not very well suited for the tasks of matching
ontologies described in di erent languages, especially when executed in their
default setting. We started another set of experiments by running some tools
(CODI, LogMap, Lily) in a manually con gured setting better suited for the
matching task. A rst glimpse, the results shows that it is possible to increase
the average F-measure up to a value of 26%. Thus, we are planning to further
investigate the in uence of con gurations on multilingual matching tasks within
more extensive experiments. 9
9 We would like to thank Ernesto Jimenez-Ruiz (LogMap [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]) and Peng Wang
(Lily [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]) for supporting us with a quick hint about a good, manually modi ed
con guration for running their systems on MultiFarm.
The role of language features We cannot neglect certain language features (like
their overlapping vocabularies) in the matching process. Once most of the
matchers take advantage of label similarities it is likely that it may be harder to nd
correspondences between Czech and Portuguese ontologies than Spanish and
Portuguese ones. In our evaluation, for most of the systems, the better
performance where incidentally observed for the pairs of languages that have some
degree of overlap in their vocabularies. This is somehow expected, however, we
could nd exceptions to this behavior. In fact, MultiFarm requires systems
exploiting more sophisticated matching strategies than label similarity and for
many ontologies in MultiFarm it is the case.
      </p>
      <p>Implications on analyzing OAEI results Aside from the topic of multilingual
ontology matching, the results implicitly emphasise the di erent characteristics of
test cases of type (i) and (ii). An example can be found when comparing results
for the OAEI Benchmark and Conference track. The Benchmark track is about
matching di erent versions (some slightly modi ed, some heavily modi ed) of
the same ontology. The Conference dataset is about matching di erent ontologies
describing the same domain. This di erence nds its counterparts in the
distinction between type (i) and type (ii) ontologies in the MultiFarm dataset. Without
taking this distinction into account, it is not easy to draw valid conclusions on
the generality of measured results.
6</p>
    </sec>
    <sec id="sec-6">
      <title>Future Work</title>
      <p>Even though we reported about diverse aspects, we could not analyse or
evaluate all interesting issues. The following listing shows possible extensions and
improvements for further evaluations.</p>
      <p>{ Executing (all or many) matching systems with a speci cally tailored
conguration;
{ Including Chinese and Russian versions of the ontologies in the evaluation
setting;
{ Exploiting automatic translation strategies and evaluate their impact on the
matching process;
{ Analysing the role of diacritics : in some languages, the same word written
with or without accent can have a di erent meaning, e.g., in French `ou'
(where) is di erent from `ou' (or).</p>
      <p>There are many things left to do, however, we have shown that the MultiFarm
dataset is a useful, comprehensive, and a di cult dataset for evaluating ontology
matching systems. We strongly recommend to apply this resource and to
compare measured results against the results presented in this paper. In particular,
we encourage developers of ontology matching systems, speci cally designed to
match ontologies described in di erent languages, to make use of the dataset
and to report about achieved results.
Some of the authors are partially supported by the SEALS project
(IST-2009238975). Ondrej Svab-Zamazal has been partially supported by the CSF grant
no. P202/10/0761.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <given-names>M.</given-names>
            <surname>Cheatham</surname>
          </string-name>
          .
          <article-title>MapSSS results for OAEI 2011</article-title>
          .
          <source>In Proc. 6th ISWC workshop on Ontology Matching (OM)</source>
          , pages
          <fpage>184</fpage>
          {
          <fpage>189</fpage>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <given-names>J.</given-names>
            <surname>David</surname>
          </string-name>
          .
          <article-title>Aroma results for OAEI 2009</article-title>
          .
          <source>In Proc. 4th ISWC workshop on Ontology Matching (OM)</source>
          , pages
          <fpage>147</fpage>
          {
          <fpage>152</fpage>
          ,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <given-names>J.</given-names>
            <surname>Euzenat</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Meilicke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Stuckenschmidt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Shvaiko</surname>
          </string-name>
          , and C. T. dos Santos.
          <article-title>Ontology alignment evaluation initiative: Six years of experience</article-title>
          .
          <source>Journal on Data Semantics</source>
          ,
          <volume>15</volume>
          :
          <fpage>158</fpage>
          {
          <fpage>192</fpage>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <given-names>J.</given-names>
            <surname>Euzenat</surname>
          </string-name>
          and
          <string-name>
            <given-names>P.</given-names>
            <surname>Shvaiko</surname>
          </string-name>
          . Ontology matching. Springer, Heidelberg (DE),
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <given-names>B.</given-names>
            <surname>Fu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Brennan</surname>
          </string-name>
          , and
          <string-name>
            <surname>D. O'Sullivan</surname>
          </string-name>
          .
          <article-title>Using pseudo feedback to improve crosslingual ontology mapping</article-title>
          .
          <source>In Proc. 8th Extended Semantic Web Conference (ESWC)</source>
          , pages
          <fpage>336</fpage>
          {
          <fpage>351</fpage>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <given-names>J.</given-names>
            <surname>Gracia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bernad</surname>
          </string-name>
          , and
          <string-name>
            <given-names>E.</given-names>
            <surname>Mena</surname>
          </string-name>
          .
          <article-title>Ontology matching with CIDER: evaluation report for OAEI 2011</article-title>
          .
          <source>In Proc. 6th ISWC workshop on Ontology Matching (OM)</source>
          , pages
          <fpage>126</fpage>
          {
          <fpage>133</fpage>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <given-names>J.</given-names>
            <surname>Huber</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Sztyler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Noessner</surname>
          </string-name>
          , and
          <string-name>
            <given-names>C.</given-names>
            <surname>Meilicke</surname>
          </string-name>
          . Codi:
          <article-title>Combinatorial optimization for data integration: results for OAEI 2011</article-title>
          .
          <source>In Proc. 6th ISWC workshop on Ontology Matching (OM)</source>
          , pages
          <fpage>134</fpage>
          {
          <fpage>141</fpage>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <given-names>E.</given-names>
            <surname>Jimenez-Ruiz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Morant</surname>
          </string-name>
          , and
          <string-name>
            <given-names>B. C.</given-names>
            <surname>Grau</surname>
          </string-name>
          .
          <article-title>LogMap results for OAEI 2011</article-title>
          .
          <source>In Proc. 6th ISWC workshop on Ontology Matching (OM)</source>
          , pages
          <fpage>163</fpage>
          {
          <fpage>170</fpage>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <given-names>Y.</given-names>
            <surname>Kalfoglou</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Schorlemmer</surname>
          </string-name>
          .
          <article-title>Ontology mapping: the state of the art</article-title>
          .
          <source>The Knowledge Engineering Review</source>
          ,
          <volume>18</volume>
          (
          <issue>1</issue>
          ):1{
          <fpage>31</fpage>
          ,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>C. Meilicke</surname>
            ,
            <given-names>R. G.</given-names>
          </string-name>
          <string-name>
            <surname>Castro</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Freitas</surname>
            ,
            <given-names>W. R. van Hage</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Montiel-Ponsoda</surname>
          </string-name>
          ,
          <string-name>
            <surname>R</surname>
          </string-name>
          . R. de Azevedo,
          <string-name>
            <given-names>H.</given-names>
            <surname>Stuckenschmidt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Svab-Zamazal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Svatek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Tamilin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Trojahn</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S.</given-names>
            <surname>Wang</surname>
          </string-name>
          .
          <article-title>Multifarm: A benchmark for multilingual ontology matching</article-title>
          .
          <source>Journal of Web Semantics. Submitted</source>
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11. E. Rahm and
          <string-name>
            <given-names>P.</given-names>
            <surname>Bernstein</surname>
          </string-name>
          .
          <article-title>A survey of approaches to automatic schema matching</article-title>
          .
          <source>The VLDB Journal</source>
          ,
          <volume>10</volume>
          (
          <issue>4</issue>
          ):
          <volume>334</volume>
          {
          <fpage>350</fpage>
          ,
          <year>2001</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <given-names>F.</given-names>
            <surname>Schadd</surname>
          </string-name>
          and
          <string-name>
            <given-names>N.</given-names>
            <surname>Roos</surname>
          </string-name>
          .
          <article-title>Maasmatch results for OAEI 2011</article-title>
          .
          <source>In Proc. 6th ISWC workshop on Ontology Matching (OM)</source>
          , pages
          <fpage>171</fpage>
          {
          <fpage>178</fpage>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <given-names>F.</given-names>
            <surname>Schadd</surname>
          </string-name>
          and
          <string-name>
            <given-names>N.</given-names>
            <surname>Roos</surname>
          </string-name>
          . YAM++
          <article-title>{ results for OAEI 2011</article-title>
          .
          <source>In Proc. 6th ISWC workshop on Ontology Matching (OM)</source>
          , pages
          <fpage>228</fpage>
          {
          <fpage>235</fpage>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <given-names>D.</given-names>
            <surname>Spohr</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Hollink</surname>
          </string-name>
          , and
          <string-name>
            <surname>P. Cimiano.</surname>
          </string-name>
          <article-title>A machine learning approach to multilingual and cross-lingual ontology matching</article-title>
          .
          <source>In Proc. 10th International Semantic Web Conference (ISWC)</source>
          , pages
          <fpage>665</fpage>
          {
          <fpage>680</fpage>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <given-names>Q.-V.</given-names>
            <surname>Tran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ichise</surname>
          </string-name>
          , and
          <string-name>
            <given-names>B.-Q.</given-names>
            <surname>Ho</surname>
          </string-name>
          .
          <article-title>Cluster-based similarity aggregation for ontology matching</article-title>
          .
          <source>In Proc. 6th ISWC workshop on Ontology Matching (OM)</source>
          , pages
          <fpage>142</fpage>
          {
          <fpage>147</fpage>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <given-names>O.</given-names>
            <surname>Svab</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Svatek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Berka</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Rak</surname>
          </string-name>
          , and
          <string-name>
            <given-names>P.</given-names>
            <surname>Tomasek</surname>
          </string-name>
          . Ontofarm:
          <article-title>Towards an experimental collection of parallel ontologies</article-title>
          .
          <source>In Poster Track of ISWC</source>
          <year>2005</year>
          ,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <given-names>P.</given-names>
            <surname>Wang</surname>
          </string-name>
          .
          <article-title>Lily results on SEALS platform for OAEI 2011</article-title>
          .
          <source>In Proc. 6th ISWC workshop on Ontology Matching (OM)</source>
          , pages
          <fpage>156</fpage>
          {
          <fpage>162</fpage>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>