<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Applying of Machine Learning Techniques to Combine String-based, Language-based and Structure-based Similarity Measures for Ontology Matching</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>Institute of Informatics Problems, Federal Research Center “Computer Science and Control” of the Russian Academy of Sciences</institution>
          ,
          <addr-line>Moscow</addr-line>
          ,
          <country country="RU">Russia</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Lomonosov Moscow State University</institution>
          ,
          <addr-line>Moscow</addr-line>
          ,
          <country country="RU">Russia</country>
        </aff>
      </contrib-group>
      <fpage>0000</fpage>
      <lpage>0003</lpage>
      <abstract>
        <p>In the areas of Semantic Web and data integration, ontology matching is one of the important steps to resolve semantic heterogeneity. Manual ontology matching is very labor-intensive, time-consuming and prone to errors. So development of automatic or semi-automatic ontology matching methods and tools is quite important. This paper applies machine learning with different similarity measures between ontology elements as features for ontology matching. An approach to combine string-based, language-based and structurebased similarity measures with machine learning techniques is proposed. Logistic Regression, Random Forest classifier and Gradient Boosting are used as machine learning methods. The approach is evaluated on two datasets of Ontology Alignment Evaluation Initiative (OAEI).</p>
      </abstract>
      <kwd-group>
        <kwd>ontology matching</kwd>
        <kwd>machine learning</kwd>
        <kwd>similarity measures</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        An ontology is “a formal, explicit specification of shared conceptualization” [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], where
conceptualisation is an abstract model of some phenomenon in the world. Ontologies
were created to facilitate the sharing of knowledge and its reuse [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. They are used for
organization of knowledge and for communication between computing systems,
people, computing systems and people [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Ontologies deal with the following kinds of
entities: classes, properties and individuals. A class (concept) of ontology is a
collection of objects, i.e., “Person” (the class of all people) or “Car” (the class of all
cars). Property (attribute) describes characteristics of a class or relations between
classes, i.e., “has as name” or “is created by”. Individual (instance) is a particular
instance or object represented by a concept, i.e., “a human cytochrome C” is an instance
of the concept “Protein” [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>
        Ontology matching is a process of establishing correspondences between
semantically related entities in different ontologies [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. A set of correspondences
(equivalence, subsumption, disjointness) between ontologies elements is called an
alignment. Ontology matching can be applied in many different subject areas: Semantic
Web, Peer-to-Peer (P2P) systems, learning systems, multi-agent systems [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>
        In the Semantic Web, ontologies are used to extract logical conclusions from data.
Many ontologies on the same subject areas have been created recently. These
ontologies have a different format and they cannot exchange information, so it is
necessary to apply ontology matching [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. In P2P systems ontology matching is used
to reduce the semantical heterogeneity (differences in the interpretation of the meaning)
between the queries of the users to system [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. In learning systems ontology matching
is a way to ease the knowledge share and reuse [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. In multi-agent systems, ontology
matching is used for interaction of different agents [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ].
      </p>
      <p>
        Ontology matching can be used also for schema mapping during data integration
[
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. Data integration is a process of combining the heterogeneous data sources into a
unified view. Schema mapping is a process of establishing correspondences between
elements of two different semantically related schemas (i.e. database schemas) [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ].
Ontology matching can help to resolve semantical heterogeneity during schema
mapping, for instance, if the schemas have ontologies as metadata or external domain
knowledge [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ].
      </p>
      <p>
        The classification of ontology matching approaches is very similar to the
classification of schema matching approaches [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. Matchers can be individual
(use single matcher criterion) or combining (combination of individual matchers). An
individual matcher can be schema-based (uses information about classes, properties
and their relationships) or instance-based (uses information about instances/content).
Schema-based matchers are divided into element-level (uses information about element
without its relationships) and structure-level (uses information about structure and
hierarchy). Combining matchers can be hybrid (creates alignment using several
matching criteria in sequentially) or composite (combines several independent
matching results). Composite matchers are divided into manual composition matchers
and automatic composition matchers [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ].
      </p>
      <p>
        This paper proposes an approach for combining individual element-level and
structure-level matchers into an automatic composition matcher based on machine
learning. Individual matchers produce similarity measures between ontology elements.
In terms of machine learning, similarity measures are used as features [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. The
composite matcher is a machine learning model trained on these features. Logistic
Regression, Random Forest and Gradient Boosting are used as the machine learning
methods in the paper. The idea of the approach is as follows: to combine a large number
of different similarity measures from other papers [
        <xref ref-type="bibr" rid="ref32">32</xref>
        ][33][34] in hope of increasing
the universality of the approach, that is, applicability to different subject areas.
      </p>
      <p>The paper is structured as follows. In Section 2 related works on application of
machine learning for ontology and schema matching are reviewed. Section 3 describes
the formal problem statement and an evaluation metric. In Section 4, similarity
measures and machine learning techniques applied are listed. Section 5 considers
implementation and evaluation issues.</p>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <p>This section reviews related works on schema matching and ontology matching because
they often applies similar techniques.</p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], a classification of approaches for schema matching is introduced and a
review of existing systems for schema matching is conducted. In [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], the authors
described the classification of ontology matching approaches based on [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], existing
matching systems, evaluation methods, similarity measures and matching strategies.
The most promising option is the apply combining matcher because it uses much more
information than an individual matcher. Many papers showed that the combining
matchers are more accurate than individual ones [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ]. Most approaches use
string-based similarity measures, i.e., N-gram [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ], Soundex [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ] [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ],
Levenshtein distance [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ] [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ], Jaro measure [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ] [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ] [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ] and others.
Languagebased similarity measures are also used, i.e., information from lexical database
WordNet [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] [
        <xref ref-type="bibr" rid="ref27">27</xref>
        ] [
        <xref ref-type="bibr" rid="ref28">28</xref>
        ] or vector representation of words from Word2vec models [
        <xref ref-type="bibr" rid="ref29">29</xref>
        ]
[
        <xref ref-type="bibr" rid="ref21">21</xref>
        ] [
        <xref ref-type="bibr" rid="ref30">30</xref>
        ] [
        <xref ref-type="bibr" rid="ref31">31</xref>
        ]. Some articles described structure-based similarity measures: differences
between numbers of properties [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ], similarity measures based on subclasses or parents
[
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] and graph-based similarity [
        <xref ref-type="bibr" rid="ref32">32</xref>
        ].
      </p>
      <p>
        Supervised machine learning for combining similarity measures is used in [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. The
authors describe an approach matching only concepts of ontologies. They used the
string-based similarity measures (prefix, suffix, Edit distance, n-gram), language-based
similarity measures (WordNet, Wu&amp;Palmer, description, Lin), similarity measures
between lists of words (for instance, name “socialNetwork” is divided into list of words
[“Social”, “Network”]), and structure similarity measures (string-based and
languagebased similarity measures between parents). Support Vector Machine (SVM) is used as
a machine learning method. The authors conducted experiments with data from
Ontology Alignment Evaluation Initiative 2007 (OAEI). The data is constructed from
three Internet directories (Google, Yahoo and Looksmart) and contains 4639 pairs of
ontologies defined using OWL language. The authors used 10-fold cross validation and
got 56.1% accuracy, 52.5% precision and 92.5% recall on average.
      </p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref51">55</xref>
        ] the authors combined the various similarity measures into a input sample for
the first time. String-based, linguistic-based and structure-based similarity measures are
used.
      </p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] language, structural and web similarity measures are used. Web similarity
measure “Web-dice” is the difference between the count of pages in a search engine
when searching for an entity.` An SVM method is selected for training. The dataset
used is OAEI benchmark tests ontologies. The authors trained two models
“SVMClass” and “SVM-Property” for matching classes and properties respectively.
      </p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref50">54</xref>
        ] 10 string-based, linguistic-based and instance-based similarity measures are
used as features. Decision Tree (DT) and Naive Bayes are used for classification. The
authors achieved 0.845 F-measure value.
      </p>
      <p>In [33] SVM, K-Nearest Neighbours (KNN), SVM, DT and AdaBoost are used as
the machine learning methods. The authors choose OAEI ontologies #301, #102 and
#103 as train dataset and ontologies #302, #303, #304 as test dataset and achieved 0.99
F-measure value.</p>
      <p>In [34] Stoilos, Soft Jaccard and Lin similarity measures are used for names, labels
and comments of entities. The authors also used information on abbreviations. Samples
from “Conference” track and benchmarks from OAEI are used as datasets. Multilayer
perceptron, Decision Trees and M5Rules are selected as machine learning methods.
The authors achieved 0.67 F-measure value.</p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref31">31</xref>
        ] the authors used string-based similarity measures, measures related with
parents and children of entities and chose “Conference” track from OAEI and EuroVoc
dataset.
      </p>
      <p>Note that known works use different datasets for their experiments and it is very
hard to compare them with each other.
3
3.1</p>
    </sec>
    <sec id="sec-3">
      <title>Ontology Matching as a Machine Learning Problem</title>
      <sec id="sec-3-1">
        <title>Formal Problem Statement of Ontology Matching</title>
        <p>Let ontology be a tuple (C, P, H). Here C is a set of classes, P is a set of properties. H
define the hierarchical relationships between classes. Other components of ontologies
like axioms and instances are not applied for ontology matching in the paper. The
objective of ontology matching is to find an alignment between classes and properties
of a source ontology O1 and a target ontology O2. An alignment is a set of tuples
( 1,  2,  ) , where  1 is an entity of  1 ,  2 is an entity of  2 , and  is the
confidence of the correspondence. A predicted alignment is the alignment obtained by
ontology matching. A true alignment is a manual alignment conducted by an domain
expert.
3.2</p>
      </sec>
      <sec id="sec-3-2">
        <title>Ontology Matching Problem as a Machine Learning Problem</title>
        <p>
          Entity pairs are extracted from source and target ontologies. Each pair of entities is
assigned with a label “0” or “1”, where “0” means that entities do not match, “1”
means that entities match. Thus, the problem is reduced to a machine learning binary
classification problem. The authors of most of the reviewed papers and OAEI used
Fmeasure for the evaluation of their approaches [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ][33].
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4 An Approach for Ontology Matching Applying Machine</title>
    </sec>
    <sec id="sec-5">
      <title>Learning Models Trained on Similarity Measures</title>
      <p>This section describes machine learning techniques applied (subsection 4.1), similarity
measures used (subsection 4.2) and algorithms constituting the approach (subsection
4.3).
4.1</p>
      <sec id="sec-5-1">
        <title>Machine Learning Techniques</title>
        <p>
          The following machine learning methods are applied in this paper: Logistic Regression,
Random Forest and Gradient Boosting. In [
          <xref ref-type="bibr" rid="ref31">31</xref>
          ][
          <xref ref-type="bibr" rid="ref43">47</xref>
          ] it is shown that a powerful ensemble
method Random Forest [
          <xref ref-type="bibr" rid="ref44">48</xref>
          ] and relatively simple and interpretable Logistic Regression
outperformed other machine learning algorithms like Gaussian Naive Bayes, K-nearest
Neighbors Algorithm, Classification and Regression Trees for ontology matching.
Gradient boosting has proven itself in many machine learning contests [
          <xref ref-type="bibr" rid="ref45">49</xref>
          ][
          <xref ref-type="bibr" rid="ref46">50</xref>
          ], so it
was also selected as a machine learning method to be applied. In the future, we also
want to test neural network (multilayer perceptron) as a machine learning method and
an approach based on automatic machine learning1.
4.2
        </p>
      </sec>
      <sec id="sec-5-2">
        <title>Similarity Measures</title>
        <p>
          String-based. We used all string-based similarity measures listed in our previous work
[
          <xref ref-type="bibr" rid="ref21">21</xref>
          ]. The listed metrics are aimed at handling various sorts of scenarios. N-gram
consider similarity of substrings and it is efficient when some characters are missing
[
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]. Dice coefficient is defined as twice the number of common words of compared
strings over the total number of words in both strings [35]. Jaccard and Generalized
Jaccard similarity are defined as the size of the intersection divided by the size of the
union of the sample sets of words [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ]. Levenshtein distance between two strings is the
minimum number of single-character edits required to change one word into the other
[36]. Jaro and Jaro-Winkler measures is edit distance measure designed for short
strings [
          <xref ref-type="bibr" rid="ref33">37</xref>
          ]. Monge-Elkan is a type of hybrid similarity measure that combines the
benefits of sequence-based and set-based methods [
          <xref ref-type="bibr" rid="ref34">38</xref>
          ]. The Smith-Waterman measure
determine similar regions between two strings [35]. The Needleman-Wunsh distance is
computed by assigning a score to each alignment between the two input strings and
choosing the score of the best alignment [
          <xref ref-type="bibr" rid="ref35">39</xref>
          ]. The Affine gap distance is an extension
of the Needleman-Wunsch measure that handles the longer gaps more gracefully [
          <xref ref-type="bibr" rid="ref36">40</xref>
          ].
The Bag distance is edit distance for sets of words [
          <xref ref-type="bibr" rid="ref48">52</xref>
          ]. Cosine similarity transforms a
string into vector so Euclidean cosine rule is used to determine similarity [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ]. Fuzzy
Wuzzy Partial Ratio finds the similarity measure between the shorter string and every
substring of length m of the longer string, and returns the maximum of those similarity
measures [
          <xref ref-type="bibr" rid="ref37">41</xref>
          ]. Soft TF-IDF and TF-IDF are numerical statistics that are intended to
reflect how important a word is to a document in a collection or corpus [
          <xref ref-type="bibr" rid="ref35">39</xref>
          ]. Partial
Token Sort2 and Token Sort are obtained by splitting the two strings into tokens and
then sorting the tokens. The score is the fuzzy wuzzy partial ratio raw score of the
transformed strings. Fuzzy Wuzzy Ratio is the ratio of the number of matching
characters to the total number of characters of two strings [
          <xref ref-type="bibr" rid="ref37">41</xref>
          ]. Editex [
          <xref ref-type="bibr" rid="ref38">42</xref>
          ] and
Soundex3 are phonetic matching measures. Tversky Index is an asymmetric similarity
measure on sets that compares a variant to a prototype [
          <xref ref-type="bibr" rid="ref39">43</xref>
          ]. Overlap coefficient is
defined as the size of the intersection divided by the smaller of the size of the two sets
[
          <xref ref-type="bibr" rid="ref40">44</xref>
          ].
        </p>
        <sec id="sec-5-2-1">
          <title>1 https://github.com/automl/auto-sklearn 2 https://anhaidgroup.github.io/py_stringmatching/v0.3.x/PartialTokenSort.html 3 http://anhaidgroup.github.io/py_stringmatching/v0.4.1/Soundex.html</title>
          <p>
            Language-based. It is possible that words differ but are close in meaning, i.e., “car”
and “auto”. WordNet can solve this problem. Wu and Palmer similarity are used for
handling this scenario [
            <xref ref-type="bibr" rid="ref41">45</xref>
            ]. If the strings consist of several words then the maximum
similarity measure of all possible pairs of sets of words is taken. But the weakness of
WordNet is that it contains only a part of all words of the language. Usage of vector
representations of words from Word2vec models [
            <xref ref-type="bibr" rid="ref42">46</xref>
            ] facilitates this problem. Cosine
similarity between two vector representations of words is calculated. If the strings
consists of several words then Sentence2vec algorithm from [
            <xref ref-type="bibr" rid="ref30">30</xref>
            ] is used.
Structure-based. Additionally, structure-based similarity measures are used: all listed
string-based and language-based similarity measures between parents of entities and
between paths of entities. These similarity measures embrace the hypothesis that
matched entities have similar parents and a similar place in hierarchy.
          </p>
          <p>Since we used the same model for the match of classes and properties, we added
feature “Type”, in which label “1” means class and label “0” means property.</p>
          <p>Such an extensive selection of similarity measures is aimed to get as much
information as possible so that a machine learning model is able to select the best
factors for prediction. Finally, we chose for each pair of entities 88 similarity measures
(29 for names, 29 for parents, 29 for paths, 1 for type), which are described in Table 1.</p>
          <p>N-gram 1, N-gram 2, N-gram 3, N-gram 4, Dice coefficient,
Jaccard similarity, Jaro measure, Monge-Elkan,
SmithWaterman, Needleman-Wunsh, Affine gap, Bag distance, Cosine
similarity, Partial Ratio, Soft TF-IDF, Editex, Generalized
Jaccard, Jaro-Winkler, Levenshtein distance, Partial Token Sort,
Fuzzy Wuzzy Ratio, Soundex, TF-IDF, Token Sort, Tversky
Index, Overlap coefficient, Longest common subsequence
Wu and Palmer similarity
Word2vec and Sentence2vec similarity
All string-based and language-based similarity measures between
parents of entities
All string-based and language-based similarity measures between
paths of entities
4.3</p>
        </sec>
      </sec>
      <sec id="sec-5-3">
        <title>Training and Matching Algorithms</title>
        <p>The approach is restricted with the following limitations: entities are matched only by
equivalence relation, classes are matched only with classes, properties are matched only
with properties, instances of ontologies are not used. The approach includes two main
algorithms: training of a machine learning model (training phase) and using it to predict
alignment (testing phase).</p>
        <p>The ontology matching algorithm using the trained model is described as follows:
Algorithm 1 Matching algorithm</p>
      </sec>
      <sec id="sec-5-4">
        <title>Input:</title>
        <p>ontology1, ontology2 - input ontologies,
THRESHOLD - threshold for create matching between entities</p>
        <p>Here ← denotes an assignment operation, and ∪ - the operation of merging lists.
The input data of the algorithm are two ontologies and a matching probability threshold
for filtering pairs of entities. If the probability is higher than the threshold, then the pair
is added to the alignment. A list of classes is extracted from each ontology. Next, two
lists of classes are fed to the input of Algorithm 2:
Algorithm 2 Creating predicted alignment from two lists of entities
create_alignment(entities1, entities2, THRESHOLD)</p>
      </sec>
      <sec id="sec-5-5">
        <title>Input:</title>
        <p>entities1, entities2 - input lists of entities (classes or properties),
THRESHOLD - threshold for create matching between entities</p>
      </sec>
      <sec id="sec-5-6">
        <title>Auxiliary functions:</title>
        <p>calculate_all_sim_measures - Algorithm 3,
predict_match - predict confidence based on similarity measures
Output: alignment - output alignment for entities1 and entities2
1 for entity1 ∈ entities1 do
2 for entity2 ∈ entities2 do
3 sim_measures ← calculate_all_sim_measures(entity1, entity2)
4 match ← predict_match(sim_measures)
5 if match &gt; THRESHOLD then
6 alignment ← alignment ∪ (entity1, entity2)
7 end if
8 end for
9 end for
10 return alignment</p>
        <p>Then, each class from the first ontology is matched with each class from the second
ontology. For example, if in the first ontology includes 10 classes and in the second
ontology includes 12 classes, then 120 pairs are matched. Each pair is fed to the input
of a machine learning model, which calculates the probability (confidence) of matching
for each pair. Then the threshold is set: if the probability is above the threshold, then
the pair is added to the final alignment. Similar actions are performed for properties.
The similarity measures for each pair are calculated in Algorithm 3:
Algorithm 3 Calculating similarity measures algorithm
calculate_all_sim_measures(entity1, entity2)</p>
      </sec>
      <sec id="sec-5-7">
        <title>Input:</title>
        <p>entity1, entity2 - input entities (classes or properties)
Auxiliary functions:
get_name - get name of entity,
get_parent - get parent of entity,
get_path - get full path of entity,
calculate_sim_measures - calculates string-based and linguistic-based similarity
measures listed in 4.2 and returns a list of 88 values
concat - merge lists
Output: sim_measures - output list of calculated similarity measures for entity1 and
entity2
1 name1 ← get_name(entity1)
2 name2 ← get_name(entity2)
3 parent1 ← get_parent(entity1)
4 parent2 ← get_parent(entity2)
5 path1 ← get_path(entity1)
6 path2 ← get_path(entity2)
7 name_sim_measures ← calculate_sim_measures(name1, name2)
8 parent_sim_measures ← calculate_sim_measures(parent1, parent2)
9 path_sim_measures ← calculate_sim_measures(path1, path2)
10 sim_measures ← concat(name_sum_measures, parent_sim_measures,
path_sim_measures)
11 return sim_measures</p>
        <p>Name, parent name, and the full hierarchical path are retrieved from each class. The
parent of a class is its super class. The full path is a string that describe the entire
hierarchy of classes: from the most general class to the current class. For example, the
class “Book” has the name “Book”, the parent name “Publication” and the full path
“Thing/Publication/Book”. Thus, a list of pairs for matching is generated. For
properties, the parent is the class that it describes. And the full path is a string describing
the complete hierarchy up to the class that describes the property. For each pair, all
similarity measures listed in Section 4.2 are calculated. Then all similarity measures
are combined into a list.</p>
        <p>The algorithm of model training is described as follows:
Algorithm 4 Creating dataset and training a machine learning model</p>
      </sec>
      <sec id="sec-5-8">
        <title>Input:</title>
        <p>train_pairs_ontologies - set of tuples (ontology1, ontology2, true_alignment)
model_name - name of machine learning method (logistic regression, random forest,
gradient boosting),
model_params - set of parameters of machine learning model,
create_dataset - Algorithm 5
Auxiliary functions: train_model - train machine learning model on training dataset
Output: model - trained model for predicting matching
10 model ← train_model(train_dataset, model_name, model_params)
11 return model
The input data is a list of ontology pairs and the true alignment between them. A model
from Section 2.5 and its parameters are also selected. The process is similar to the first
algorithm: the names of objects, the names of parents and full paths are retrieved, and
the similarity measures are calculated. First, a dataset is created for the classes, then for
properties, and after that the datasets are combined.</p>
        <p>The algorithm for creating a dataset is described in Algorithm 5:
Algorithm 5 Creating dataset from two lists of entities - create_dataset(entities1,
entities2, true_alignment, type_entity)</p>
      </sec>
      <sec id="sec-5-9">
        <title>Input:</title>
        <p>true_alignment - set of matched pairs of entities,
entities1, entities2 - input lists of entities (classes or properties),
type_entity - type of input entities (class or property)
Auxiliary functions: train_model - train machine learning model on training dataset
Output: train_dataset - output list of tuples with pairs of entities, their matchings and
similarity measures</p>
        <p>The input is a true alignment, two lists of entities and the type of input entities. Each
entity from the first list is mapped to each entity from the second list. Then, if a pair of
entities is contained in the true alignment, then the pair is assigned label “1”, otherwise
- label “0”. Also, each pair indicates the type of entity (either “Class” or “Property”)
because the same model was used to map classes and properties. Then all pairs are
combined into one dataset. Further, the model is trained on the created dataset with the
selected parameters.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Implementation and Evaluation Results</title>
      <sec id="sec-6-1">
        <title>Datasets</title>
        <p>Two datasets are selected for evaluation experiments (called as Dataset #1 and Dataset
#2 below). These datasets are sets of ontologies and their true alignments taken from
Ontology Alignment Evaluation Initiative (OAEI). Some pairs of ontologies and their
true alignments are selected for training the machine learning models and their testing.
This selection is called a partition. Ontologies from OAEI are used in many papers.
These papers include [33] and [34]. [33] presents an approach to combining similarity
measures without instances of ontologies and user feedback. KNN, SVM, DT and
AdaBoost were used as machine learning models. The authors achieve on some
alignments the value of F-measure 0.99. [34] proposed a new ontology matching
approach. The authors used five different similarity measures: syntactic, semantic,
abbreviation and context similarity. Multilayer Perceptron, REPTree, M5Rules are
used as the machine learning models. Average F-measure is 0.67. Dataset #1 is a
partition from third experiment of [33]. Dataset #2 is a partition from [34]. The used
pairs of ontologies and their true alignments are described in Tables 3 and 4. All
ontologies are defined using OWL-DL4 language in the RDF and XML format.</p>
        <p>Dataset #1 is a set of ontologies about Bibliographic references from Benchmark test
library. Ontology #101 is the reference ontology. Other ontologies (#102-#103,
#301#304) are compared with the reference ontology. Dataset has 7 ontologies and 6 true
alignments: 3 alignments for training and 3 alignments for testing.</p>
        <p>Dataset #2 consists several ontologies from Benchmark test library and all
ontologies from Conference track of OAEI. Conference track contains 16 ontologies,
which dealing with conference organization, and 21 true alignments. Dataset has 27
ontologies and 26 alignments: 8 alignments for training and 18 alignments for testing.</p>
        <p>The pairs of entities from each pair of ontologies and their alignments are extracted.
Dataset #1 has 14148 training samples (156 positive and 13992 negative samples) and
14940 testing samples (172 positive and 14768 negative samples) and Dataset #2 has
55348 training samples (284 positive and 55064 negative samples) and 114045 testing
samples (253 positive and 113792 negative samples). A positive sample is a pair of
entities which are matching, and a negative example is a pair of non-matching entities.
Note that the datasets are very unbalanced.
5.2</p>
      </sec>
      <sec id="sec-6-2">
        <title>Implementation</title>
        <p>The approach was implemented using Python 3.5. This language is widely used for
implementation of machine learning workflows and possesses a lot of useful program
libraries.</p>
        <sec id="sec-6-2-1">
          <title>4 https://www.w3.org/TR/owl-features/</title>
          <p>Entity from Ontology #101
Entity from Ontology #302
Collection
TechReport
Report
date
Book
TechReport
Publication
Resource
publishedOn</p>
          <p>
            Ontologies are represented as RDF/OWL files. The owlready25 library was used for
syntactic parsing of ontologies. Alignments are defined in RDF format. For parsing
alignments, the BeautifulSoup 6 library was used. As implementation of logistic
regression and random forest machine learning techniques sklearn7 library is used. As
a gradient boosting implementation the XGBoost8 library is used. Dataset is formed as
a dataframe of the pandas9 library. To evaluate F-measure, the Alignment API10 library
was used. Computation experiments: training of machine learning models and the
selection of their parameters were performed at the Hybrid high-performance
computing cluster [
            <xref ref-type="bibr" rid="ref47">51</xref>
            ]. WordNet dictionary is taken from the nltk11 library. Word2vec
model was trained on GoogleNews12 news. N-gram implementation is taken from the
ngram13 library. Similarity measures based on edit distance implementation is taken
from the editdistance14 library.
5.3
          </p>
        </sec>
      </sec>
      <sec id="sec-6-3">
        <title>Experiments</title>
        <p>The best parameters for the models were selected by the brute force method (a grid of
values was created for each parameter): the models were trained on all combinations of
parameters and the model with the best F-measure value using threshold 0.5 was
selected.</p>
        <p>For logistic regression, the following parameters were selected: inverse of
regularization strength, weights of classes and norm used in the penalization. For
random forest the number of trees in the forest, the maximum depth of the tree, the
5 https://owlready2.readthedocs.io/en/latest/
6 https://pypi.org/project/beautifulsoup4/
7 https://scikit-learn.org/stable/
8 https://xgboost.readthedocs.io
9 https://pandas.pydata.org
10 http://alignapi.gforge.inria.fr
11 https://www.nltk.org
12 https://github.com/mmihaltz/word2vec-GoogleNews-vectors
13 https://pythonhosted.org/ngram/
14 https://pypi.org/project/editdistance/
number of features to consider when looking for the best split and class weights were
selected. For XGBoost, minimum sum of instance weight, minimum loss reduction
requited to make a further partition, subsample ratio for training instances, subsample
ratio of columns when constructing each tree and maximum depth of a tree were
selected.</p>
        <p>After training and searching for the best parameters, a threshold was selected with
the highest F-measure value for each alignment. For each machine learning model, a
grid of values for parameters was created manually. For numerical parameters, a grid
of 3-5 values with different steps was created, i.e. for numbers of estimators in random
forest: 10, 100, 200, 500, 1000. For parameters with options, all possible options were
taken (2-4 options).</p>
        <p>
          The values of F-measure for each alignment are presented in tables 3, 4 and 5. The
best models for the Dataset #1 are logistic regression and random forest. Gradient
boosting is a bit less accurate. However, the gradient boosting is the best model on
average on Dataset #2. It is more accurate than logistic regression at 0.02 and than
random forest at 0.01. The values of F-measure on Dataset #1 are comparable with the
classical methods [
          <xref ref-type="bibr" rid="ref33">37</xref>
          ] [
          <xref ref-type="bibr" rid="ref34">38</xref>
          ] [
          <xref ref-type="bibr" rid="ref50">54</xref>
          ] [
          <xref ref-type="bibr" rid="ref51">55</xref>
          ] but lower than [33]. This may be associated with
a specific set of training and test datasets, and it is also possible that the metrics that
were not implemented in this work have an impact. In [33] the importance of each
similarity measures is not described, but there is a hypothesis that the main contribution
comes from similarity measures associated with comments to entities, and two
structural measures from [
          <xref ref-type="bibr" rid="ref51">55</xref>
          ]. The study of this issue is future work. The values of
Fmeasure on Dataset #2 are comparable with [34].
        </p>
        <p>The computational complexity of the approach is O(n1n2 + m1m2), where n1, n2 are
the number of classes in the ontology O1 and O2, and m1 and m2 are the number of
properties. The computation time and the used memory depending on the size of the
ontology are showed on figures 1 and 2. The calculations were performed on MacBook
Air 1.8 GHz 8GB RAM. The dependence of training and testing time on the ontology
size is showed on figure 1. 20 points (evenly distributed between 10 and 1000) were
used to build the figure. Training of random forest is longer than training of logistic
regression and gradient boosting. The reason for the jumps on the figure is that the
dataset is sampled randomly. Unfortunately, the used machine did not have enough
capacity to calculate the time for training and testing gradient boosting with an ontology
size of more than 600. The testing time of logistic regression and gradient boosting is
much less than a random forest, therefore in the figure the graphs are close to zero. In
general, there is a quadratic dependence. The dependence of memory usage on the
ontology size is showed on figure 2. 20 points were also used to build the figure.
Memory was measured using the memory profiler 15 package: the amount of used
memory was measured when running the training and testing script. It is hard to
understand why increasing the size of the ontology does not increase the amount of
memory used, perhaps this is due to the internal work of the Python language. It is
noticeable that the most memory is used by gradient boosting.
15 https://pypi.org/project/memory-profiler/</p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>Conclusions and Future Work</title>
      <p>We combined string-based, language-based, and structural-based similarity measures
using three different machine learning models and apply them for ontology matching
problem. The approach is implemented and evaluated using datasets selected from
Ontology Alignment Evaluation Initiative (OAEI).</p>
      <p>
        Due to the large number of similarity measures, there is hope that there is a potential
for a more universal use of the approach. Universality refers to the applicability of the
different subject areas. It is necessary to test the approach on ontologies with other
subject areas. As a future work we would like to add similarity measures based on
comments of entities, more structure-based similarity measures, such as a path length,
a number of children, a number of properties of a class. It is also necessary to test the
similarity measure from [
        <xref ref-type="bibr" rid="ref49">53</xref>
        ]. Neural network (multilayer perceptron) is planned to be
used as a machine learning model. Evaluation issues to be resolved are checking the
effectiveness of learning two different models separately for classes and properties and
testing different strategies to resolve a problem of the strong imbalance of classes as
well as strategies for significant reduce of a number of pairs of entities for matching.
      </p>
      <p>
        Acknowledgments. The research is financially supported by Russian Foundation
for Basic Research, projects 18-07-01434, 18-29-22096. The calculations were
performed by Hybrid high-performance computing cluster of FRC CS RAS [
        <xref ref-type="bibr" rid="ref47">51</xref>
        ].
33.
34.
35.
36.
      </p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Gruber</surname>
          </string-name>
          , T.:
          <article-title>A Translation Approach to Portable Ontology Specifications</article-title>
          . In:
          <article-title>Knowledge Acquisition - Special issue: Current issues in knowledge modeling</article-title>
          , vol.
          <volume>5</volume>
          , issue 2 (
          <year>1993</year>
          ). doi:
          <volume>10</volume>
          .1006/knac.
          <year>1993</year>
          .1008
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Fensel</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          : Ontologies:
          <article-title>Silver Bullet for Knowledge Management and Electronic Commerce</article-title>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>662</fpage>
          -09083-1
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Gruninger</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>J.: Ontology</given-names>
          </string-name>
          <string-name>
            <surname>Applications</surname>
          </string-name>
          and Design - Introduction. In:
          <article-title>Communications of the ACM (</article-title>
          <year>2002</year>
          ). doi:
          <volume>10</volume>
          .1145/503124.503146
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Euzenat</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shvaiko</surname>
            ,
            <given-names>P .</given-names>
          </string-name>
          : Ontology Matching. Springer-Verlag Berlin Heidelberg, Berlin (
          <year>2007</year>
          ). doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>642</fpage>
          -38721-0
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Otero-Cerdeira</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rodríguez-Martínez</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gómez- Rodríguez</surname>
          </string-name>
          , A.:
          <article-title>Ontology Matching: A Literature Review</article-title>
          .
          <source>In: Expert Systems with Applications</source>
          , vol.
          <volume>42</volume>
          , issue 2, pp.
          <fpage>949</fpage>
          -
          <lpage>971</lpage>
          (
          <year>2015</year>
          ). doi:
          <volume>10</volume>
          .1016/j.eswa.
          <year>2014</year>
          .
          <volume>08</volume>
          .032
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Shvaiko</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Euzenat</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>A Survey of Schema-Based Matching Approaches</article-title>
          .
          <source>In: Journal on Data Semantics IV</source>
          , pp.
          <fpage>146</fpage>
          -
          <lpage>171</lpage>
          (
          <year>2005</year>
          ). doi:
          <volume>10</volume>
          .1007/11603412_
          <fpage>5</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Atencia</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Euzenat</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pirro</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rousset</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Alignment-Based Trust for Resource Finding in Semantic P2P Networks</article-title>
          .
          <source>In: The Semantic Web - ISWC</source>
          <year>2011</year>
          : 10th International Semantic Web Conference, pp.
          <fpage>51</fpage>
          -
          <lpage>66</lpage>
          (
          <year>2011</year>
          ).doi:
          <volume>10</volume>
          .1007/978 -3-
          <fpage>642</fpage>
          -25073-
          <issue>6</issue>
          _
          <fpage>4</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Arch-int</surname>
          </string-name>
          , N.,
          <article-title>Arch-int, S.: Semantic Ontology Mapping for Interoperability of Learning Resource Systems using a rule-based reasoning approach</article-title>
          .
          <source>In: Expert Systems with Applications</source>
          , vol.
          <volume>40</volume>
          , issue 18, pp.
          <fpage>7428</fpage>
          -
          <lpage>7443</lpage>
          (
          <year>2013</year>
          ). doi: https://doi.org/10.1016/j.eswa.
          <year>2013</year>
          .
          <volume>07</volume>
          .027
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Mascardi</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ancona</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bordini</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ricci</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>CooL-AgentSpeak: Enhancing AgentSpeak-DL Agents with Plan Exchange and Ontology Services</article-title>
          .
          <source>In: WI-IAT '11 Proceedings of the 2011 IEEE/WIC/ACM International Conferences on Web Intelligence and Intelligent Agent Technology</source>
          , vol.
          <volume>2</volume>
          , pp.
          <fpage>109</fpage>
          -
          <lpage>116</lpage>
          (
          <year>2011</year>
          ). doi:
          <volume>10</volume>
          .1109/WIIAT.
          <year>2011</year>
          .255
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Dong</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Srivastava</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Big Data Integration</article-title>
          . In: 2013 IEEE 29th International Conference on (
          <year>2015</year>
          ). doi:
          <volume>10</volume>
          .1109/ICDE.
          <year>2013</year>
          .6544914
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11. E. Rahm,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bernstein</surname>
          </string-name>
          .:
          <article-title>A survey of approaches to automatic schema matching</article-title>
          .
          <source>In: The International Journal on Very Large Data Bases</source>
          , vol.
          <volume>10</volume>
          ,
          <string-name>
            <surname>issue</surname>
            <given-names>4</given-names>
          </string-name>
          ,
          <year>December 2001</year>
          , pp.
          <fpage>334</fpage>
          -
          <lpage>350</lpage>
          . doi:
          <volume>10</volume>
          .1007/ s007780100057
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Hlaing</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Ontology based schema matching and mapping approach for structured databases</article-title>
          .
          <source>In: Proceedings of the 2nd International Conference on Interaction Sciences: Information Technology, Culture and Human</source>
          , pp.
          <fpage>853</fpage>
          -
          <lpage>859</lpage>
          (
          <year>2009</year>
          ). doi:
          <volume>10</volume>
          .1145/1655925.1656080
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Nathalie</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Schema Matching Based on Attribute Values and Background Ontology</article-title>
          .
          <source>In: 12th AGILE International Conference on Geographic Information Science</source>
          (
          <year>2009</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Ichise</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <article-title>: Machine Learning Approach for Ontology Mapping using Multiple Concept Similarity Measures</article-title>
          . In: Seventh IEEE/ACIS International Conference on Computer and Information Science (
          <year>2008</year>
          ). doi:
          <volume>10</volume>
          .1109/ICIS.
          <year>2008</year>
          .51
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Mao</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Peng</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Spring</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Neural Network based Constraint Satisfaction in Ontology Mapping</article-title>
          .
          <source>In: Proceedings of the Twenty-Third AAAI Conference on Artificial Intelligence</source>
          , vol.
          <volume>2</volume>
          , pp
          <fpage>1207</fpage>
          -
          <lpage>1212</lpage>
          (
          <year>2008</year>
          ). http://www.dit.unitn.it/~p2p/RelatedWork/Matching/AAAI10-MaoM.pdf
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Do</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Melnik</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rahm</surname>
          </string-name>
          , E.:
          <article-title>Comparison of Schema Matching Evaluations</article-title>
          . In:
          <article-title>Revised Papers from the NODe 2002 Web and Database-Related Workshops on Web</article-title>
          ,
          <source>WebServices, and Database Systems</source>
          , pp.
          <fpage>221</fpage>
          -
          <lpage>237</lpage>
          (
          <year>2002</year>
          ). doi:
          <volume>10</volume>
          .1007/3-540-36560-5_
          <fpage>17</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Do</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rahm</surname>
          </string-name>
          , E.:
          <article-title>COMA: a system for flexible combination of schema matching approaches</article-title>
          .
          <source>In: VLDB '02 Proceedings of the 28th international conference on Very Large Data Bases</source>
          , pp.
          <fpage>610</fpage>
          -
          <lpage>621</lpage>
          (
          <year>2002</year>
          ). doi:
          <volume>10</volume>
          .1016/B978-155860869-6/
          <fpage>50060</fpage>
          -
          <lpage>3</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18. L.
          <string-name>
            <surname>Xu</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Embley</surname>
          </string-name>
          .:
          <article-title>Automating Schema Mapping for Data Integration</article-title>
          . (
          <year>2003</year>
          ). http:// www.deg.byu.edu/papers/AutomatingSchemaMatching.journal.pdf
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Lambrix</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tan</surname>
          </string-name>
          , H.:
          <article-title>SAMBO-A system for aligning and merging biomedical ontologies</article-title>
          .
          <source>In: Journal of Web Semantics</source>
          , vol.
          <volume>4</volume>
          , issue 3, pp.
          <fpage>196</fpage>
          -
          <lpage>206</lpage>
          (
          <year>2006</year>
          ). doi:
          <volume>10</volume>
          .1016/j.websem.
          <year>2006</year>
          .
          <volume>05</volume>
          .003
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Ngo</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Enhancing Ontology Matching by Using Machine Learning, Graph Matching and Information Retrieval Techniques</article-title>
          . In: University Montpellier II -
          <source>Sciences et Techniques du Languedoc</source>
          (
          <year>2012</year>
          ).
          <source>doi: 10.1.1.302.587</source>
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>Bulygin</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Combining Lexical and Semantic Similarity Measures with Machine Learning Approach for Ontology and Schema Matching Problem</article-title>
          .
          <source>In: Selected Papers of the XX International Conference on Data Analytics and Management in Data Intensive Domains</source>
          , pp.
          <fpage>245</fpage>
          -
          <lpage>249</lpage>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22.
          <string-name>
            <surname>Gal</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Modica</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jamil</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Eyal</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Automatic Ontology Matching Using Application Semantics</article-title>
          . In: AI Magazine -
          <article-title>Special issue on semantic integration</article-title>
          , vol.
          <volume>26</volume>
          , issue 1, pp.
          <fpage>21</fpage>
          -
          <lpage>31</lpage>
          (
          <year>2005</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <surname>Hariri</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sayyadi</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Abolhassani</surname>
          </string-name>
          , H.:
          <article-title>Combining Ontology Alignment Metrics Using the Data Mining Techniques</article-title>
          .
          <source>In: Proceedings of the 2nd International Workshop on Contexts and Ontologies: Theory, Practice and Applications</source>
          (
          <year>2006</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          24.
          <string-name>
            <surname>Stoilos</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stamou</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kolias</surname>
            ,
            <given-names>S.:</given-names>
          </string-name>
          <article-title>A String Metric for Ontology Alignment</article-title>
          .
          <source>In: The Semantic Web - ISWC</source>
          <year>2005</year>
          , pp.
          <fpage>624</fpage>
          -
          <lpage>637</lpage>
          (
          <year>2005</year>
          ). doi:
          <volume>10</volume>
          .1007/11574620_
          <fpage>45</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          25.
          <string-name>
            <surname>Cheatham</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hitzler</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>String Similarity Metrics for Ontology Alignment</article-title>
          .
          <source>In: The Semantic Web - ISWC</source>
          <year>2013</year>
          , pp.
          <fpage>294</fpage>
          -
          <lpage>309</lpage>
          (
          <year>2013</year>
          ). doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>642</fpage>
          -41338-4_
          <fpage>19</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          26.
          <string-name>
            <surname>Saruladha</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Aghila</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sathiya</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>A Comparative Analysis of Ontology and Schema Matching Systems</article-title>
          . In:
          <source>International Journal of Computer Applications</source>
          , vol.
          <volume>34</volume>
          , issue 8, pp.
          <fpage>14</fpage>
          -
          <lpage>21</lpage>
          (
          <year>2011</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          27.
          <string-name>
            <surname>Jean-Mary</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shironoshita</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kabuka</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Ontology Matching with Semantic Verification</article-title>
          .
          <source>In: Web Semant</source>
          , vol.
          <volume>7</volume>
          , issue 3, pp.
          <fpage>235</fpage>
          -
          <lpage>251</lpage>
          (
          <year>2009</year>
          ). doi:
          <volume>10</volume>
          .1016/j.websem.
          <year>2009</year>
          .
          <volume>04</volume>
          .001
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          28.
          <string-name>
            <surname>Seddiqui</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Aono</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Anchor-flood: Results for OAEI 2009</article-title>
          .
          <source>In: Proceedings of the 4th International Workshop on Ontology Matching collocated with the 8th International Semantic Web Conference</source>
          (
          <year>2009</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          29.
          <string-name>
            <surname>Kolyvakis</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kalousis</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kiritsis</surname>
            ,
            <given-names>D.:</given-names>
          </string-name>
          <article-title>DeepAlignment: Unsupervised Ontology Matching with Refined Word Vectors</article-title>
          .
          <source>In: Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          , vol.
          <volume>1</volume>
          (
          <year>2018</year>
          ). doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>N18</fpage>
          -1072
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          30.
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lai</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>He</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhao</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lv</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          :
          <article-title>Ontology Matching with Word Embeddings</article-title>
          .
          <source>In: Chinese Computational Linguistics and Natural Language Processing Based on Naturally Annotated Big Data</source>
          , pp
          <fpage>34</fpage>
          -
          <lpage>45</lpage>
          (
          <year>2014</year>
          ). doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>319</fpage>
          -12277-
          <issue>9</issue>
          _
          <fpage>4</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          31.
          <string-name>
            <surname>Nagy</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vargas-Vera</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Motta</surname>
          </string-name>
          , E.:
          <article-title>DSSim-ontology mapping with uncertainty</article-title>
          .
          <source>In: 1st International Workshop on Ontology Matching</source>
          (
          <year>2006</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          32.
          <string-name>
            <surname>Nkisi-Orji</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wiratunga</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Massie</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hui</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Heaven</surname>
          </string-name>
          , R.:
          <article-title>Ontology alignment based on word embedding and random forest classification</article-title>
          .
          <source>In: Energy Transfer Processes in Polynuclear Lanthanide Complexes</source>
          , pp.
          <fpage>557</fpage>
          -
          <lpage>572</lpage>
          (
          <year>2018</year>
          ). doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>030</fpage>
          -10925- 7_
          <fpage>34</fpage>
          <string-name>
            <surname>Nezhadi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shadgar</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Osareh</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Ontology Alignment Using Machine Learning Techniques</article-title>
          .
          <source>In: International Journal of Computer Science &amp; Information Technology</source>
          , vol.
          <volume>3</volume>
          , pp.
          <fpage>139</fpage>
          -
          <lpage>150</lpage>
          (
          <year>2011</year>
          ). doi:
          <volume>10</volume>
          .5121/ijcsit.
          <year>2011</year>
          .3210 Alboukaey,
          <string-name>
            <given-names>N.</given-names>
            ,
            <surname>Joukhadar</surname>
          </string-name>
          ,
          <string-name>
            <surname>A.</surname>
          </string-name>
          :
          <article-title>Ontology Matching as Regression Problem</article-title>
          .
          <source>In: Journal of Digital Information Management</source>
          , vol.
          <volume>16</volume>
          , issue
          <volume>1</volume>
          (
          <year>2018</year>
          ). http://dline.info/fpaper/jdim/v16i1/jdimv16i1_4.
          <string-name>
            <surname>pdf Cohen</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ravikumar</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fienberg</surname>
            ,
            <given-names>S.:</given-names>
          </string-name>
          <article-title>A Comparison of String Metrics for Matching Names and Records</article-title>
          . Euzenat,
          <string-name>
            <surname>J.:</surname>
          </string-name>
          <article-title>An API for ontology alignment</article-title>
          .
          <source>In: The Semantic Web - ISWC</source>
          <year>2004</year>
          : Third International Semantic Web Conference (
          <year>2004</year>
          ). doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>540</fpage>
          -30475-3_
          <fpage>48</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          37.
          <string-name>
            <surname>David</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Guillet</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Briand</surname>
          </string-name>
          , H.:
          <article-title>Association Rule Ontology Matching Approach</article-title>
          . In:
          <source>International Journal on Semantic Web and information systems</source>
          , vol.
          <volume>3</volume>
          , issue 2, pp.
          <fpage>27</fpage>
          -
          <lpage>49</lpage>
          (
          <year>2007</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          38.
          <string-name>
            <surname>Straccia</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Troncy</surname>
          </string-name>
          , R.: oMAP:
          <article-title>Combining Classifiers for Aligning Automatically OWL Ontologies</article-title>
          .
          <source>In: Web Information Systems Engineering</source>
          , pp.
          <fpage>133</fpage>
          -
          <lpage>147</lpage>
          (
          <year>2005</year>
          ). doi:
          <volume>10</volume>
          .1007/11581062_
          <fpage>11</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          39.
          <string-name>
            <surname>Needleman</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wunsch</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <string-name>
            <given-names>A General</given-names>
            <surname>Method</surname>
          </string-name>
          <article-title>Applicable to Search for Similarities in Amino Acid Sequence of 2 Proteins</article-title>
          . In
          <source>: Journal of Molecular Biology</source>
          , vol.
          <volume>48</volume>
          . issue 3, pp.
          <fpage>443</fpage>
          -
          <lpage>53</lpage>
          (
          <year>1970</year>
          ). doi:
          <volume>10</volume>
          .1016/
          <fpage>0022</fpage>
          -
          <lpage>2836</lpage>
          (
          <issue>70</issue>
          )
          <fpage>90057</fpage>
          -
          <lpage>4</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          40.
          <string-name>
            <surname>Doan</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Halevy</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ives</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          :
          <article-title>Principles of Data Integration</article-title>
          . (
          <year>2012</year>
          ). doi:
          <volume>10</volume>
          .1016/C2011-0-06130-6
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          41.
          <string-name>
            <given-names>Appa</given-names>
            <surname>Rao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            ,
            <surname>Srinivas</surname>
          </string-name>
          ,
          <string-name>
            <surname>G.</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Venkata</given-names>
            <surname>Rao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            ,
            <surname>Prasad</surname>
          </string-name>
          <string-name>
            <surname>Reddy</surname>
          </string-name>
          ,
          <string-name>
            <surname>P.:</surname>
          </string-name>
          <article-title>A partial ratio and ratio based fuzzy-wuzzy procedure for characteristic mining of mathematical formulas from documents</article-title>
          . (
          <year>2018</year>
          ). doi:
          <volume>10</volume>
          .21917/ijsc.
          <year>2018</year>
          .0242
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          42.
          <string-name>
            <surname>Zobel</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dart</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Phonetic String Matching: Lessons from Information Retrieval</article-title>
          .
          <source>In: SIGIR '96 Proceedings of the 19th annual international ACM SIGIR conference on Research and development in information retrieval</source>
          , pp.
          <fpage>166</fpage>
          -
          <lpage>172</lpage>
          (
          <year>1996</year>
          ). doi:
          <volume>10</volume>
          .1145/243199.243258
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          43.
          <string-name>
            <surname>Tversky</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Features of Similarity</article-title>
          .
          <source>In: Psychological Review</source>
          , vol.
          <volume>84</volume>
          , issue 4, pp.
          <fpage>327</fpage>
          -
          <lpage>352</lpage>
          (
          <year>1977</year>
          ).
          <source>doi: 10.1037/0033-295X.84.4.327</source>
        </mixed-citation>
      </ref>
      <ref id="ref40">
        <mixed-citation>
          44.
          <string-name>
            <surname>Vijaymeena</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kavitha</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>A Survey on Similarity Measures in Text Mining</article-title>
          . (
          <year>2016</year>
          ). doi:
          <volume>10</volume>
          .5121/mlaij.
          <year>2016</year>
          .3103
        </mixed-citation>
      </ref>
      <ref id="ref41">
        <mixed-citation>
          45.
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Palmer</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Verbs Semantics and Lexical Selection</article-title>
          .
          <source>In: Proceedings of the 32nd annual meeting on Association for Computational Linguistics</source>
          (
          <year>1994</year>
          ). doi:
          <volume>10</volume>
          .3115/981732.981751
        </mixed-citation>
      </ref>
      <ref id="ref42">
        <mixed-citation>
          46.
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Corrado</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dean</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          :
          <article-title>Efficient Estimation of Word Representations in Vector Space</article-title>
          .
          <source>In: Proceedings of the International Conference on Learning Representations</source>
          (
          <year>2013</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref43">
        <mixed-citation>
          47.
          <string-name>
            <surname>Jurisch</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Igler</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>RDF2Vec-based Classification of Ontology Alignment Changes</article-title>
          . (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref44">
        <mixed-citation>
          48.
          <string-name>
            <surname>Breiman</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Random Forests</article-title>
          .
          <source>In: Machine Learning</source>
          , vol.
          <volume>45</volume>
          , issue 1, pp.
          <fpage>5</fpage>
          -
          <lpage>32</lpage>
          . doi:
          <volume>10</volume>
          .1023/A:1010933404324
        </mixed-citation>
      </ref>
      <ref id="ref45">
        <mixed-citation>
          49.
          <string-name>
            <surname>Volkovs</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wei</surname>
            <given-names>Yu</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            ,
            <surname>Poutanen</surname>
          </string-name>
          ,
          <string-name>
            <surname>T.</surname>
          </string-name>
          :
          <article-title>Content-based Neighbor Models for Cold Start in Recommender Systems</article-title>
          .
          <source>In: Proceedings of the Recommender Systems Challenge</source>
          (
          <year>2017</year>
          ). doi:
          <volume>10</volume>
          .1145/3124791.3124792
        </mixed-citation>
      </ref>
      <ref id="ref46">
        <mixed-citation>
          50.
          <string-name>
            <surname>Sandulescu</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chiru</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Predicting the future relevance of research institutions - The winning solution of the KDD Cup</article-title>
          <year>2016</year>
          . (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref47">
        <mixed-citation>
          51. Federal Research Center Computer Science and Control of Russian Academy of Sciences. Available at: http://hhpcc.frccsc.
          <source>ru (accessed</source>
          <volume>09</volume>
          /12/2018)
        </mixed-citation>
      </ref>
      <ref id="ref48">
        <mixed-citation>
          52.
          <string-name>
            <surname>Nobarian</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Derakhshi</surname>
            ,
            <given-names>M.:</given-names>
          </string-name>
          <article-title>The Review of Fields Similarity Estimation Methods</article-title>
          .
          <source>In: International Journal of Machine Learning and Computing</source>
          , vol.
          <volume>2</volume>
          (
          <year>2012</year>
          ). doi:
          <volume>10</volume>
          .7763/IJMLC.
          <year>2012</year>
          .V2.200
        </mixed-citation>
      </ref>
      <ref id="ref49">
        <mixed-citation>
          53.
          <string-name>
            <surname>Znamenskij</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Stable assessment of the quality of similarity algorithms of character strings and their normalizations</article-title>
          .
          <source>In: Program systems: theory and applications</source>
          , vol.
          <volume>9</volume>
          ,
          <issue>issue</issue>
          39, pp.
          <fpage>561</fpage>
          -
          <lpage>578</lpage>
          (
          <year>2018</year>
          ). doi:
          <volume>10</volume>
          .25209/2079-3316 -2018-9-4-
          <fpage>561</fpage>
          -578
        </mixed-citation>
      </ref>
      <ref id="ref50">
        <mixed-citation>
          54.
          <string-name>
            <surname>Eckert</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Meilicke</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stuckenschmidt</surname>
          </string-name>
          , H.:
          <article-title>Improving Ontology Matching using Meta-level Learning</article-title>
          .
          <source>In: The Semantic Web: Research and Applications</source>
          , pp.
          <fpage>158</fpage>
          -
          <lpage>172</lpage>
          (
          <year>2009</year>
          ). doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>642</fpage>
          -02121-3_
          <fpage>15</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref51">
        <mixed-citation>
          55.
          <string-name>
            <surname>Euzenat</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Guégan</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Valtchev</surname>
            ,
            <given-names>P.:</given-names>
          </string-name>
          <article-title>OLA in the OAEI 2005 alignment contest</article-title>
          .
          <source>In: Proceedings of the K-CAP 2005 Workshop on Integrating Ontologies</source>
          (
          <year>2005</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>