<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Cross-lingual ontology matching with CIDER-LM: results for OAEI 2022</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Javier Vela</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jorge Gracia</string-name>
          <email>jogracia@unizar.es</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Transformers</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>CIDER-LM</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Sentence-BERT</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="editor">
          <string-name>Cross-lingual Ontology Matching, Ontology Alignment, Natural Language Processing, Language Models,</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science and Systems Engineering</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>María de Luna 1</institution>
          ,
          <addr-line>50018 Zaragoza</addr-line>
          ,
          <country country="ES">Spain</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2022</year>
      </pub-date>
      <abstract>
        <p>In this paper, the CIDER-LM cross-lingual matching system is presented, as well as the results it achieved during the OAEI (Ontology Alignment Evaluation Initiative) 2022 campaign. This is the first appearance of CIDER-LM in OAEI where it only participated in MultiFarm, the track for cross-lingual ontology alignment evaluation. The matching system uses a pre-trained multilingual language model based on transformers, fine-tuned using the openly available portion of the MultiFarm dataset. The model calculates the vector embeddings of the labels associated to every ontology entity and its context. The confidence degree between matching entities is computed as the cosine similarity between their associated embeddings. CIDER-LM is novel in the use of multilingual language models for cross-lingual ontology matching. Its initial version obtained promising results in the OAEI'22 MultiFarm track, attaining a modest precision but the best overall performance in recall.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        sis [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], CIDER-LM uses language models as BERT [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] based on the Transformer architecture [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
Transformers are encoder-decoder neural networks with a self attention mechanism. They are
very popular for solving Natural Language Processing (NLP) tasks. BERT is a language model
that leverages the transformer encoder to predict tokens in a sentence when given a context.
Numerous studies have been published exploring how BERT can be fine-tuned to solve a wide
variety of NLP tasks. The original BERT was pre-trained using a document-level corpus in
English, extracted from the BooksCorpus1 (800M words) and the English Wikipedia2 (2,500M
https://github.com/javiervela/ (J. Vela); http://jogracia.url.ph/web/ (J. Gracia)
      </p>
      <p>© 2022 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
CEUR
Workshop
Proceedings
words).</p>
      <p>
        The use of BERT and similar models has proven to be useful in ontology matching, for
example, in the biomedical domain [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], but they have been largely unexplored for cross-lingual
matching until now. Multilingual language models can be used to that end, which can represent
tokens in several languages in the same embedding space.
      </p>
      <p>
        Sentence-BERT (SBERT) [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], which is a modification of the pre-trained BERT, produces
semantically meaningful embeddings for sentences. These sentence embeddings are aligned
in the same embedding space, in contrast to other models that create the embeddings at the
token level. In addition, SBERT monolingual models can be made multilingual using knowledge
distillation [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. The resulting models produce vector embeddings that can be compared to find
the similarity between sentences in two diferent languages.
      </p>
      <p>
        CIDER-LM uses SBERT to associate embeddings to the labels of the ontology entities (and
their context), which are compared using cosine similarity to create a matching confidence
score between them. The choice of sentence embeddings instead of token embeddings (as
other multilingual language models do) is motivated by the fact that many ontology labels are
conformed by multi-word expressions, thus embeddings that represent whole sentences (and
not only atomic tokens) are more suitable to capture their semantic content and to compute
similarities between them [
        <xref ref-type="bibr" rid="ref6 ref8">8, 6</xref>
        ].
      </p>
    </sec>
    <sec id="sec-2">
      <title>1.1. State, purpose, general statement</title>
      <p>
        CIDER-LM is a cross-lingual ontology matching system that utilizes a transformer-based
pretrained multilingual language model, fine-tuned using the openly available portion of the
MultiFarm dataset3. The model calculates the vector embeddings of the labels associated to
every ontology entity and its context. The confidence degree between two matching entities
is computed as the cosine similarity between their associated embeddings. The generated
alignments are one-to-one mappings between entities from two input ontologies. The input
ontologies must be in OWL or RDF-S format, and the output provided is expressed in Alignment
Format4. The type of discovered correspondence is “equivalence”, with a confidence degree in
[
        <xref ref-type="bibr" rid="ref1">0, 1</xref>
        ]. CIDER-LM works with ontology classes and properties, not yet with instances.
      </p>
    </sec>
    <sec id="sec-3">
      <title>1.2. Specific techniques used</title>
      <p>
        CIDER-LM integrates a fine-tuned version of distiluse-base-multilingual-cased-v25, a multilingual
pre-trained model from Sentence Transformers6. This ontology aligner is implemented in
Python and is wrapped using the Matching EvaLuation Toolkit (MELT), a framework for ontology
matching and evaluation [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. The system is packaged as a Docker image implementing the Web
Interface7.
      </p>
      <p>An overall view of the CIDER-LM matching pipeline is shown in Figure 1. In short, the
process is as follows. First, the aligner receives a source ( 1) and target ( 2) ontology. Both</p>
      <sec id="sec-3-1">
        <title>3https://www.irit.fr/recherches/MELODI/multifarm/</title>
        <p>4http://alignapi.gforge.inria.fr/format.html
5https://huggingface.co/sentence-transformers/distiluse-base-multilingual-cased-v2
6https://www.sbert.net/docs/pretrained_models.html#multi-lingual-models
7https://dwslab.github.io/melt/matcher-packaging/web#web-interface-http-matching-interface
ontologies are read into Python objects using the library Owlready28 and fed individually into
a semantic reasoner, which extends the ontologies by inferring semantic relations not initially
declared. The classes and properties labels extracted from both ontologies are verbalized using
their ontological context. Then, the verbalized labels are passed to the Transformer model
that obtains an embedding for each of the entities in the ontologies (i.e., a vector capturing the
semantics of the entity). Each embedding from  1 is compared to every embedding from  2
using cosine similarity, forming a bipartite graph. The Maximum Weight Bipartite Extractor
algorithm obtains an initial alignment that is reduced by a threshold filter, obtaining the final
alignment. The following paragraphs describe each of the involved techniques with some more
detail.</p>
        <sec id="sec-3-1-1">
          <title>1.2.1. Ontology Reasoning</title>
          <p>CIDER-LM performs a preliminary reasoning of both the source and the target ontologies using
HermiT OWL Reasoner9. This step expands the semantic relations in the ontologies, inferring
new knowledge not initially asserted. However, including the reasoning in the pipeline increases
the execution time of CIDER-LM considerably. This step is optional, but we kept it for OAEI-22.</p>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>8https://owlready2.readthedocs.io/en/v0.37/ 9http://www.hermit-reasoner.com/</title>
        <p>1.2.2. Verbalization of entities
CIDER-LM considers ontology classes and properties. For both types of entities, their associated
labels are extracted to build the corresponding sentence embeddings. In order to make the
calculated embedding more representative of the semantics of the analyzed ontology entity, the
labels of other entities that are part of its ontological context are considered. The verbalization
process is needed so that the language model can process the set of labels (coming from the
entities and its ontological context) as an individual sentence.</p>
        <p>In its current version, CIDER-LM builds the ontological context of an entity based on its
neighboring entities, treating classes and properties diferently (we use ‘+’ as the string concatenation
operator):</p>
        <p>Class Verbalization. For a class  and its label () , the set of parent classes and child classes
in the hierarchy of the ontology are   and   , respectively. The verbalization of the class is the
concatenation of the following:
• () + ‘, ’ ;
• for every class in   : () + ‘ i s a ’ + (   ) + ‘, ’
• for every class in   : (  ) + ‘ i s a ’ + () + ‘, ’</p>
        <p>Property Verbalization. For a property  and its label () , the sets of its domain and range
classes in the ontology are   and   , respectively. The verbalization of the property is the
concatenation of the following:
• () + ‘, ’
• for every entity in   : () + ‘ h a s d o m a i n ’ + (  ) + ‘, ’
• for every entity in   : () + ‘ h a s r a n g e ’ + (  ) + ‘, ’</p>
        <p>Supposing that a certain entity has more than one label assigned, the first label is chosen to
concatenate in the verbalization. The verbalized sentences from labels are concatenated with
particles in English, independently of the language of the ontology. During the preliminary
evaluation of the system, we found evidence that the particular language used for concatenating
the labels was not very relevant, even if it difered from the language of the ontology. This will
require further exploration in the future.</p>
        <sec id="sec-3-2-1">
          <title>1.2.3. Fine-tuned language model</title>
          <p>CIDER-LM relies on d i s t i l u s e - b a s e - m u l t i l i n g u a l - c a s e d - v 2 , which is pre-trained on Semantic
Textual Similarity (STS) and uses the SBERT architecture. The model is the knowledge-distilled
version of the Universal Sentence Encoder and supports more than 50 languages. Given a
sentence, the model obtains a vector embedding in a 512 dimensional dense vector space. A
checkpoint of the d i s t i l u s e - b a s e - m u l t i l i n g u a l - c a s e d - v 2 model has been downloaded from
HuggingFace, using the SentenceTransformers Python framework.</p>
          <p>In fact, CIDER-LM uses a fine-tuned version of the model, specialized on the task of obtaining
similarities between two entity labels in diferent languages. Given a set of pairs of entity labels
obtained from the training set and a true confidence of 1 or 0, indicating if the entities are a
match or not; the model is trained on reducing the CosineSimilarityLoss between the predicted
and true confidence . We use the SenteceTransformers framework and a basic training pipeline to
obtain the fine-tuned model used on the matching system.</p>
          <p>The cosine similarity metric is used to find the distance between the vector embeddings
associated with the entities that come from the two diferent ontologies in the common embedding
space. This ofers a measure of how similar two sentences (the verbalised sets of entity labels)
are, resulting in the confidence degree associated to the possible matching between the two
entities.</p>
        </sec>
        <sec id="sec-3-2-2">
          <title>1.2.4. Maximum Weight Bipartite Extractor</title>
          <p>CIDER-LM obtains the matching confidence for every pair of entities from the target and source
ontology. Once the confidence degrees have been determined, the alignment can be considered
as a bipartite graph with an edge (the confidence in a match) from every node (entity) from the
source ontology to the target ontology. Using the implementation of the Hungarian algorithm
in the scipy10 Python library, the maximum weight matching in bipartite graphs problem is
solved.</p>
        </sec>
        <sec id="sec-3-2-3">
          <title>1.2.5. Threshold filter</title>
          <p>After having a complete alignment, a threshold filter is used to remove all the correspondences
with low confidence from the alignment. Every correspondence with a confidence lower than
the default threshold value is removed. The threshold value can be used to direct the results
obtained by the system: a higher value will promote precision, while a lower value will favour
recall. CIDER-LM applies a default threshold value of 0.5 that, according to the results, promotes
recall.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>1.3. Adaptations made for the evaluation</title>
      <p>To participate in the OAEI campaign, we have wrapped the Python implementation of
CIDERLM around the M a t c h e r C L I Java class from the MELT framework. Wrapping enables MELT
evaluation and the packaging plugin, which is used to create the Docker container image for
submission to the OAEI.</p>
      <p>CIDER-LM performs a preliminary substitution of ‘&amp; x s d ; d a t e ’ for ‘&amp; x s d ; d a t e T i m e ’ and
‘h t t p : / / w w w . w 3 . o r g / 2 0 0 1 / X M L S c h e m a # d a t e ’ for ‘h t t p : / / w w w . w 3 . o r g / 2 0 0 1 / X M L S c h e m a # d a t e T i m e ’
because the dataset of ontologies used for the evaluation was not initially recognized by the
HermiT reasoner as OWL Version 2, which is the only version of the format taken by the
reasoner. Performing the substitution fixes the error reading the ontology.</p>
    </sec>
    <sec id="sec-5">
      <title>1.4. Link to the system and parameters file</title>
      <p>The implementation of the system is hosted in a GitHub repository11. The container image
with the matching system is also available on GitHub in the Packages section.
10https://scipy.org/
11https://github.com/sid-unizar/CIDER-LM
2. Results
In its first participation in the OAEI campaign (2022), CIDER-LM contributed to the MultiFarm
track only. The reason is that this tool is primarily aimed at cross-lingual ontology matching.
However, the method is able to produce monolingual mappings as well, so we do not discard
participation in other tracks in future OAEI editions.</p>
      <p>The MultiFarm evaluation involves two ontologies from the Conference domain (edas and
ekaw), which are translated into eight languages from English, resulting in 55 x 24 matching
tasks. Details about the test ontologies, the evaluation process, and the complete results for
the MultiFarm track can be found on the OAEI’22 website12. The results reported by the
OAEI organisers on October 9th of 2022 describe the precision, recall and F-measure of the
alignments produced by each of the participant systems. The aggregated results for completing
the matching task are shown in Table 1.
3. General comments
The following sections contain some remarks and comments on the results obtained and the
evaluation process.</p>
    </sec>
    <sec id="sec-6">
      <title>3.1. Comments on the results</title>
      <p>The obtained results in MultiFarm are intermediate in terms of F-measure (third-best result out
of five participants), and are also very good in terms of recall, having attained the best result of
any OAEI edition for the MultiFarm “diferent ontologies” sub-task 13.</p>
      <p>
        The results of CIDER-LM largely improve those obtained by its predecessor tool CIDER-CL [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]
for the MultiFarm diferent ontologies subtask, which were  = 0.16 ,  = 0.19 , and  = 0.17
      </p>
      <p>The results of CIDER-LM in the OAEI multilingual track are still modest, especially in
precision, but the fact that even the best systems do not score very high illustrates the dificulty
of the problem. For instance, the F-measure attained in MultiFarm was never higher than 0.47
in any OAEI edition.
12https://oaei.ontologymatching.org/2022/multifarm/index.html
13Only in OAEI’12 higher values can be found, but from two matchers that gave nearly every possible combination
as result, thus resulting in close to zero values of precision and F-measure
3.2. Discussions on the way to improve the proposed system
The analysis of the OAEI’22 results shows that the CIDER-LM matching system has the
potential to improve its current results in several ways. For instance, the current version builds
embeddings based on the entity labels, the labels of parents and children for classes, and the
labels of domain and range for properties. A future version could include more features in the
verbalization of labels.</p>
      <p>The current results show a clear unbalance between precision and recall. A higher threshold
value would help promote the precision measure, further achieving a higher F-measure. The
results also show that the execution time of the tool is greater than that of the other participants.
Removing the use of a reasoner would greatly reduce the time, with only a small impact on the
performance of the results, as seen in our internal experiments.</p>
      <p>Furthermore, a more careful study of the fine-tuning process could lead to an improvement
of the CIDER-LM alignment results. Involving more sophisticated techniques on the training
and validation of the fine-tuned model can reduce overfitting and provide a more general model
that behaves better with ontologies diferent from the ones seen in training.</p>
    </sec>
    <sec id="sec-7">
      <title>3.3. Comments on the OAEI procedure</title>
      <p>The MELT wrapper for the Python matching system has proven to be easy to comprehend
and implement, and useful for encapsulating the tool and later creating the web server
container image for the OAEI submission. We consider the inclusion of the MELT framework a
significant advancement in OAEI to allow the integration and participation of non-Java-based
implementations.
4. Conclusion
This paper presented the first version of CIDER-LM, which explores for the first time the
potential of multilingual language models on the task of finding cross-lingual alignments
between ontologies in diferent languages. The system uses SBERT, a multilingual language
model based on the Transformer architecture. It was evaluated on the OAEI’22 MultiFarm track,
achieving intermediate results in terms of the F-measure and very good results in terms of recall.
Although there is much room for further improvements, we consider that CIDER-LM results
have proved the viability of using multilingual language models for this task.</p>
      <p>In future versions, more features will be considered to build the ontological context, new
verbalization strategies will be analyzed, and a more careful study of the fine-tuning process will
be carried out, to attain a better and more general model for cross-lingual ontology matching.
Acknowledgments
This article is the result of a collaboration grant 2021-22 at the Department of Computer Science
and Systems Engineering, University of Zaragoza, funded by the Ministry of Education and
Professional Training (Spain). It has also been partially supported by the Engineering Research
Institute of Aragon (I3A), by the Spanish project PID2020-113903RB-I00 (AEI/FEDER, UE), by
DGA/FEDER, and by the Agencia Estatal de Investigación of the Spanish Ministry of Economy
and Competitiveness and the European Social Fund through the “Ramón y Cajal” program
(RYC2019-028112-I).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J.</given-names>
            <surname>Gracia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Asooja</surname>
          </string-name>
          ,
          <article-title>Monolingual and cross-lingual ontology matching with CIDER-CL: Evaluation report for OAEI 2013</article-title>
          ,
          <source>in: Proc. of 8th Ontology Matching Workshop (OM'13)</source>
          , at 12th
          <source>International Semantic Web Conference (ISWC'13)</source>
          , volume
          <volume>1111</volume>
          ,
          <string-name>
            <surname>CEUR-WS</surname>
          </string-name>
          ,
          <fpage>ISSN1613</fpage>
          -0073,
          <string-name>
            <surname>Syndey</surname>
          </string-name>
          (Australia),
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>P.</given-names>
            <surname>Sorg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Cimiano</surname>
          </string-name>
          ,
          <article-title>Exploiting Wikipedia for cross-lingual and multilingual information retrieval</article-title>
          ,
          <source>Data &amp; Knowledge Engineering</source>
          <volume>74</volume>
          (
          <year>2012</year>
          )
          <fpage>26</fpage>
          -
          <lpage>45</lpage>
          .
          <source>doi:1 0 . 1 0</source>
          <volume>1 6</volume>
          / j . d
          <source>a t a k . 2 0 1 2 . 0 2 . 0 0 3 .</source>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          , BERT:
          <article-title>Pre-training of Deep Bidirectional Transformers for Language Understanding, in: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Association for Computational Linguistics</article-title>
          , Stroudsburg, PA, USA,
          <year>2019</year>
          , pp.
          <fpage>4171</fpage>
          -
          <lpage>4186</lpage>
          .
          <source>doi:1 0 . 1 8</source>
          <volume>6 5 3</volume>
          / v 1 / N 1 9
          <article-title>- 1 4 2 3</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>A.</given-names>
            <surname>Vaswani</surname>
          </string-name>
          , G. Brain,
          <string-name>
            <given-names>N.</given-names>
            <surname>Shazeer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Parmar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Uszkoreit</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Jones</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. N.</given-names>
            <surname>Gomez</surname>
          </string-name>
          , Ł. Kaiser,
          <string-name>
            <surname>I. Polosukhin</surname>
          </string-name>
          , Attention Is All You Need,
          <source>in: Proc. of 31st Conference on Neural Information Processing Systems (NIPS</source>
          <year>2017</year>
          ),
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Y.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Antonyrajah</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Horrocks</surname>
          </string-name>
          ,
          <article-title>Biomedical Ontology Alignment with BERT</article-title>
          ,
          <source>in: Proc. of 16th International Workshop on Ontology Matching (OM'21)</source>
          <article-title>co-located with the 20th International Semantic Web Conference (ISWC</article-title>
          <year>2021</year>
          ), CEUR-WS,
          <year>2021</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>12</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>N.</given-names>
            <surname>Reimers</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Gurevych</surname>
          </string-name>
          ,
          <article-title>Sentence-bert: Sentence embeddings using siamese bert-networks</article-title>
          ,
          <source>in: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics</source>
          ,
          <year>2019</year>
          . URL: https://arxiv.org/abs/
          <year>1908</year>
          .10084.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>N.</given-names>
            <surname>Reimers</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Gurevych</surname>
          </string-name>
          ,
          <article-title>Making monolingual sentence embeddings multilingual using knowledge distillation</article-title>
          ,
          <source>in: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics</source>
          ,
          <year>2020</year>
          . URL: https://arxiv.org/abs/
          <year>2004</year>
          .09813.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>S.</given-names>
            <surname>Neutela</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. H. de Boerb</surname>
          </string-name>
          ,
          <article-title>Towards Automatic Ontology Alignment using BERT, in: Proc. of the AAAI 2021 Spring Symposium on Combining Machine Learning and Knowledge Engineering (AAAI-MAKE</article-title>
          <year>2021</year>
          ), CEUR-WS, Palo Alto, California, USA,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>S.</given-names>
            <surname>Hertling</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Portisch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Paulheim</surname>
          </string-name>
          , MELT
          <article-title>- matching evaluation toolkit</article-title>
          ,
          <source>in: Semantic Systems. The Power of AI and Knowledge Graphs - 15th International Conference, SEMANTiCS</source>
          <year>2019</year>
          , Karlsruhe, Germany, September 9-
          <issue>12</issue>
          ,
          <year>2019</year>
          , Proceedings,
          <year>2019</year>
          , pp.
          <fpage>231</fpage>
          -
          <lpage>245</lpage>
          .
          <source>doi:1 0 . 1 0</source>
          <volume>0 7 / 9 7 8 - 3 - 0 3 0 - 3 3 2 2 0 - 4</volume>
          \ _ 1
          <fpage>7</fpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>