<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>ADVIS Lab</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Dept of Computer Science</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Illinois at Chicago</institution>
          ,
          <addr-line>Chicago IL 60607</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2022</year>
      </pub-date>
      <fpage>23</fpage>
      <lpage>24</lpage>
      <abstract>
        <p>AgreementMakerDeep (AMD) is a new flexible and extensible ontology matching system. It exploits the contextual and structural information of ontologies by infusing knowledge to pre-trained masked language model, and then filter the output mappings using knowledge graph embedding techniques. AMD learns from classes and their relations between classes by constructing vector representations into the low dimensional embedding space with knowledge graph embedding methods. The results demonstrate that AMD achieves a competitive performance in many OAEI tracks, but AMD has limitations for property and instance matching.</p>
      </abstract>
      <kwd-group>
        <kwd>Ontology matching</kwd>
        <kwd>Knowledge graph embedding</kwd>
        <kwd>pre-train language model</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>https://ellenzhuwang.github.io (Z. Wang)</p>
    </sec>
    <sec id="sec-2">
      <title>2. Specific Techniques Used</title>
      <p>The architecture of AMD is shown in Fig. 1, including ontology parsing, textual matching with
large pre-train language models, knowledge graph embedding, model learning and candidate
selection.</p>
      <p>
        Ontology parsing. owlready2 [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] is used to extract meta information of classes from
the source and target ontology, such as super/sub-classes, labels, annotations, partof and
disjointwith. BeautifulSoup [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] is used to extract synonyms.
      </p>
      <p>
        Textual matching. We apply several text per-processing techniques like stop-words removal
and tokenization on class labels and annotations. AMD uses sentence-BERT [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] to compute
cosine similarity between two concept labels and annotations in unsupervised tasks. We consider
to use textual matching results as our mapping candidations.
      </p>
      <p>
        Knowledge graph embedding. We characterize the structure information of ontologies by
relations translated from one class to another class using a modified TransR [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] model into
relational embedding spaces.
2.0.1. Problem Formulation.
      </p>
      <p>Given two ontologies O and O’, we construct knowledge graph X and Y, and define the
correspondence between two concepts as following triplets  , ′ = &lt; ,  ,  ′ &gt;, where r is the relation
between c and c’. The problem is to find mapping set M =
{(  ,   ) ×  |
 ≡   }. In this study,
we focus on one-to-one alignment and the relation between concepts is equality.</p>
      <p>Let  ⃗(  )= { 1,  2, ...  } and  ⃗(  )= { 1′,  2′, ...  ′} be two d-dimensional vectors sets of size m and
n, we compute their distance with simple cosine similarity by d( ⃗(  ),  ⃗(  )) = 1-sim( ⃗(  ),  ⃗(  ))
as follows:
( ⃗(
 ),  ⃗(  )) = ∑ arg max ( ⃗(</p>
      <p>),  ⃗(  ))
=1

(1)
We define the probability of the aligned labels between concepts   and   by (  |  ) as follows:
(  |  ) =  ( ⃗(
(2)
(4)
(5)
where  is the sigmoid function.</p>
      <sec id="sec-2-1">
        <title>2.1. Masked Language Modeling</title>
        <p>
          In bio-ML track, we train on ontologies corpus for semi-supervised tasks. We use standard
transformer architecture[
          <xref ref-type="bibr" rid="ref19">19</xref>
          ] following Roberta[
          <xref ref-type="bibr" rid="ref20">20</xref>
          ]. The text encoder takes a sequence of tokens
from triples {ℎ,  , } as inputs, and computes a numbers of L layers to obtain contextualized
representations   ∈ ℝ × , where N is the number of tokens in our vocabulary and d is the
dimension.
        </p>
        <p>Concept prediction Concepts are the dominant elements in ontology matching problem,
therefore, predicting the concepts forces the model to learn the of semantic information. At the
same time, we infuse structural knowledge by triples to enable the ability of the model to learn
the contextualized representations for each ontology. Here, the concept prediction is to predict
head or tail concept, and the diference is position embedding of the masked tokens.</p>
        <p>For all the concepts  ∈  , we randomly select 30% of them to predict. And for each selected
concept   , the token   ℎ or    is replaced with the special token [MASK] in probability of 80%,
another random token in 10% and the rest remain itself. The loss of concept prediction is defined
as:</p>
        <p>++
 2 = − ∑ log  ( + | &lt;+ ) (3)</p>
        <p>+</p>
        <p>Relation prediction Relations express the way of connection head and tail concepts, and
also provide enrich hierarchy as structural information. We considers to concatenate all tokens
in  to predict, because relation labels usually have few words or tokens and are meaningless
by separated tokens. For the training tasks, relation prediction is similar process as concepts
prediction by masking randomly. Thus, the loss of relation prediction is written as:
++
 2 = − ∑ log  ( + | &lt;+ )</p>
        <p>+
Therefore, the masked language modeling loss function can now be written as,</p>
        <p>=  2 +  2
where we take a linear combination of both the loss terms.</p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Knowledge graph embedding</title>
        <p>
          In AMD, we apply a modified TransR method which translates concepts and relations into
concept space and relation-specify concept spaces, since there are multiple relations in the
ontologies e.g subclassof and disjointwith. In the original TransR, the projected vectors are
defined as   =   ,  ′ =  ′  , and the score function as   (,  ′) = ‖  +  −  ′‖22 [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ]. Inspired by
Sun et al. [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ], the absolute scores of positive triples are lower than the negative ones, so we
modify the loss function by using two  hyper-parameters as follows:
ℒ
= ∑ ∑
( )∈ (  )∈ 
(0, (
 ( ) −  1) − (  (  ) +  2))
(6)
where T denotes ℎ,  ,  and   represent negative triples,  1,  2,  &gt; 0 and  2 &gt;  1. Negative
triples are generated from negative sampling method by following AMD[
          <xref ref-type="bibr" rid="ref14">14</xref>
          ] and Multi-OM [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ].
        </p>
        <p>
          During the process that computes vectors, we need to generate negative triples. Following the
work of Sun et al. [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] and Li et al. [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ], we refine the uniform negative sampling by choosing from
the k-nearest neighbors in the embedding space, and setting constraints of select candidates
excluding from the subclassOf or disjointWith related concepts. In this way, we can avoid vector
sparsity and obtain better quality of vector representations for the concepts.
        </p>
        <p>Candidate selection We select candidates based on a threshold of the classes knowledge
graph embedding vectors similarity, and then compare the similarity with baseline if the pairs
are in baseline result sets.</p>
      </sec>
      <sec id="sec-2-3">
        <title>2.3. Parameter settings</title>
        <p>Ontology pre-processing. Ontologies always are in the format of owl or rdf, but the inputs
of masked language model and knowledge graph embedding models require the format of word
or token embeddings. Firstly, we extract meta information from ontologies using o w l r e a d y 2
1, such as ID, labels, resource, descriptions of class(or called concepts). The nature language
information of relations are extracted from restriction, property, subClass or superClass. Since
the ontologies in the tasks were developed by diferent organizations, we process the ontology
parsing from diferent tags, e.g rdf:ID=”isPartOf” and rdf:resource =”UNDEFINED_part_of”.</p>
        <p>
          MLM pre-training settings. In practice, we use RoBERTa implementation by
Huggingface2 as the base pre-trained model in our all experiments. LaKERMap is initialized with the
r o b e r t a . b a s e parameters, and the base model size is 12 layers and 768-dimensional hiddden
states( = 12,  = 768) . For the MLM training task, we use the words or tokens in knowledge
triples as our corpora for fine tuning. We select the first 5 mapping pairs from lexical matching
method in few-shot learning. Hyper-parameters are the same in [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ].
        </p>
        <p>
          KGE training settings. We use the outputs of MLM as word or token embedding in the
initialization for knowledge graph embbedding training process, as the dimension of d is set to
768. The remainder of our hyper-parameters in KGE are setup followed AMD [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ].
        </p>
        <p>The threshold for textual matching is 0.925, and the threshold for candidate selection is 0.9.</p>
      </sec>
      <sec id="sec-2-4">
        <title>2.4. Datasets</title>
        <p>We use the datasets provide by OAEI. AMD is able to be executed by organizers in four schema
matching tracks, including Conference, Anatomy, bio-ML and Common Knowledge Graph track.
However, AMD supports most of tracks in our local environment setups exclude for interactive
matching track.
1https://github.com/pwin/owlready2
2https://huggingface.co/roberta-base</p>
      </sec>
      <sec id="sec-2-5">
        <title>2.5. Adaptations made for the evaluation</title>
        <p>Our framework uses Python with Pytorch 3 and RDFLib 4, and is packed for SEALS using MELT.
We use the best parameter set in local alignments for the OAEI submission, see section 2.3.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Results</title>
      <sec id="sec-3-1">
        <title>3.1. Anatomy</title>
        <p>The Anatomy track results of AMD are shown in Table 1. In this year, AMD returns 1299
correspondences in 160 seconds. The result shows that AMD can be competitive among the top
promising matching systems. The mapping candidates generation runtime is still 3 seconds
which is same as last year. However, to improve overall performance in terms of recall and
F1-score, we conduct a filtering process with more time consuming.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Conference</title>
        <p>The Conference track results of AMD are shown in Table 1. As expected, the performance
of AMD in the conference track is not good, with the F-measure only slightly higher when
comparing baseline method(StringEquiv). AMD shows a lack of ability to extract and match
the properties in M2 and M3 evaluation variants. However, AMD has higher values in term of
Precision in most tasks. We have 0.01 improvement in term of F1-score in this year.
3.3. bio-ML</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. General comments</title>
      <sec id="sec-4-1">
        <title>4.1. Comments on the result</title>
        <p>Overall, the results show that AMD is able to complete several tasks in diferent domains on
class-level matching in a timely manner. In this year, we have improvements in anatomy and
3https://pytorch.org
4https://github.com/RDFLib
bio-ML tracks in terms of evaluation metrics. By contrast to last year, we solved memory issues
for large scale ontologies. Moreover, we consider to enable semi-supervised capability of AMD,
and it is beneficial to train on triples in intra-ontology and inter-ontology.</p>
        <p>However, AMD is still under development that it is only able to return class correspondences,
and is not able to match properties and instances in the current stage for some tracks.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Improvements</title>
        <p>The current development of AMD touches on several aspects. Besides considering properties
and instances matching, we will utilize joint embedding to combine contextualized knowledge
graph embeddings like coKE and additional knowledge resources such as WebIsA [21] as a
lexicon database. Moreover, we will adapt AMD with more diferent data types parsing and
parameters selections for diferent tracks.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusions</title>
      <p>In this paper, we have introduced an ontology matching system called AMD. In this year, we
consider to use BERT-like pre-train language model to obtain contextualized representations.
To improve the overall performance, we adapted a modified transR model to fit the ontology
matching problem: thus, we learn low-dimensional representations for each class and relation
to capture the hidden semantics of ontologies, rather than measuring the similarities between
classes directly, as in other traditional systems. AMD makes full use of the textual and structure
knowledge of ontologies. The results demonstrate the high eficiency and the promising
performance of our proposed matching method as compared to other systems results in several
tracks.
arXiv:1907.11692 (2019).
[21] J. Seitner, C. Bizer, K. Eckert, S. Faralli, R. Meusel, H. Paulheim, S. P. Ponzetto, A large
database of hypernymy relations extracted from the web., in: Proceedings of the Tenth
International Conference on Language Resources and Evaluation (LREC’16), 2016, pp.
360–367.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>I. F.</given-names>
            <surname>Cruz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Palandri Antonelli</surname>
          </string-name>
          , C. Stroe,
          <source>AgreementMaker: Eficient Matching for Large Real-World Schemas and Ontologies, PVLDB</source>
          <volume>2</volume>
          (
          <year>2009</year>
          )
          <fpage>1586</fpage>
          -
          <lpage>1589</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>I. F.</given-names>
            <surname>Cruz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Palandri Antonelli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Stroe</surname>
          </string-name>
          ,
          <article-title>Eficient Selection of Mappings and Automatic Quality-driven Combination of Matching Methods</article-title>
          , volume
          <volume>551</volume>
          <source>of CEUR Workshop Proceedings</source>
          ,
          <year>2009</year>
          , pp.
          <fpage>49</fpage>
          -
          <lpage>60</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>D.</given-names>
            <surname>Faria</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Pesquita</surname>
          </string-name>
          , E. Santos,
          <string-name>
            <given-names>M.</given-names>
            <surname>Palmonari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I. F.</given-names>
            <surname>Cruz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. M.</given-names>
            <surname>Couto</surname>
          </string-name>
          ,
          <article-title>The AgreementMakerLight Ontology Matching System</article-title>
          , in: International Conference on Ontologies,
          <source>DataBases, and Applications of Semantics (ODBASE)</source>
          , Springer,
          <year>2013</year>
          , pp.
          <fpage>527</fpage>
          -
          <lpage>541</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , Y. Qu,
          <article-title>Bootstrapping Entity Alignment with Knowledge Graph Embedding</article-title>
          , in: IJCAI, volume
          <volume>18</volume>
          ,
          <year>2018</year>
          , pp.
          <fpage>4396</fpage>
          -
          <lpage>4402</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>J.</given-names>
            <surname>Euzenat</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Shvaiko</surname>
          </string-name>
          , Ontology Matching, Springer-Verlag, Heidelberg (DE),
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>W.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Duan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <surname>G.</surname>
          </string-name>
          <article-title>Qi, Multi-view embedding for biomedical ontology matching</article-title>
          .,
          <source>OM@ ISWC</source>
          <volume>2536</volume>
          (
          <year>2019</year>
          )
          <fpage>13</fpage>
          -
          <lpage>24</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>A.</given-names>
            <surname>Laadhar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ghozzi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Megdiche</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ravat</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Teste</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Gargouri</surname>
          </string-name>
          ,
          <article-title>Partitioning and Local Matching Learning of Large Biomedical Ontologies</article-title>
          ,
          <source>in: ACM SIGAPP Symposium on Applied Computing</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>2285</fpage>
          -
          <lpage>2292</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>S.</given-names>
            <surname>Melnik</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Garcia-Molina</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Rahm</surname>
          </string-name>
          ,
          <article-title>Similarity Flooding: A Versatile Graph Matching Algorithm</article-title>
          and Its Application to Schema Matching,
          <year>2002</year>
          , pp.
          <fpage>117</fpage>
          -
          <lpage>128</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>I. F.</given-names>
            <surname>Cruz</surname>
          </string-name>
          , W. Sunna,
          <article-title>Structural Alignment Methods with Applications to Geospatial Ontologies, Transactions in GIS, Special Issue on Semantic Similarity Measurement</article-title>
          and
          <source>Geospatial Applications</source>
          <volume>12</volume>
          (
          <year>2008</year>
          )
          <fpage>683</fpage>
          -
          <lpage>711</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>P.</given-names>
            <surname>Kolyvakis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kalousis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kiritsis</surname>
          </string-name>
          , Deepalignment:
          <article-title>Unsupervised ontology matching with refined word vectors</article-title>
          ,
          <source>in: Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          , Volume
          <volume>1</volume>
          (
          <string-name>
            <surname>Long</surname>
            <given-names>Papers)</given-names>
          </string-name>
          ,
          <year>2018</year>
          , pp.
          <fpage>787</fpage>
          -
          <lpage>798</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>P.</given-names>
            <surname>Ristoski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Rosati</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Di</surname>
          </string-name>
          <string-name>
            <surname>Noia</surname>
          </string-name>
          , R. De Leone, H. Paulheim,
          <article-title>RDF2Vec: RDF Graph Embeddings Their Applications</article-title>
          ,
          <source>Semantic Web</source>
          <volume>10</volume>
          (
          <year>2019</year>
          )
          <fpage>721</fpage>
          -
          <lpage>752</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>M.</given-names>
            <surname>Cheatham</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Hitzler</surname>
          </string-name>
          ,
          <article-title>String similarity metrics for ontology alignment</article-title>
          , in: International semantic web conference, Springer,
          <year>2013</year>
          , pp.
          <fpage>294</fpage>
          -
          <lpage>309</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>J.</given-names>
            <surname>Hao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>Universal Representation Learning of Knowledge Bases by Jointly Embedding Instances and Ontological Concepts</article-title>
          ,
          <source>in: ACM SIGKDD International Conference on Knowledge Discovery &amp; Data Mining</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>1709</fpage>
          -
          <lpage>1719</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I. F.</given-names>
            <surname>Cruz</surname>
          </string-name>
          , Agreementmakerdeep results for oaei
          <year>2021</year>
          ., in: ISWC International Workshop on Ontology Matching (OM),
          <source>CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>124</fpage>
          -
          <lpage>130</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>J.-B. Lamy</surname>
          </string-name>
          , Owlready:
          <article-title>Ontology-oriented programming in python with automatic classification and high level constructs for biomedical ontologies</article-title>
          ,
          <source>Artificial intelligence in medicine 80</source>
          (
          <year>2017</year>
          )
          <fpage>11</fpage>
          -
          <lpage>28</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>L.</given-names>
            <surname>Richardson</surname>
          </string-name>
          , Beautiful soup documentation,
          <source>April</source>
          (
          <year>2007</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>N.</given-names>
            <surname>Reimers</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Gurevych</surname>
          </string-name>
          ,
          <article-title>Sentence-bert: Sentence embeddings using siamese bert-networks</article-title>
          ,
          <source>in: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics</source>
          ,
          <year>2019</year>
          . URL: http://arxiv.org/abs/
          <year>1908</year>
          .10084.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <article-title>Learning entity and relation embeddings for knowledge graph completion</article-title>
          ,
          <source>in: AAAI Conference on Artificial Intelligence</source>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>A.</given-names>
            <surname>Vaswani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Shazeer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Parmar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Uszkoreit</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Jones</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. N.</given-names>
            <surname>Gomez</surname>
          </string-name>
          , Ł. Kaiser,
          <string-name>
            <surname>I. Polosukhin</surname>
          </string-name>
          ,
          <article-title>Attention is all you need</article-title>
          ,
          <source>in: Advances in neural information processing systems</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>5998</fpage>
          -
          <lpage>6008</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ott</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Joshi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Levy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zettlemoyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Stoyanov</surname>
          </string-name>
          ,
          <article-title>Roberta: A robustly optimized bert pretraining approach</article-title>
          , arXiv preprint
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>