<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>DESKMatcher</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Bibliography</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Data and Web Science Group, University of Mannheim</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>SAP SE Product Engineering Financial Services</institution>
          ,
          <addr-line>Walldorf</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>State</institution>
          ,
          <addr-line>Purpose, General Statement</addr-line>
        </aff>
      </contrib-group>
      <abstract>
        <p>This paper describes DESKMatcher, a label-based ontology matcher. It utilizes background knowledge from the financial services and enterprise domain to better find matches in these domains. The background knowledge utilized for the enterprise domain was in the form of documentation of terms used in SAP software (textual). Therefore, Word2Vec and GloVe were used for these corpora. The Financial Industries Business Ontology (FIBO) was used as more specific background knowledge for the financial services domain. Vector space embeddings for this corpus were trained using RDF2Vec and KGloVe. Individual matchers utilizing one set of embeddings (generated from a combination of method and corpus) are pipelined together after a string-based matchers, searching only for matches between entities that have not been assigned to a match in a previous step. Results on the OAEI tracks are expected to be sub-par, because low overlap between corpus and task vocabulary is expected.3 DESKMatcher (Enterprise Domain Specific K nowledge Matcher) is an elementlevel, label-based matcher which utilizes vector space embeddings trained by applying multiple techniques on three background knowledge datasets specific to the enterprise and financial services domain, namely the Financial Industry Business Ontology (FIBO), the SAP Glossary, as well as SAP Term. The matcher was implemented for domain-specific matching in the financial services domain where classic schema matching problems are common and can be modelled as ontology matching problems [11]. However, in this paper we evaluate in how far the matcher generalizes to non-business/other domains. The matcher has not been adapted for other tasks.</p>
      </abstract>
      <kwd-group>
        <kwd>Ontology Matching</kwd>
        <kwd>Ontology Alignment</kwd>
        <kwd>Domain Specific Background Knowledge</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>1.1</p>
    </sec>
    <sec id="sec-2">
      <title>Presentation of the System</title>
      <sec id="sec-2-1">
        <title>Specific Techniques Used</title>
        <p>
          The DESKMatcher system is implemented as a matching pipeline of subsequent
matching steps using multiple domain-specific datasets that were embedded with
RDF2Vec or word2vec depending on their inherent structure. In the following a
quick introduction to the datasets used as well as to RDF2Vec will be given.
External Domain-Specific Datasets Below, we quickly introduce the sources of
background knowledge that have been used:
1. The Financial Services Business Ontology (FIBO) [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ] is used as the most
specific source of background knowledge. It is an ontology specific to the
ifnancial services domain maintained by the EDM council, with the possibility
for outside authors to contribute4. The FIBO version used contained roughly
88,000 triples with roughly 12,000 unique URIs.
2. The SAP Glossary is a textual corpus describing terms that are relevant
for SAP’s Enterprise Resource Planning (ERP) software. The resource is
not available as ontology but instead in the form of a losely structured text
corpus. The glossary was last released in 2017. The set contained definitions
for roughly 48,000 terms using roughly 14,000 unique words.
3. The SAP Term is larger than the SAP Glossary but follows the same objective.
        </p>
        <p>It is frequently updated. The resource is not available as ontology but instead
in the form of a losely structured text corpus. For this work we used the
version as of March 2020. The set contained definitions for roughly 62,000
terms using roughly 16,000 unique words.</p>
        <p>
          Embedding Approaches Used In word2vec [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] Mikolov et al. present two vector
space embedding approaches for textual corpora: Skip-Gram (SG) and continuous
bag of words (CBOW). Embeddings are generated by building a neural network
that models randomly drawn context windows given a word (SG) and vice versa
(CBOW). RDF2Vec [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ] is an embedding approach for knowledge graphs that
has already been used before in the area of ontology matching [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]. Random
walks are generated starting at each node in the knowledge graph. The set
of generated walks is then regarded as sentences and a word2vec algorithm is
applied. Thereby, a vector is obtained for each node and for each edge (that
appear in the random walks) in the knowledge graph.5 GloVe [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] is another
embedding approach for textual corpora presented 2014 by Pennington et al.
Embeddings are generated based on co-occurence probabilities of words in the
input corpus. KGloVe [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ] is an approach to generate embeddings on knowledge
graphs presented by Cochez et al. in 2017. Node “co-occurence probabilities” are
approximated in a first step, by applying a version of the Bookmark Coloring
Algorithm (BCA) [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]. The probabilities are then fed to the standard GloVe
model, which yields embeddings for each node in the graph. Embeddings for
4 see https://github.com/edmcouncil/fibo/blob/master/CONTRIBUTING.md
5 More information about RDF2Vec and its application can be found online: http:
//rdf2vec.org/
FIBO were trained using the jrdf2Vec6 [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] framework, as well as Cochez et al.’s
implemenation of their own KGloVe [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]. SAP Glossary and SAP Term were
embedded with word2vec (using the gensim7 library [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ]) and GloVe as made
available by Penningotn et al8.
        </p>
        <p>
          Configuration of Embedding approaches Skip-gram was chosen over CBOW. This
was based on Mikolov et al.’s results, that Skip-gram is better in semantic tasks
[8, p. 7], which has also been indicated in [16, p. 4]. Generally, higher dimensions
lead to higher performance, however the gain in performance per added dimension
seems to greatly decrease after 200 dimensions, wherever dimensions are reported.
Therefore the dimensions were fixed at 2009. Based on recommended parameter
settings from previous work, the window-size was fixed to 5, negative sampling
with 15 noise words and a smoothing exponent of 0.75 (as per Mikolov et al.’s
recommendation in the original paper) was used. The Skip-gram embeddings
were generated using the implementation in the gensim library [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ].
The walks required for the RDF2Vec model were generated using jRDF2Vec [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ],
while the training of the actual embeddings was conducted using gensim’s
Skipgram implementation (same as for the text corpora). The walk strategy used to
generate walks, is exactly one of the strategies proposed by Ristoski et al. in their
original paper (Breadth-first [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ]). 100 walks were generated per entity, using a
depth of 4, which lead to “sentences” with a maximum length of 12.
To generate the GloVe embeddings, the original authors’ C implementation was
used 10. For GloVe three parameters needed to be set: minCount was set to 4 in
accordance to the value used in Skip-gram. windowsize was set to 15. xmax was
set to 10 for this small corpus setting, due to the authors chosing 100 on their
large corpus [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ].
        </p>
        <p>The implementation by Cochez et al. 11, was used to generate the shufled
cooccurrence files needed as input for the final step of GloVe.</p>
        <p>Based on their results for best performance, the PageRank weighting scheme
for context generation would have been chosen, which unfortunately did not
execute without fatal errors, even after several attempts to tinker with the code.
Therefore the uniform weighting was chosen, because it was reported to be the
second best approach.</p>
        <p>For the BCA, that is used to generate the “co-occurence probabilities”,
parameters α (which probability fraction is retained on a node) and ϵ (minimum value of
probability to be distributed, values below being discarded) were chosen identical
to the number Cochez et al. chose (α = 0.1 and ϵ = 0.00001).</p>
        <p>The output co-occurence matrix was then put into GloVe using the same
parameters as above.
6 see https://github.com/dwslab/jrdf2vec
7 see https://radimrehurek.com/gensim/
8 see https://nlp.stanford.edu/projects/glove/
9 In order to add another level of consistency between the approaches, the dimensions
were also fixed to 200 in all of the other embedding generation approaches.
10 available under https://github.com/stanfordnlp/GloVe
11 https://github.com/miselico/globalRDFEmbeddingsISWC
Six embedding sets were therefore generated in total: two for each of the three
corpora.</p>
        <p>
          Matching Process Only the label and the entity type (class, datatype, property,
object property, or individual) are considered. The entity types are used as a
iflter to only be matched against each other so that a homogeneous alignment is
created, which proved to be a valuable heuristic in development. Matches are
mainly determined based on the entity label. In the first step of the pipeline,
simple matches are detected by a string matcher assuming n:m arity. Following
steps try to apply increasingly less specific background knowledge in the form of
embeddings trained on respectively less specific corpora, assuming only 1:1 arity
(by ignoring entities already appearing in predicted matches). The specificity was
assumed from the vocabulary size of a corpus. Per corpus, Word2Vec/RDF2Vec
were applied before GloVe/KGloVe embeddings12. So the embedding sets were
applied in the order FIBO-RDF2Vec, FIBO-KGloVe, SAP Glossary-Word2Vec,
SAP Glossary-Word2Vec, SAP Term-Word2Vec, SAP Term-GloVe.
Implementation The system has been implemented and packaged with the
Matching and Evaluation Toolkit (MELT), a framework for matcher development,
tuning, evaluation, and packaging [
          <xref ref-type="bibr" rid="ref10 ref4">4, 10</xref>
          ]. As the matcher heavily depends on
the python environment, the ML server module [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] of MELT has been forked
to wrap additional python code. Eventually, the system was packaged with the
framework. MELT greatly facilitated matcher development and also allowed for
an easy inclusion of correspondence-level expanations.
2
2.1
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Results</title>
      <sec id="sec-3-1">
        <title>Anatomy</title>
        <p>For this track, DESKMatcher was barely able to exceed the StringEquiv baseline
and heavily underperformed on Precision and in turn F1. Because the knowledge
to train the embeddings was not taken from the same domain, these results are
not surprising.
2.2</p>
      </sec>
      <sec id="sec-3-2">
        <title>Conference</title>
        <p>The Recall of 0.5 was rather below average compared to other matching systems,
whereas Precision and F1 were far below that of the others. An overlap between
the Conference vocabulary present in the track and Business vocabulary from
the background knowledge might have been expected, which in turn would have
caused DESKMatcher to perform better.
12 The decision whether to apply Word2Vec/RDF2Vec or GloVe/KGloVe embeddings
ifrst was taken arbitrarily. An improvement would be to investigate which embedding
approaches actually are most suited for matching tasks.</p>
      </sec>
      <sec id="sec-3-3">
        <title>Knowledge Graph</title>
        <p>
          DESKMatcher was able to perform all test cases of the knowledge graph track [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ].
In order to increase the performance, the embeddings are not used for instance
matching. With an F1 of 0.81, the matching system could outperform several
systems on this track such as all 2020 LogMap [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] matching systems. Yet the F-score
is still close to the baselineLabel matcher and below the baselineAltLabel
matcher.
3
3.1
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>General Comments</title>
      <sec id="sec-4-1">
        <title>Comments on the results (strength and weaknesses)</title>
        <p>This system uses very specific domain knowledge from the financial services and
business domains, which are not exactly covered by any of the tracks. Therefore, it
was expected, that it should not be able to perform well. Even though expectations
were set low, the results appear to be even worse. The system’s strength lies in it
being able to improve recall, which causes its greatest weakness: bad precision
that in turn leads to bad F1.
3.2</p>
      </sec>
      <sec id="sec-4-2">
        <title>Discussions on the way to improve the proposed system</title>
        <p>The greatest weakpoint of bad precision needs to be removed. Possible solutions
would be a more strict linking process. A very greedy linking approach was
chosen, to be able to find any matches at all. Additionally, the embedding sets
can be pre-evaluated in a diferent way and discarded or used accordingly; using
multiple embedding sets for one corpus did not show any positive results in the
datasets evaluated here.
4</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Conclusions</title>
      <p>In this paper, we presented the DESKMatcher, a matching system for the financial
services domain. The inner workings of the systems have been explained and the
performance numbers in the 2020 campaign of the OAEI have been discussed.
The system did not perform competitively in the campaign due to low vocabulary
overlap in the datasets that have been used. We strive to improve the system in
the future.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Berkhin</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Bookmark-coloring algorithm for personalized PageRank computing</article-title>
          .
          <source>Internet Mathematics</source>
          <volume>3</volume>
          (
          <issue>1</issue>
          ),
          <fpage>41</fpage>
          -
          <lpage>62</lpage>
          (
          <year>2006</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Cochez</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ristoski</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ponzetto</surname>
            ,
            <given-names>S.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Paulheim</surname>
          </string-name>
          , H.:
          <article-title>Global RDF Vector Space Embeddings</article-title>
          . In: d'Amato,
          <string-name>
            <given-names>C.</given-names>
            ,
            <surname>Fernandez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Tamma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            ,
            <surname>Lecue</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            ,
            <surname>Cudré-Mauroux</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            ,
            <surname>Sequeda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            ,
            <surname>Lange</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            ,
            <surname>Heflin</surname>
          </string-name>
          ,
          <string-name>
            <surname>J</surname>
          </string-name>
          . (eds.)
          <source>The Semantic Web - ISWC</source>
          <year>2017</year>
          , vol.
          <volume>10587</volume>
          , pp.
          <fpage>190</fpage>
          -
          <lpage>207</lpage>
          . Springer International Publishing,
          <string-name>
            <surname>Cham</surname>
          </string-name>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <given-names>EDM</given-names>
            <surname>Council</surname>
          </string-name>
          <article-title>: About FIBO</article-title>
          . https://edmcouncil.org/general/custom.asp?page= aboutfiboreview (
          <year>2019</year>
          ), (accessed
          <year>2020</year>
          -
          <volume>09</volume>
          -02)
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Hertling</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Portisch</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Paulheim</surname>
          </string-name>
          , H.:
          <article-title>MELT - Matching EvaLuation Toolkit</article-title>
          . In: Semantics 2019
          <string-name>
            <given-names>SEM2019</given-names>
            <surname>Proceedings. Karlsruhe</surname>
          </string-name>
          (
          <year>2019</year>
          , to appear)
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Hertling</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Portisch</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Paulheim</surname>
          </string-name>
          , H.:
          <article-title>Supervised ontology and instance matching with MELT</article-title>
          .
          <source>In: OM@ISWC</source>
          <year>2020</year>
          (
          <year>2020</year>
          ), to appear
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Hofmann</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Perchani</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Portisch</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hertling</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Paulheim</surname>
          </string-name>
          , H.:
          <article-title>Dbkwik: Towards knowledge graph creation from thousands of wikis</article-title>
          . In: Nikitina,
          <string-name>
            <given-names>N.</given-names>
            ,
            <surname>Song</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            ,
            <surname>Fokoue</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Haase</surname>
          </string-name>
          , P. (eds.)
          <source>Proceedings of the ISWC</source>
          <year>2017</year>
          <article-title>Posters &amp; Demonstrations and Industry Tracks co-located with 16th International Semantic Web Conference (ISWC</article-title>
          <year>2017</year>
          ), Vienna, Austria, October 23rd - to - 25th,
          <year>2017</year>
          .
          <source>CEUR Workshop Proceedings</source>
          , vol.
          <year>1963</year>
          .
          <article-title>CEUR-WS.org (</article-title>
          <year>2017</year>
          ), http://ceur-ws.
          <source>org/</source>
          Vol-1963/paper540.pdf
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Jiménez-Ruiz</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          :
          <article-title>Logmap family participation in the oaei 2020</article-title>
          .
          <source>OM@ISWC</source>
          <year>2020</year>
          (
          <year>2020</year>
          ), to appear
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Corrado</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dean</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          :
          <source>Eficient Estimation of Word Representations in Vector Space (Jan</source>
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Pennington</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Socher</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manning</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          : Glove:
          <article-title>Global Vectors for Word Representation</article-title>
          .
          <source>In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP)</source>
          . pp.
          <fpage>1532</fpage>
          -
          <lpage>1543</lpage>
          . Association for Computational Linguistics, Doha, Qatar (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Portisch</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hertling</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Paulheim</surname>
          </string-name>
          , H.:
          <article-title>Visual analysis of ontology matching results with the melt dashboard</article-title>
          .
          <source>In: The Semantic Web: ESWC 2020 Satellite Events</source>
          (
          <year>2020</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Portisch</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hladik</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Paulheim</surname>
          </string-name>
          , H.:
          <source>Evaluating Ontology Matchers on Real-World Financial Services Data</source>
          Models p.
          <volume>5</volume>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Portisch</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hladik</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Paulheim</surname>
          </string-name>
          , H.:
          <article-title>Rdf2vec light - A lightweight approach for knowledge graph embeddings</article-title>
          . CoRR abs/
          <year>2009</year>
          .07659 (
          <year>2020</year>
          ), https://arxiv.org/ abs/
          <year>2009</year>
          .07659
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Portisch</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Paulheim</surname>
          </string-name>
          , H.:
          <article-title>ALOD2Vec Matcher</article-title>
          .
          <source>OM@ISWC. CEUR Workshop Proceedings</source>
          (vol.
          <volume>2288</volume>
          ), pp.
          <fpage>132</fpage>
          -
          <lpage>137</lpage>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Řehůřek</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sojka</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Software framework for topic modelling with large corpora</article-title>
          .
          <source>In: Proceedings of the LREC 2010 Workshop on New Challenges for NLP Frameworks</source>
          . pp.
          <fpage>45</fpage>
          -
          <lpage>50</lpage>
          . ELRA, Valletta, Malta (May
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Ristoski</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Paulheim</surname>
          </string-name>
          , H.:
          <article-title>RDF2Vec: RDF Graph Embeddings for Data Mining</article-title>
          . In: Groth,
          <string-name>
            <given-names>P.</given-names>
            ,
            <surname>Simperl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            ,
            <surname>Gray</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Sabou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Krötzsch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Lecue</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            ,
            <surname>Flöck</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            ,
            <surname>Gil</surname>
          </string-name>
          ,
          <string-name>
            <surname>Y</surname>
          </string-name>
          . (eds.)
          <source>The Semantic Web - ISWC</source>
          <year>2016</year>
          , vol.
          <volume>9981</volume>
          , pp.
          <fpage>498</fpage>
          -
          <lpage>514</lpage>
          . Springer International Publishing,
          <string-name>
            <surname>Cham</surname>
          </string-name>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Sheikh</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Illina</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fohr</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Linares</surname>
          </string-name>
          , G.:
          <article-title>Document Level Semantic Context for Retrieving OOV Proper Names</article-title>
          .
          <source>In: 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)</source>
          . pp.
          <fpage>6050</fpage>
          -
          <lpage>6054</lpage>
          .
          <source>Proceeding of IEEE ICASSP</source>
          <year>2016</year>
          , IEEE, Shanghai, China (Mar
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>