<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Leveraging Multilingual Descriptions for Link Prediction: Initial Experiments</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Genet Asefa Gesese</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mehwish Alam</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Fabian Hoppe</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Harald Sack</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>FIZ Karlsruhe</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Karlsruhe Institute of Technology, Institute AIFB</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Leibniz Institute for Information Infrastructure</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>In most Knowledge Graphs (KGs), textual descriptions of entities are provided in multiple natural languages. Additional information that is not explicitly represented in the structured part of the KG might be available in these textual descriptions. Link prediction models which make use of entity descriptions usually consider only one language. However, descriptions given in multiple languages may provide complementary information which should be taken into consideration for the tasks such as link prediction. In this poster paper, the bene ts of multilingual embeddings for incorporating multilingual entity descriptions into the task of link prediction in KGs are investigated.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Introduction
Various Knowledge Graphs (KGs) such as DBpedia and Wikidata have been
published to share linked data and have been crucial for many tasks. However,
according to the Open World Assumption, KGs are never complete. Due to this
fact, di erent KG completion models which map KGs to a low dimensional
vector space based on the task of link prediction have been proposed. However,
only some of these models such as DKRL [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], MKBE[
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] , Jointly[
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] , SSP [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]
, and LiteralE [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] leverage the textual descriptions of entities for the link
prediction task [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. Furthermore, most popular KGs contain descriptions in two or
more languages for a single entity due to the multilingual community working
on these KGs (as in Wikidata) or the multilingual nature of its sources (as in
DBpedia). The cultural context and bias associated with each of these
descriptions induces a di erence with regards to content. However, despite the fact that
entity descriptions are available in multiple natural languages, all the existing
models including DKRL consider only one language. Figure 1 presents an
example scenario showing the di erences in contents of multilingual descriptions of a
single entity. In this example, the description in German contains information
which does not appear in the English or French descriptions. For instance, the
fact that the team is the record winner of the U-19 Asian Cup with twelve titles
is only mentioned in the German description.
      </p>
      <p>Copyright © 2020 for this paper by its authors. Use permitted under Creative
Commons License Attribution 4.0 International (CC BY 4.0).</p>
      <p>
        This paper presents part of an ongoing work which is an empirical evaluation
of an already published position paper [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Speci cally, in this poster paper, the
performance of the existing model DKRL in leveraging multilingual descriptions
using multilingual embeddings has been analysed and the results of the initial
experiments are discussed.
"Die sudkoreanische U-20-Fu ballnationalmannschaft ist ... bei
U-20Weltmeisterschaften und U-19-Asienmeisterschaften..."@de
/m/05f5sr9 ("South Korea
national under-20 football team")
"L'equipe de Coree du Sud de
football des moins de 20 ... de la
Federation de Coree du Sud de
football (KFA).@fr).
"The Korea Republic national
under-20 football team
represents South Korea in international
youth football competitions."@en
As discussed in [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], an entity alignment model named KDCoE [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] has
demonstrated the advantage of multilingual word embeddings by using a cross-lingual
Bilbowa word embedding [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] to encode multilingual descriptions for the task of
cross lingual learning. In this paper, the same approach is adopted to encode
multilingual entity descriptions for a link prediction task on a monolingual dataset.
In particular, the experiments have been performed with one of the existing
models DKRL [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] using pretrained multilingual word embeddings by MUSE3.
DKRL is an extension of TransE [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], which learns two kinds of vector
representations for an entity, i.e., structure-based and description-based representations.
DKRL adopts TransE for the structure-based representation and uses CNN to
encode entity descriptions for the description-based representations. These two
kinds of entity representations are learned simultaneously into the same vector
space without forcing them to be uni ed. In our experiments, in order to
effectively utilize multilingual descriptions, the embeddings of the words in the
descriptions obtained by MUSE are passed as inputs to the encoder.
      </p>
      <p>MUSE has been chosen because this paper deals with multilingual
descriptions and MUSE aligns embeddings (speci cally, FastText embeddings) of words</p>
    </sec>
    <sec id="sec-2">
      <title>3 https://github.com/facebookresearch/MUSE</title>
      <p>
        in di erent languages into the same vector space. Figure 2 shows the CNN
encoder part of DKRL with pretrained word embeddings from multilingual
descriptions as inputs.
In this section, the experiments conducted to incorporate textual descriptions
in English, French, and German into DKRL (for the task of link prediction)
are presented. Table 1 shows the dataset created out of FB15K-237 [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] for the
experiments by removing those triples for which either the head or tail entity
does not have descriptions in at least one of the three languages mentioned
above or have less than 3 words after preprocessing. Since FB15K-237 is a
dataset generated from Freebase and the entity descriptions in Freebase are
old, the descriptions in all the three languages have been constructed by
taking the information from the summary part of their respective Wikipedia pages.
During preprocessing, stop words are removed and all phrases are marked
using entity names and also by applying Spacy's4 named entity recognizer. The
created dataset is available at
https://github.com/ISE-FIZKarlsruhe/LinkPrediction-with-Multilingual-Entity-Descriptions.
      </p>
      <p>For the experiments with DKRL and TransE, the code published at https:
//github.com/xrb92/DKRL by the authors of the DKRL paper and the code</p>
    </sec>
    <sec id="sec-3">
      <title>4 https://spacy.io/</title>
      <p>for TransE available at https://github.com/thunlp/OpenKE has been used
respectively. The DKRL model has been trained on three varieties of the dataset,
given the names DKRLe, DKRLeg, and DKRLegf . For DKRLe, only English
descriptions are used whereas for DKRLeg the combination of descriptions in
German and English are used. On the other hand, descriptions in all the three
languages are used to train DKRLegf . The minimum, maximum, and average
number of words are 3, 615, and 107.351 for the descriptions in DKRLe, 9, 970,
and 140.591 in DKRLeg, and 18, 1460, and 192.091 in DKRLegf respectively.</p>
      <p>
        In DKRLe, the words are initialized using FastText pretrained embeddings
and for the other two models MUSE pretrained embeddings are used. As shown
in Table 2, TransE [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] has also been trained on the new dataset for fair
comparison with DKRL. The hyperparameters are chosen from embedding size f50, 100,
150g, margin f1.0, 2.0, 3.0, 4.0, 5.0g, learning rate f0.01, 0.1, 1.0g (following the
same procedure as in the paper of TransE). The optimal parameters for TransE
on this dataset is embedding size: 100, margin: 4.0, learning rate: 0.1, and epoch:
1000. For all the other models DKRLe, DKRLeg, and DKRLegf , the same
procedure as in the original study DKRL has been adopted. The optimal parameters
are entity and relation embedding size: 100, learning rate 0.001, margin: 1.0,
window-size: 2, dimension of feature map: 100, and word embedding size: 300,
for all the three varieties. DKRLe is trained for 1000 epochs where as DKRLeg
and DKRLegf are trained for 1200 epochs.
      </p>
      <p>As shown in Table 2, incorporating descriptions into the link prediction task
brings improvement over TransE. However, when comparing the di erent
varieties of DKRL, it is seen that combining multiple descriptions has only a slight
improvement. For instance, hits@10 is the same for DKRLeg and DKRLegf . One
potential reason for such results could be the out of vocabulary words in the
pretrained word embeddings by MUSE. There are 18.4% and 20% out of vocabulary
words for DKRLeg and DKRLegf respectively and they are randomly initialized.
In this paper, which is based on an already published position paper, preliminary
results from an ongoing work to discuss the bene ts of leveraging multilingual
embeddings for the task of link prediction are presented. As a future work,
more experiments will be conducted by aligning pretrained FastText embeddings
which have bigger vocabulary size, using MUSE, to avoid the problem which
rises due to out of vocabulary words. Moreover, another way to improve the
results will be investigated which is to learn description-based embeddings of an
entity separately for each language and then fusing the vectors using di erent
approaches.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Bordes</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Usunier</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Garcia-Duran</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Weston</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yakhnenko</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          :
          <article-title>Translating Embeddings for Modeling Multi-Relational Data</article-title>
          . In: NIPS (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tian</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chang</surname>
            ,
            <given-names>K.W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Skiena</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zaniolo</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Co-training embeddings of knowledge graphs and entity descriptions for cross-lingual entity alignment</article-title>
          . arXiv preprint arXiv:
          <year>1806</year>
          .
          <volume>06478</volume>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Gesese</surname>
            ,
            <given-names>G.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Alam</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sack</surname>
          </string-name>
          , H.:
          <article-title>Semantic entity enrichment by leveraging multilingual descriptions for link prediction</article-title>
          .
          <source>In: DL4KG@ ESWC</source>
          (
          <year>2020</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Gesese</surname>
            ,
            <given-names>G.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Biswas</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Alam</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sack</surname>
          </string-name>
          , H.:
          <article-title>A survey on knowledge graph embeddings with literals: Which model links better literal-ly? arXiv preprint</article-title>
          arXiv:
          <year>1910</year>
          .
          <volume>12507</volume>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Gouws</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bengio</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Corrado</surname>
          </string-name>
          , G.:
          <article-title>Bilbowa: Fast bilingual distributed representations without word alignments</article-title>
          . In: ICML (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Kristiadi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Khan</surname>
            ,
            <given-names>M.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lukovnikov</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lehmann</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fischer</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Incorporating literals into knowledge graph embeddings</article-title>
          .
          <source>In: International Semantic Web Conference</source>
          . pp.
          <volume>347</volume>
          {
          <fpage>363</fpage>
          . Springer (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Pezeshkpour</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Singh</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Embedding multimodal relational data for knowledge base completion</article-title>
          .
          <source>In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing</source>
          . pp.
          <volume>3208</volume>
          {
          <fpage>3218</fpage>
          .
          <article-title>Association for Computational Linguistics (Oct-Nov</article-title>
          <year>2018</year>
          ), https://www.aclweb.org/anthology/D18- 1359
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Toutanova</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Observed versus latent features for knowledge base and text inference</article-title>
          .
          <source>In: 3rd Workshop on Continuous Vector Space Models and their Compositionality</source>
          .
          <source>Association for Computational Linguistics</source>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Xiao</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Meng</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhu</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          :
          <article-title>Ssp: semantic space projection for knowledge graph embedding with text descriptions</article-title>
          .
          <source>In: Thirty-First AAAI Conference on Arti cial Intelligence</source>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Xie</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jia</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Luan</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sun</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Representation learning of knowledge graphs with entity descriptions</article-title>
          .
          <source>In: AAAI</source>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Xu</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Qiu</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          :
          <article-title>Knowledge graph representation with jointly structural and textual encoding</article-title>
          . pp.
          <volume>1318</volume>
          {
          <issue>1324</issue>
          (08
          <year>2017</year>
          ). https://doi.org/10.24963/ijcai.
          <year>2017</year>
          /183
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>