<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A Comprehensive Survey of Knowledge Graph Embeddings with Literals: Techniques and Applications</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Genet Asefa Gesese</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Russa Biswas</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Harald Sack</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>FIZ Karlsruhe</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Karlsruhe Institute of Technology, Institute AIFB</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Leibniz Institute for Information Infrastructure</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <fpage>31</fpage>
      <lpage>40</lpage>
      <abstract>
        <p>Knowledge Graphs are organized to describe entities from any discipline and the interrelations between them. Apart from facilitating the inter-connectivity of datasets in the LOD cloud, KGs have been used in a variety of applications such as Web search or entity linking, and recently are part of popular search systems and Q&amp;A applications etc. However, the KG applications su er from high computational and storage cost. Hence, there arises the necessity of having a representation learning of the high dimensional KGs into low dimensional spaces preserving structural as well as relational information. In this study, we conduct a comprehensive survey based on techniques of KG embedding models which consider the structured information of the graph as well as the unstructured information in form of literals such as text, numerical values etc. Furthermore, we address the challenges in their embedding models followed by a discussion on di erent application scenarios.</p>
      </abstract>
      <kwd-group>
        <kwd>Knowledge Graph</kwd>
        <kwd>Embedding</kwd>
        <kwd>Literals</kwd>
        <kwd>Knowledge Graph embedding survey</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Various Knowledge Graphs (KGs) have been published for the purpose of sharing
linked data. Some of the most popular general purpose KGs are DBpedia [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ],
Freebase [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], Wikidata [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ], and YAGO[
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. KGs have become quite invaluable
for various applications mainly in the area of arti cial intelligence. For instance,
in a more general sense, KGs can be used to support decision making process
and to improve di erent machine learning applications. Spam detection, movie
recommendation, and market basket analysis are some of the ML applications
which can bene t from KGs [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ]. General purpose KGs as e.g., Wikidata, often
comprise millions of entities, represented as nodes, with hundreds of millions of
facts, represented as edges connecting those nodes. However, a signi cant number
of important graph algorithms needed for the e cient manipulation and analysis
of graphs have proven to be NP-complete [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. Although KGs are e ective in
representing structured data, the underlying symbolic nature of the way data is
encoded as triples (i.e. &lt; subject; predicate; object &gt;) usually makes KGs hard
to manipulate [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ]. In order to address these issues and use a KG e ciently, it is
recommended to convert it into a low dimensional vector space while preserving
the graph properties. To this end, various attempts have been made so far to learn
vector representation (embeddings) for KGs. However, most of these approaches,
including the state of the art TransE [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], are structure based embeddings which
do not include any literal information. This is a major disadvantage because a
lot of information encoded in the literals will be left unused when capturing the
semantics of a certain entity.
      </p>
      <p>Literals can bring advantages to the process of learning KG embeddings in
two major ways. The rst is in learning embeddings for novel entities i.e., entities
which are not linked to any other entity in the KG but have some literal values
associated with them. In most existing structure based embedding models, it
is not possible to learn embeddings for such novel entities. However, this can
be addressed by utilizing the information held in literals to learn embeddings.
The other advantage of literals is improving the representation of entities in
structure based embedding models where an entity is required to appear in at
least minimum number of relational triples. Some approaches have been proposed
to make use of literals for KG embeddings. The focus of this paper is to discuss
these di erent embedding approaches and their advantages and drawbacks in
the light of di erent application scenarios. Our contributions include:
{ A detailed analysis of the existing literal enriched KG embedding models
and their approaches. In addition, a method is proposed to categorize them
into di erent groups.
{ The research gaps in the area of KG embeddings in using literals are
indicated as directions for further future works.</p>
      <p>The rest of this paper is organized as follows. Sect. 2 presents a brief overview
of related work. In Sect. 3, the problem formulation is provided. In Sect. 4,
the analysis of the di erent KG embedding techniques with literals is discussed.
In Sect. 5, various tasks used to evaluate the embedding models discussed in
Sect. 4 are explained. The survey is concluded in Sect. 6 by providing directions
for future work for KG embedding with literals.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <p>
        Few attempts have been made to conduct surveys on the techniques and
applications of KG embeddings [
        <xref ref-type="bibr" rid="ref12 ref24 ref3">12, 3, 24</xref>
        ]. However, none of these surveys include all
the existing KG embedding models which make use of literals. The rst survey
[
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] is conducted with focus on network embedding models. The second [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] and
the third [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ] surveys discuss only RESCAL [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] and KREAR [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] as methods
which use attributes of entities for KG embeddings, and focuses mostly on the
structure based embedding methods.
      </p>
      <p>However, RESCAL is a matrix-factorization method for relational learning
which encodes each object/data property as a slice of the tensor ending up
increasing the dimensionality of the tensor automatically. Thus, this method is
not e cient to utilize literals in KG embedding. Similarly, KREAR is not a
proper KG embedding model with literals since it takes only those data
properties which have categorical values and ignores those which take any random
literals as values. This shows that there is a gap in the KG embedding surveys.
Taking this into consideration, in this paper, a survey on KG embedding models,
which make use of literals is provided.
3</p>
    </sec>
    <sec id="sec-3">
      <title>Problem Formulation</title>
      <p>In this section, a brief introduction is provided on fundamental KG and its
embeddings followed by a formal de nition of KG embedding with literals.
3.1</p>
      <sec id="sec-3-1">
        <title>De nitions and preliminaries</title>
        <p>Relations (or Properties). Based on the nature of the objects, relations are
classi ed into two main categories:
{ Object Relations { relations that link entities to entities
{ Data Type Relations { relations that link entities to data values
(literals).The triples consisting of literals as objects are often referred to as
attributive triples.
3.2</p>
      </sec>
      <sec id="sec-3-2">
        <title>Types of literals</title>
        <p>Literals in a KG encode information that is not captured by the relations or links
between the entities. There are di erent types of literals present in the KGs:
{ Text { Wide variety of di erent information can be stored in KG in the
form of free text such as names, labels, titles, descriptions, comments, etc.
In most of the KG embedding models with literals, text information has
been further categorized into Short text and Long text for better capture
of the semantics in the model. The literals which are fairly short such as
for relations like names, titles, etc. are considered as Short text. On the
other hand, for strings that are much longer such as descriptions of entities,
comments, etc. are considered as Long text.
{ Numeric { Information encoded in the form of real numbers, decimal
numbers such as height, year or date, population, etc. also provide useful insight.
It is worth considering the numbers as distinct entities in the embeddings
models, as it has its own semantics to be covered which cannot be covered
by string distance metrics. For e. g. 777 is more similar to 788 than 77.
{ Units of Measurement { (Numeric) literals often denote units of
measurements to a de nite magnitude. For e. g. Wikidata property wdt:P2048
takes values in mm, cm, m, km, inch, foot, and pixel. Hence, discarding the
units and considering only the numeric values without normalization results
in loss of semantics, especially in the case if units are not comparable, as e.g.
units of length and units of weight.
{ Images { Images also provide latent useful information for modelling of
the entities. For example, a person's details such as age, gender etc. can be
deduced via visual analysis of an image depicting the person.
{ Others { Useful information encoded in the form of other literals such as
external URIs which could lead to an image, text, audio or video les.
Since the information present in the KGs is diverse, modelling of the entities is
a challenging task.</p>
        <p>{ RQ1 { How to combine the structured (triples with object relations) and
unstructured information (attributive triples) in the KGs into the
representation learning?
{ RQ2 { How to capture and combine the heterogeneity of the types of literals
present in the KGs into representation learning?
4</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Knowledge Graph Embeddings with Literals</title>
      <p>In this study, the investigated KG embedding models with literals are divided
into the following di erent categories based on the literals utilized: (i) Text, (ii)
Numeric, (iii) Image , and (iv) Multi-modal. A KG embedding model which
utilizes at least two types of literals is considered as multi-modal. This section
consists of an analysis of the models in each category, with their similarities and
di erences, followed by a discussion of potential drawbacks.
4.1</p>
      <sec id="sec-4-1">
        <title>Text Literals</title>
        <p>
          Subsequently, four KG models considering text literals are discussed, namely,
Extended RESCAL [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ], DKRL [
          <xref ref-type="bibr" rid="ref28">28</xref>
          ], KDCoE [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ], and KGloVe with literals [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ].
        </p>
        <p>Extended RESCAL improves the original RESCAL approach by
processing literal values more e ciently and deal with the sparsity nature of the
tensors. In this method, attributive triples are handled in a separate matrix
factorization, which is performed jointly with the tensor factorization of the
nonattributive triples. Attributive triples containing only text literals are encoded
in an entity-attributes matrix in such a way that given a triple, one or more
&lt; data type relation; value &gt; pairs are created by tokenizing and stemming the
object literal. Despite the advantage that this approach handles multi-valued
literals, it does not consider the sequence of words in the literal values.</p>
        <p>
          DKRL generates embeddings of entities and relations of a KG by
combining structure-based and description-based representations. The structure based
representation of entities and relations are obtained via TransE [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ], in which the
relation in each triple (head, relation, tail), is regarded as the translation from
head entity to tail entity. On the other hand, continuous bag of words (CBOW)
and a deep convolutional neural network model (CNN) have been used to
generate the description based representations of the head and tail entities. In case
of CBOW, short text is generated from the description based on keywords and
their corresponding word embeddings are summed up to generate the entity
embedding. In the CNN model, after preprocessing of the description, pre-trained
word vectors from Wikipedia are provided as input. The CNN has ve layers
and after every convolutional layer pooling is applied to decrease the parameter
space of CNN and lter noises. Max-pooling is applied for the rst pooling and
mean pooling for the last one. CNN model works better than CBOW because it
preserves the sequence of words.
        </p>
        <p>KDCoE focuses on the creation of an alignment between entities of
multilingual KGs by creating new inter-lingual links (ILLs). The model leverages
a weakly aligned multilingual KG for semi-supervised cross-lingual learning
using entity descriptions. It performs co-training of a multilingual KG embedding
model (KGEM) and a multilingual literal description embedding model (DEM)
iteratively in order for each model to propose a new ILL alternately. KGEM
adopts TransE whereas DEM uses an attentive gated recurrent unit encoder
(AGRU) to encode the multilingual entity descriptions.</p>
        <p>
          KGloVe with literals works by rst creating a cooccurrence matrix from
the underlying graph by performing Personalized PageRank (PPR) on the (weighted)
graph followed by the same optimisation used in the GloVe [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ] approach. Two
cooccurrence matrices are generated independently and merged in the end. The
rst matrix is generated using KGloVe [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] technique and Named Entity
Recognition is performed prior to the creation of the second matrix.
        </p>
        <p>The basic di erences between these models lie in the methods used to exploit
the information given in the text literals and combine them with
structurebased representation. One major advantage of KDCoE over text literal based
embedding models is that it considers descriptions present in multilingual KGs.
Also, both DKRL and KDCoE embedding models are designed to perform well
for the novel entities which have only attributive triples in the KGs. Other types
of text literals are not widely considered.
4.2</p>
      </sec>
      <sec id="sec-4-2">
        <title>Numeric literals</title>
        <p>
          In this section, four models which make use of numeric literals, namely,
MTKGNN [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ], KBLRN [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ], LiteralE [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ], and TransEA [
          <xref ref-type="bibr" rid="ref26">26</xref>
          ] are discussed.
        </p>
        <p>MT-KGNN trains a relational network (RelNet) for triple classi cation
and an attribute network (AttrNet) for attribute value regression in order to
learn embeddings for entities, object properties, and data properties. Only data
properties with non-discrete literal values are considered in this approach. RelNet
is a simple binary (pointwise) classi cation whereas the AttrNet is a regression
task. In RelNet, a concatenated triple is passed through a nonlinear transform
and then a sigmoid function is applied to get a linear transform. In the case of
AttrNet, two regression tasks are performed for head and tail data properties
respectively. Finally, the two networks are trained in a multi-task fashion using
a shared embedding space.</p>
        <p>
          KBLRN combines the relational, latent (learned by adapting TransE), and
numerical features together. It uses a probabilistic PoE (Product of Experts)
method to combine these feature types and train them jointly end to end. Each
relational feature is formulated by adopting the rule mining approach AMIE
+[
          <xref ref-type="bibr" rid="ref9">9</xref>
          ], to be evaluated in the KG to compute the value of the features.
Numerical features are used with the assumption that, for some relation types, the
di erences between the head and tail is seen as characteristics for the relation
itself. In PoE, one expert is trained for each (relation type, feature type) pair.
The parameters of the entity embedding model are shared by all the experts
in order to create dependencies between them. For numerical features, a radial
basis function is applied as activation function if the di erence of values is in a
speci c range.
        </p>
        <p>LiteralE is designed in order to incorporate literals into existing latent
feature models, which are designed for link prediction. Given a base model, for
instance Distmult, LiteralE modi es the scoring function f used in Distmult
by replacing the vector representations of the entities ei in f with literal enriched
representations eliit. In order to generate eliit, LiteralE uses a learnable
transformation function g which takes ei and its corresponding literal vectors li as inputs
and maps them to a new vector. For g, linear transformations, non-linear
transformations, simple multi-layer NNs, and non-linear transformations with gating
mechanisms are proposed. The modi ed scoring function f is trained following
the same procedure as in the base model.</p>
        <p>TransEA has two component models; a newly proposed attribute
embedding model and a directly adopted translation-based structure embedding model,
TransE. For the attribute embedding, it uses all attributive triples containing
numeric values as input and applies a linear regression model to learn
embeddings of entities and attributes. The loss function for TransE is de ned by taking
the sum of the respective loss functions of the component models with a
hyperparameter to assign a weight for each of the models. Finally, the two models are
jointly optimized in the training process by sharing the embeddings of entities.</p>
        <p>Despite their support for numerical literals, all the embedding methods
discussed fail to interpret the semantics behind data types of literals and units.
For e. g., `1999e ' and `the year 1999' could be considered same because type
semantics are discarded. Moreover, none of the models apply normalization for
literal values, hence the semantic similarity between two literal values such as,
200 mm and 2 cm is not captured. Also, most of the models do not have proper
mechanism to handle multi-valued literals.
4.3</p>
      </sec>
      <sec id="sec-4-3">
        <title>Image</title>
        <p>
          IKRL [
          <xref ref-type="bibr" rid="ref27">27</xref>
          ] learns embeddings by jointly training a structure-based (by adapting
TransE) with an image-based representation. For the image-based
representation, an image encoder is applied to generate embedding for each instance of a
multi-valued image relation. Attention-based multi-instance learning is used to
integrate the representations learned for each image instance by automatically
calculating the attention that should be given to each instance. Given a triple,
the overall energy function is de ned by combining four energy functions which
are based on two kinds of entity representations. The rst energy function is same
as TransE and the second uses their corresponding image-based representations
for both head and tail entities. The third function is based on the
structurebased representation of the head entity and the image-based representation of
the tail entity whereas the fourth function is the exact opposite.
4.4
        </p>
      </sec>
      <sec id="sec-4-4">
        <title>Multi-modal</title>
        <p>
          Numeric literals and text: LiteralE with blocking [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] proposes to improve
the e ectiveness of the data linking task by combining LiteralE with a CER
blocking[
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] strategy. Unlike LiteralE, it also considers literals from URI in xes
of the head entities and data relations of attributive triples. The CER blocking
is based on a two-pass indexing scheme. In the rst pass, Levenshtein distance
metric is used to process literal objects and URI in xes whereas in the second
pass semantic similarity computation with Wordnet is applied to process
object/data relations. All the extracted literals are tokenized into word lists so as to
create the indices.
        </p>
        <p>
          EAKGAE [
          <xref ref-type="bibr" rid="ref22">22</xref>
          ] jointly learns entity embeddings of two KGs using structure
embedding (by adapting TransE) and attribute character embedding. Given a
triple (h; r; a), the data property r is interpreted as a translation from the head
entity h to the literal value a i.e. h + r = fa(a) where fa(a) is a compositional
function. Three di erent compositional functions SUM, LSTM, and
N-grambased functions have been proposed. SUM is de ned as a summation of all
character embeddings of the attribute value. In LSTM, the nal hidden state
is taken as a vector representation of the attribute value. The N-gram-based
function, which shows better performance than the others, uses the summation
of n-gram combination of the attribute value.
        </p>
        <p>The common drawback with both methods is that text and numeric literals
are treated in the same way. They also do not consider literal data type semantics
or multi-valued literals in their approach. Furthermore, since EAKGAE is using
character-based attribute embedding, it fails to capture the semantics behind
the cooccurrence of syllables.</p>
        <p>
          Numeric literals, Text, and Images: MKBE [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ] is a a multi-modal
knowledge graph embedding, in which the text, numeric and image literals are
modelled together. It extends DistMult, which creates embedding for entities and
relations, by adding neural encoders for di erent data types. For image triples,
a xed-length vector is encoded using CNN. On the other hand, textual
attributes are encoded using sequential embedding approaches like LSTMs. Given
the vectors representations of the entities, relations and attributes, the same
scoring function from DistMult is used to determine the correctness probability
of triples.
5
        </p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Applications</title>
      <p>In this section, the di erent KG application scenarios used by the techniques
discussed in Sect. 4 are presented.</p>
      <p>Link prediction. Link prediction aims to predict new links for a KG given
the existing links among the KG entities. The models Extended RESCAL,
LiteralE, TransEA, KBLRN, DKRL, KDCoE, EAKGAE, IKRL, and MKBE have</p>
      <p>FB15K FB15K-237 YAGO-10</p>
      <p>KBLN 0.739 0.301 0.487
MTKGNN(DistMult+MultiTask) 0.669 0.285 0.481</p>
      <p>DistMult+LiteralE 0.583 0.314 0.504
DistMult+LiteralE-Gate 0.723 0.300</p>
      <p>ComplEx+LiteralE 0.765 0.299 0.509</p>
      <p>ConvE+LiteralE 0.66 0.314 0.506
Model MRR</p>
      <p>Numeric
KBLN 0.503
S+N 0.549</p>
      <p>Image
IKRL 0.509</p>
      <p>S+I 0.566
been evaluated on the link prediction task. However, it is not possible to compare
the obtained evaluation results because the experiments have been carried out
on di erent datasets. The authors of the LiteralE and MKBE models conducted
some experiments to compare their proposed models/submodels with already
existing ones. LiteralE has been compared with KBLN, which is a submodel of
KBLRN designed without taking into consideration the relational information of
graph feature methods. Besides KBLN, LiteralE has been compared with a new
modi ed version of MTKGNN, where its ER-MLP part is replaced with
DistMult to make it compatible with their speci c implementation environment. The
results taken from LiteralE are shown in Table 1a. From the result, it can be
seen that DistMult+LiteralE delivers better MRR values when it is compared
with both KBLN and MTKGNN on the datasets FB15k-237 and YAGO-10. The
authors of LiteralE argue that the performance of DistMult+LiteralE is lower
than the others on the FB15K dataset because this dataset includes a lot of
inverse relations and hence claim that it is not an appropriate dataset for link
prediction. The other experiments conducted are in MKBE where the submodels,
structures along with numeric and image literals respectively arecompared with
KBLN and IKRL respectively as shown in Table 1b. Thereby, it can be inferred
that the two submodels of MKBE perform better than their counterparts.</p>
      <p>Triple Classi cation. A potential triple is classi ed as 0 (false) or 1 (true).
MTKGNN, KGlove with literals, and IKRL have been evaluated on this task.
Since they do not use a common evaluation dataset, it is not possible to compare
the reported results directly.</p>
      <p>Entity Classi cation. Given a KG and an entity, the entity type is
predicted using a multilabel classi cation algorithm with KG entity types as given
classes. DKRL has been evaluated on this task.</p>
      <p>Entity Alignment. Semantically similar entities are determined from
multiple KGs using speci c similarity metrics. EAKGAE has been evaluated on
an entity alignment task. In addition, KDCoE has also been evaluated on a
cross-lingual entity alignment task which determines similar entities in di erent
languages. Despite the fact that both these models use the same task for
evaluation, their experimental results cannot be compared since they are based on
di erent datasets.</p>
      <p>Other Machine Learning problems. Attribute value prediction,
nearest neighbor analysis, data linking, and document classi cation are other
applications scenarios used for the evaluation of the models discussed in Sect. 4.
In MTKGNN, attribute value prediction is applied using an attribute-speci c
Linear Regression classi er for evaluation. Nearest neighbor analysis has been
performed in LiteralE to compare DistMult+LiteralE with the base model
DistMult. Data linking and document classi cation tasks have been used in LiteralE
with blocking and KGlove with literals respectively.
6</p>
    </sec>
    <sec id="sec-6">
      <title>Conclusion and Future Directions</title>
      <p>To sum up, in this paper, a comprehensive survey of KG embedding models with
literals is presented. The survey provides a detailed analysis and categorization
of the embedding techniques of these models along with their application
scenarios and limitations. As mentioned in Section 4, these embedding models have
di erent drawbacks. None of them consider the e ect that data types and units
have on the semantics of literals. Most of them also do not have a proper
mechanism to handle multi-valued literals. Thus, lling these gaps will be taken as a
direction for future work.</p>
      <p>Moreover, only few approaches have been proposed for multi-modal KG
embeddings and none of them take into consideration literals with URIs linking to
items such as audio, video, or pdf les. This clearly indicates that more work has
to be invested to address di erent types of literals. Regarding the comparison of
the quality of the models, as discussed in Section 5, it was only possible to use
the experimental results conducted for some of the models as most use di erent
datasets and application scenarios. However, as a future work, experiments for
all of the models on di erent applications will be performed to enable better
comparability.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Bollacker</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Evans</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Paritosh</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sturge</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Taylor</surname>
          </string-name>
          , J.:
          <article-title>Freebase: A Collaboratively Created Graph Database for Structuring Human Knowledge</article-title>
          .
          <source>In: ACM SIGMOD international conference on Management of data</source>
          (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Bordes</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Usunier</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Garcia-Duran</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Weston</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yakhnenko</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          :
          <article-title>Translating Embeddings for Modeling Multi-Relational Data</article-title>
          . In: NIPS (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Cai</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zheng</surname>
            ,
            <given-names>V.W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chang</surname>
          </string-name>
          , K.C.C.
          <article-title>: A Comprehensive Survey of Graph Embedding: Problems, Techniques, and</article-title>
          <string-name>
            <surname>Applications. TKDE</surname>
          </string-name>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tian</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chang</surname>
            ,
            <given-names>K.W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Skiena</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zaniolo</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Co-training Embeddings of Knowledge Graphs and Entity Descriptions for Cross-Lingual Entity Alignment</article-title>
          . arXiv preprint arXiv:
          <year>1806</year>
          .
          <volume>06478</volume>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Cochez</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Garofalo</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Len</surname>
            <given-names>en</given-names>
          </string-name>
          , J.,
          <string-name>
            <surname>Pellegrino</surname>
            ,
            <given-names>M.A.</given-names>
          </string-name>
          :
          <article-title>A First Experiment on Including Text Literals in KGloVe</article-title>
          . arXiv preprint arXiv:
          <year>1807</year>
          .
          <volume>11761</volume>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Cochez</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ristoski</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ponzetto</surname>
            ,
            <given-names>S.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Paulheim</surname>
          </string-name>
          , H.:
          <article-title>Global rdf Vector Space Embeddings</article-title>
          . In: International Semantic Web Conference. Springer (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>de Assis Costa</surname>
            , G., de Oliveira,
            <given-names>J.M.P.:</given-names>
          </string-name>
          <article-title>A Blocking Scheme for Entity Resolution in the Semantic Web</article-title>
          . In: AINA (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>de Assis Costa</surname>
            , G., de Oliveira,
            <given-names>J.M.P.</given-names>
          </string-name>
          :
          <article-title>Towards Exploring Literals to Enrich Data Linking in Knowledge Graphs</article-title>
          . In: AIKE (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Galarraga</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Te ioudi</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hose</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Suchanek</surname>
            ,
            <given-names>F.M.</given-names>
          </string-name>
          :
          <article-title>Fast Rule Mining in Ontological Knowledge Bases with AMIE+</article-title>
          .
          <source>VLDB</source>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Garc</surname>
            a-Duran,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Niepert</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          : Kblrn:
          <article-title>End-to-End Learning of Knowledge Base Representations with Latent, Relational, and Numerical Features</article-title>
          . In: UAI (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Garey</surname>
            ,
            <given-names>M.R.</given-names>
          </string-name>
          , Johnson, D.S.:
          <article-title>Computers and Intractability; A Guide to the Theory of NP-Completeness</article-title>
          .
          <string-name>
            <given-names>W. H.</given-names>
            <surname>Freeman</surname>
          </string-name>
          &amp; Co., New York, NY, USA (
          <year>1990</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Goyal</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ferrara</surname>
          </string-name>
          , E.: Graph Embedding Techniques, Applications, and
          <string-name>
            <surname>Performance</surname>
            :
            <given-names>A</given-names>
          </string-name>
          <string-name>
            <surname>Survey. Knowl</surname>
          </string-name>
          .-Based
          <string-name>
            <surname>Syst</surname>
          </string-name>
          . (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Kristiadi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Khan</surname>
            ,
            <given-names>M.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lukovnikov</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lehmann</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fischer</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Incorporating Literals into Knowledge Graph Embeddings</article-title>
          .
          <source>CoRR</source>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Lehmann</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Isele</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jakob</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jentzsch</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kontokostas</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mendes</surname>
            ,
            <given-names>P.N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hellmann</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Morsey</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Van Kleef</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Auer</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          , et al.:
          <article-title>Dbpedia{A Large-scale, Multilingual Knowledge Base Extracted from Wikipedia</article-title>
          . Semantic
          <string-name>
            <surname>Web</surname>
          </string-name>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Lin</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sun</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Knowledge Representation Learning with Entities, Attributes and Relations</article-title>
          . ethnicity (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Mahdisoltani</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Biega</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Suchanek</surname>
            ,
            <given-names>F.M.:</given-names>
          </string-name>
          <article-title>Yago3: A Knowledge Base from Multilingual Wikipedias</article-title>
          . In: CIDR (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Nickel</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tresp</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kriegel</surname>
            ,
            <given-names>H.P.:</given-names>
          </string-name>
          <article-title>A Three-Way Model for Collective Learning on Multi-Relational Data</article-title>
          . In: ICML (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Nickel</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tresp</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kriegel</surname>
            ,
            <given-names>H.P.</given-names>
          </string-name>
          :
          <article-title>Factorizing Yago: Scalable Machine Learning for Linked Data</article-title>
          .
          <source>In: Proceedings of the 21st international conference on World Wide Web. ACM</source>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Pennington</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Socher</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Manning</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Glove: Gobal Vectors for Word Representation</article-title>
          . In: EMNLP (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Pezeshkpour</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Singh</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Embedding multimodal relational data for knowledge base completion</article-title>
          . arXiv preprint arXiv:
          <year>1809</year>
          .
          <volume>01341</volume>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>Tay</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Luu</surname>
            ,
            <given-names>A.T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Phan</surname>
            ,
            <given-names>M.C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hui</surname>
          </string-name>
          , S.C.
          <article-title>: Multi-task Neural Network for Nondiscrete Attribute Prediction in Knowledge Graphs</article-title>
          .
          <source>CoRR</source>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22.
          <string-name>
            <surname>Trsedya</surname>
            ,
            <given-names>B.D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Qi</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          , Zhang, R.:
          <article-title>Entity Alignment between Knowledge Graphs Using Attribute Embeddings</article-title>
          . In: AAAI (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <surname>Vrandecic</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          , Krotzsch, M.:
          <string-name>
            <surname>Wikidata</surname>
            :
            <given-names>A Free</given-names>
          </string-name>
          <string-name>
            <surname>Collaborative Knowledge Base</surname>
          </string-name>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          24.
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mao</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Guo</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Knowledge Graph Embedding: A Survey of Approaches and Applications</article-title>
          .
          <source>TKDE</source>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          25.
          <string-name>
            <surname>Wilcke</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bloem</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>De Boer</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          :
          <article-title>The Knowledge Graph as the Default Data Model for Learning on Heterogeneous Knowledge. Data Science (</article-title>
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          26.
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          :
          <article-title>Knowledge Graph Embedding with Numeric Attributes of Entities</article-title>
          . In: Rep4NLP@ACL (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          27.
          <string-name>
            <surname>Xie</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chua</surname>
            ,
            <given-names>T.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Luan</surname>
            ,
            <given-names>H.B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sun</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Image-embodied knowledge representation learning</article-title>
          .
          <source>In: IJCAI</source>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          28.
          <string-name>
            <surname>Xie</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jia</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Luan</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sun</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Representation Learning of Knowledge Graphs with Entity Descriptions</article-title>
          . In: AAAI (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>