<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Learning Fine-Grained Semantics for Multi-Relational Data</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Nitisha Jain</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ralf Krestel</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Hasso Plattner Institute University of Potsdam</institution>
          ,
          <addr-line>14482 Potsdam</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The semantics of relations play a central role in the understanding and analysis of multi-relational data. Real-world relational datasets represented by knowledge graphs often contain polysemous relations between di erent types of entities, that represent multiple semantics. In this work, we present a data-driven method that can automatically discover the distinct semantics associated with high-level relations and derive an optimal number of sub-relations having ne-grained meaning. To this end, we perform clustering over vector representations of entities and relations obtained from knowledge graph embedding models.1 Copyright 2020 for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0). 1 https://hpi.de/naumann/projects/web-science/ontology-engineering.html</p>
      </abstract>
      <kwd-group>
        <kwd>relation disambiguation knowledge graph embeddings</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Relations between di erent words or phrases are important for the semantic
understanding of text. For real-world data, the relations are oftentimes
polysemous by nature, i.e., they exhibit distinct meanings in di erent contexts. Similar
to the task of word-sense disambiguation, relation disambiguation is needed to
interpret the speci c contextual semantics of relations in such cases. Relation
semantics are particularly important in the context of knowledge graphs (KGs)
that are widely used as multi-relational databases and constructed from natural
language texts, where relation polysemy occurs frequently [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. Since the
ontologies for most large scale KGs have been curated based on texts through manual
or semi-automated e orts, relations between the entities are often abstracted
for simpli cation and avoidance of redundancies. This may result in cases where
a single high-level relation serves as a generic notion between various di erent
types of KG entities and has more than one semantic meaning associated with it.
Due to the diversity of the kinds of associations between entities, abstract
relations may not be su ciently representative of the underlying semantics that they
are supposed to capture. E.g., in Yago3 [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] the majority of the relations have
multiple entity types (concepts) associated with them, with generic relations
such as owns and created exhibiting very high type plurality. Some examples of
the entity types that are associated with these relations are shown in Table 1
in the form of (head, tail) pairs. It can be seen that created is associated with
(company, computer game), as well as (painter, artwork ), despite the di erent
contexts and dissimilar entity types. If such relations are split into ne-grained
sub-relations, the precise contextual semantics as per the associated entity types
can be clearly represented. De ning ne-grained semantics between the entities
of a KG is important for facilitating several related applications such as question
answering [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] and knowledge base completion [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
      <p>
        However, due to the broad spectrum of semantic distances between the
different entity types, it is a non-trivial task to determine an optimal way to split
a relation into sub-relations, both in terms of the number of sub-relations, as
well as the entity types that the sub-relations should represent. E.g., while some
entity types are semantically similar to one another such as television and movie
for created relation, other types are quite di erent, for instance, company and
writer for created, or airport and club in the case of owns relation. We propose a
data-driven, scalable method called SemSplit that can automatically determine
an optimal con guration on a per-relation basis for a given dataset. Some
previous works have discussed the identi cation of multiple relation semantics [
        <xref ref-type="bibr" rid="ref10 ref6">6,
10</xref>
        ]. However, they either perform manual clustering or propose rigid techniques
with prede ned number of clusters for all relations across a dataset, and thus,
unlike our method, such approaches are not scalable or suitable for large and/or
dynamic datasets.
      </p>
      <p>
        The proposed method, SemSplit, leverages knowledge graph embeddings that
provide semantic representations of the KG entities and relations in a continuous
vector space. Popular KG embedding models such as TransE [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] are trained
such that the vectors h, r and t associated with a triple &lt;h; r; t&gt; satisfy h +
r = t or r = t - h. Previous work has shown that these embeddings are able
to capture relation similarity, i.e., relations having similar meanings stay close
in the embedding space [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. On the contrary, we found that relations having
multiple semantics are not well represented in the vector space, i.e., for the
triples of a relation r, the t - h vectors denoted by r, are not necessarily in
the neighborhood of the actual relation vector. This is illustrated in Figure 1
where the original relation vector r for owns (that was calculated from all its
fact triples) does not serve as the center of a single cluster for all the r vectors.
This is because the di erent type pairs connected by the same relation are quite
di erent from one another semantically, and thus, form their own clusters. In
this work, we perform clustering on the r vectors and identify an optimal
number of clusters that represent diverse semantics exhibited by the entity type
pairs connected by the relation. These clusters can be employed for splitting the
relation into several ne-grained sub-relations.
2
      </p>
    </sec>
    <sec id="sec-2">
      <title>SemSplit: Clustering Relations</title>
      <p>For every unique relation r in a given KG, the rst step is to obtain the
corresponding set of fact triples &lt;h; r; t&gt;. For every such triple, the entity types
of the head and tail entities, denoted as Th and Tt, are extracted from the KG
ontology. The triples are categorized by the unique type pairs (Th; Tt) that will
be used to derive the labels of the clusters during the clustering stage. The set
of all unique type pairs for relation r is denoted by Sr. For each type pair in Sr,
the vectors r = t h are calculated and saved for all the triples associated
with it. Here, the vectors t and h for the entities are derived from pre-trained
KG embeddings.</p>
      <p>For the clustering, we leverage the r vectors for nding suitable clusters
that will represent the di erent semantic meanings for a relation and hence derive
the sub-relations for a given relation. Since j vectors capture the semantics
of the type pairs, vectors that are close in the embedding space would convey
similar semantic meaning. A maximal splitting of the relation, with one
subrelation for every di erent type pair, would be ine cient and lead to a large
number of unnecessary sub-relations. For example, the created relation has type
pairs (artist, painting) and (artist, music) that have the same head entity type
and hence related meanings, while the type pair (company, computer game)
portrays a di erent context. The SemSplit clustering performs the challenging
task of nding an optimal number and composition of clusters Copt for the type
pairs, that can convey the distinct semantics of the relations based on the KG
triples, by combining similar type pairs while separating the dissimilar ones.</p>
      <p>
        In order to automatically determine the optimal con guration, clustering is
performed for several iterations with a varying number of clusters. It starts with
L = jSij clusters, where every cluster corresponds to a distinct type pair for
the relation, and the cluster labels are assigned accordingly. To narrow down
the search space for the optimal clusters in further iterations, SemSplit
leverages the semantic similarity of type pairs. For this, the cosine similarity scores
between all combinations of type pairs are calculated with the help of ConVec
embeddings [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] that represent the semantics of the types quite well. For each
subsequent iteration, the most similar type pairs (with ties broken arbitrarily)
are merged and assigned a combined cluster label that serves as ground truth,
(cOrePat)ed fff(((caworrtmiitspetra,,ncmyo,mocvpoiuem)tpe(ruwtgreairtmegrea,)mgte,el)fegv(,aisrfito(iwnst)r,i(tmaerre,tdisiutc,mtimo)no(avoliec)che(ahprolaalcdyteeerrr,,)gmmoovviiee))gg,, 0.14 (1) 0.29 (9) 0.49 (5)
(oSwPn)s ff((ccoommppaannyy,, caliurpbo)r(tc)o(morpgaannyi,zacotimonp,aaniyr)po(crto)ugn,tfry(s,ocvluerbe)iggn, building)g, 0.03 (1) 0.32 (6) 0.52 (3)
isA liatedTo f(artist, club)g, f(cricketer, club)g, f(player, club) (hockey player, club) 0.09 (1) 0.43 (7) 0.61 (4)
(HA) (hockey player, university) (hockey player, team)g, f(o ceholder, club)g
along with the labels of other type pairs. This process of reducing the number of
clusters and updating the labels by merging type pairs is followed until all the
type pairs have been progressively combined back together in one single cluster.
Finally, the optimal number of clusters N = jCoptj is determined based on a
cluster quality score calculated after each iteration.
3
      </p>
    </sec>
    <sec id="sec-3">
      <title>Evaluation</title>
      <p>Dataset. For this work, we have prepared a dataset that is derived from the
Yago3 knowledge base and ontology. The class types (concepts) of all entities
were extracted and the 53 most frequent concepts (having at least 10,000 entities
associated with them) were taken into consideration. Thereafter, we extracted
the fact triples that were comprised of entities associated with the chosen
concepts. This resulted in a dataset of 1,492,078 triples with 917,325 unique entities
and 31 relations.</p>
      <p>
        Cluster Quality. To analyze the performance of SemSplit, we employed several
algorithms to obtain the clusters in high dimensional vector space : Spectral (SP),
Optics (OP) and Hierarchical Agglomerative (HA) clustering. The entity and
relation vectors were obtained from pre-trained ConvE KG embeddings [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. The
quality of the clusters, and thereby, the resultant sub-relations is measured in
terms of homogeneity score [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] that favors a clustering scheme where every cluster
represents a unique dominant type. Table 2 shows examples of the optimal cluster
con gurations obtained by SemSplit with di erent algorithms. The performance
of Copt is compared to two baseline con gurations | original relation cluster
(Corig) and maximal splitting clusters (Cmax), in terms of the cluster homogeneity
scores and the number of clusters. The results indicate that while it is favorable
to split the original relation cluster into multiple sub-relation clusters, a naive
splitting leading to maximal sub-relation clusters is also not the ideal solution
for representing the ne-grained semantics. It can be seen that the Copt clusters
obtained by SemSplit perform better than both the original relation as well
as the naive maximal splitting of the relations in terms of homogeneity scores,
thus indicating the e cacy of the method for nding optimal ne-grained
subrelations.
      </p>
    </sec>
    <sec id="sec-4">
      <title>Conclusion</title>
      <p>In this paper, we have studied the need for relation disambiguation for knowledge
graphs due to the inherent relation polysemy in these datasets. We have proposed
a scalable, data-driven method SemSplit that automatically determines an
optimal con guration for deriving sub-relations with concrete semantics. First
experiments have con rmed the importance of learning ne-grained relation semantics
for real-world data and shown promising results for SemSplit performance. We
plan to perform a systematic analysis of the utility and impact of our method
on semantic tasks, such as relation extraction and question answering over KGs.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Bordes</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Usunier</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Garcia-Duran</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Weston</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yakhnenko</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          :
          <article-title>Translating embeddings for modeling multi-relational data</article-title>
          .
          <source>In: Advances in Neural Information Processing Systems</source>
          . pp.
          <volume>2787</volume>
          {
          <issue>2795</issue>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Cui</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Xiao</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Song</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hwang</surname>
            ,
            <given-names>S.w.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>W.:</given-names>
          </string-name>
          <article-title>KBQA: Learning Question Answering over QA Corpora and Knowledge Bases</article-title>
          .
          <source>Proc. of the VLDB Endowment</source>
          <volume>10</volume>
          (
          <issue>5</issue>
          ),
          <volume>565</volume>
          {
          <fpage>576</fpage>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Dettmers</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Minervini</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stenetorp</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Riedel</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Convolutional 2D knowledge graph embeddings</article-title>
          .
          <source>In: Proc. of the 32nd AAAI Conf. on Arti cial Intelligence</source>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Do</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tran</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Venkatesh</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Knowledge Graph Embedding with Multiple Relation Projections</article-title>
          .
          <source>In: Proc. of the 24th International Conf. on Pattern Recognition</source>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Jain</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kumar</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chakrabarti</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          , et al.:
          <article-title>Type-sensitive Knowledge Base Inference without Explicit Type Supervision</article-title>
          .
          <source>In: Proc. of the 56th Annual Meeting of the Association for Computational Linguistics</source>
          . pp.
          <volume>75</volume>
          {
          <issue>80</issue>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Lin</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sun</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhu</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          :
          <article-title>Learning Entity and Relation Embeddings for Knowledge Graph Completion</article-title>
          .
          <source>In: Proc. of the 29th AAAI Conference on Arti cial Intelligence</source>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Mahdisoltani</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Biega</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Suchanek</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>YAGO3:A Knowledge Base from Multilingual Wikipedias</article-title>
          .
          <source>In: Proc. of the 7th Biennial Conference on Innovative Data Systems Research</source>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Min</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shi</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grishman</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lin</surname>
            ,
            <given-names>C.Y.</given-names>
          </string-name>
          :
          <article-title>Ensemble Semantics for Large-scale Unsupervised Relation Extraction</article-title>
          .
          <source>In: Proc. of the Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning</source>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Sherkat</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Milios</surname>
            ,
            <given-names>E.E.</given-names>
          </string-name>
          :
          <article-title>Vector Embedding of Wikipedia Concepts and Entities</article-title>
          .
          <source>In: Proc. of the International Conference on Applications of Natural Language to Information Systems</source>
          . pp.
          <volume>418</volume>
          {
          <fpage>428</fpage>
          . Springer (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhuang</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Qu</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lin</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>He</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          :
          <article-title>Knowledge Graph Embedding with Hierarchical Relation Structure</article-title>
          .
          <source>In: Proc. of the 2018 Conference on Empirical Methods in Natural Language Processing</source>
          . pp.
          <volume>3198</volume>
          {
          <issue>3207</issue>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>