<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Learning Ontology Axioms over Knowledge Graphs via Representation Learning</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Leyuan Zhao</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Xiaowang Zhang</string-name>
          <email>xiaowangzhang@tju.edu.cn</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Kewen Wang</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Zhiyong Feng</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Zhe Wang</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>College of Intelligence and Computing, Tianjin University</institution>
          ,
          <addr-line>Tianjin 300350</addr-line>
          ,
          <country country="CN">China</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>School of Information and Communication Technology, Gri th University</institution>
          ,
          <addr-line>Brisbane, QLD 4111</addr-line>
          ,
          <country country="AU">Australia</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Tianjin Key Laboratory of Cognitive Computing and Application</institution>
          ,
          <addr-line>Tianjin</addr-line>
          ,
          <country country="CN">China</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>This presents a representation learning model called SetE by modeling a predicate into a subspace in a semantic space where entities are vectors. Within SetE, a type as unary predicate is encoded as a set of vectors and a relation as binary predicate is encoded as a set of pairs of vectors. A new approach is proposed to compute the subsumption of predicates in a semantic space by employing linear programming methods to determine whether entities of a type belong to a sup-type and thus an algorithm for learning OWL axioms is developed. Experiments on real datasets show that SetE can e ciently learn various forms of axioms with high quality.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Ontology construction is a core task of ontology engineering. It has been a
research challenge in both knowledge representation and machine learning
communities. This is because ontologies are often based on logical formalisms such as
description logics (DLs), and contain more complex logical structures than graph
databases or RDF triples. DL-Learner [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] is among the rst practical systems
to learn ontological expressions, including complex DL class descriptions. Many
methods for learning new rst order formulas and rules have been developed in
Inductive Logic Programming (ILP) but they are often unable to handle very
large ontologies. Recently, some attempts have been made to e ectively learn
rules, such as [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], over KG through techniques in knowledge representation
learning, but the rules they learn are not typical ontological axioms. What is more,
conventional embedding models (e.g.,TransE, TransR, DistMult and SimplE)
mainly focus on KG completion, which only embed entities and relations without
modeling unary predictes. TransC [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] rstly di erentiate types (unary predictes)
and entities, it encodes each type as a sphere and can learn the SubClassOf
relationship between types. However, the encoding of relations and types in TransC
* Copyright 2019 for this paper by its authors. Use permitted under Creative
Commons License Attribution 4.0 International (CC BY 4.0).
is split, which prevents it from learning the relation SubPropertyOf and other
complex axioms (e.g.SubClassOf(ObjectSomeValuesFrom(P; C); D)).
      </p>
      <p>In this paper, we propose a novel uni ed embedding (called SetE) for KG
unary predicates (types) and binary predicates (relations) treating types as sets
of entities and relations as sets of entity pairs. On this basis, the subsumption
is transformed to relative position of set boundaries which can be e ciently
computed by linear programming (LP). We provide an algorithm for learning
positive OWL axioms over large-scale knowledge graphs.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Our Approach</title>
      <p>In this section, we will introduce SetE and the learning algorithm.</p>
      <p>Baguette
Quartz
..
..</p>
      <p>0.2
0.8
….</p>
      <p>….
(a) A simple illustration of the embeddings
of Baguette and Quartz.
t
f(e,t)
p</p>
      <p>g(s,p,o)
e
s</p>
      <p>o
(b) The structure of SetE
Following the same intuition, the entity pair &lt; s; o &gt; can be considered as an
instance of the relation p, so we model the fact &lt; s; p; o &gt; as follows. Where s
and o are head and tail entity of the relation p, concate(s; o) means concatenate
the two vecotrs s and o.</p>
      <p>g(s; p; o) = concate(s; o)T p =
2n
X[concate(s; o)]i [p]i
Learning Ontology Axioms over Knowledge Graphs via Embdding
To train the model, we introduce type boundary Bt 2 R. So that for all entity e
of type t, there has f (e; t) &gt; Bt; for e 2= t, there has f (e; t) &lt; Bt. The relation
boundary Br is the same. Like previous models, we generate negative samples
and use SGD to train SetE.</p>
      <p>LP to Subsumption Subsumption in KG has SubClassOf and SubPropertyOf.
We take SubClassOf as an example to show how this can be transformed into
LP under our model. The axiom SubClassOf(C; D) means that all entities that
are instances of C must be instances of D. i.e., f (e; tC ) &gt; Bt implies f (e; tD) &gt;
Bt, where tC and tD are type embeddings of C and D. So we convert this to
linear programming that computes the minimum value of f (e; tD) subject to
e 2 C (f (e; tC ) &gt; Bt). If the minimum value is greater than the boundary
Bt, that is for all entity e in type C, e always satisfy D, so we get the axiom
SubClassOf(C; D).</p>
      <p>
        Learning Ontology Axioms Based on previous analysis, we use liner
programming on embeddings to learn the following forms: A1, SubClassOf(C; D);
A2, SubPropertyOf(P; Q); A3,SubClassOf( ObjectSomeValuesFrom(P; C); D);
A4, SubClassOf(ObjectIntersectionOf( C; D); Range(F )). The algorithm learning
Sub ClassOf(C; D) is as follows.Line 3 means that if values in tC are smaller
than or equal to tD in every dimension, then we can directly get that for any
e, if f (e; tC ) &gt; Bt then f (e; tD) &gt; Bt. At last, Filter() returns axioms whose
SC(standard con dence, de ned in [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]) are greater than M inSC.
Algorithm 1 Learning SubClassOf Axioms from a KG
Input: a KG K, and two real numbers LBt and M inSC 2 [0; 1]
Output: a set O of SubClassOf axioms
1: E := SetE(K); O := ;.
2: for type embeddings tC and tD in E do
3: if (Pin=1([tC]i 6 [tD]i)?1 : 0) == n then
4: Add SubClassOf(C; D) to O
5: else if LP(tC; tD) &gt; LBt then
6: Add SubClassOf(C; D) to O
7: end if
8: end for
9: O := Filter(O; M inSC)
10: return O
3
      </p>
    </sec>
    <sec id="sec-3">
      <title>Experiments and Evaluation</title>
      <p>
        The experiment on YAGO39K aims to evaluate the e ectiveness of SetE by
comparing with the state-of-the-art model TransC[
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] in SubClassOf classi
cation. We retain four metrics: Accuracy, Precision, Recall and F1-score. TransC
was trained with the con guration in their report. SubClassOfs were removed
from the training set. To re ect real data that the negative samples far exceeds
the positive(e.g.,#negative :#positive is 226:1 in DBpedia 2016 OWL), we add
the proportion of negative samples during the experiment.
      </p>
      <p>Result in Table 1 indicates that SetE outperforms TransC and is getting
better when improving the proportion of negative samples. The Precision of
SetE is much higher than TransC (up to 87.03% for rate 1:10). It shows that
SetE is more cautious in making positive judgments, i.e., SetE distinguishes
positive samples better.
1:1
TransC 1:4</p>
      <p>1:10
1:1
1:4
1:10
57.95
50.96
48.17</p>
    </sec>
    <sec id="sec-4">
      <title>Conclusion</title>
      <p>In this paper, we present a new model SetE to speci cally represent types and
relations in a semantic space which can reduce subsumption into linear
programming. Our proposal utilizes the logical relationship to characterize the semantic
features of expressive types in learning shows certain interpretability. In the
future, we will improve the quality of expressive axioms learned and considerate
even negated axioms.</p>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgments</title>
      <p>This work is supported by the National Key Research and Development Program
of China (2017YFC0908401) and the National Natural Science Foundation of
China (61976153,61972455). Xiaowang Zhang is supported by the Peiyang Young
Scholars in Tianjin University (2019XRX-0032).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1. Buhmann, L.,
          <string-name>
            <surname>Lehmann</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Westphal</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>DL-Learner - A framework for inductive learning on the semantic web</article-title>
          .
          <source>J. Web Semant., (39)</source>
          ,
          <volume>15</volume>
          {
          <fpage>24</fpage>
          (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Ding</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Guo</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Improving knowledge graph embedding using simple constraints</article-title>
          .
          <source>In: Proc. of ACL</source>
          <year>2018</year>
          , pp.
          <volume>110</volume>
          {
          <fpage>121</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Lv</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hou</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          :
          <article-title>Di erentiating concepts and instances for knowledge graph embedding</article-title>
          .
          <source>In: Proc. of EMNLP</source>
          <year>2018</year>
          , pp.
          <year>1971</year>
          {
          <year>1979</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Omran</surname>
            ,
            <given-names>P.G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          :
          <article-title>Scalable rule learning via learning representation</article-title>
          .
          <source>In: Proc. of IJCAI</source>
          <year>2018</year>
          , pp.
          <volume>2149</volume>
          {
          <fpage>2155</fpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>