<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Look beyond the Surface: A Demo for Explaining Knowledge Graph Embeddings and Entity Similarity</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Huu Tan Mai</string-name>
          <email>huu.mai@telecom-paris.fr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Youmna Ismaeil</string-name>
          <email>youmna.ismaeil@de.bosch.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Trung-Kien Tran</string-name>
          <email>trungkien.tran@de.bosch.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Hendrik Bloeckeel</string-name>
          <email>hendrik.blockeel@kuleuven.be</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Daria Stepanova</string-name>
          <email>daria.stepanova@de.bosch.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="editor">
          <string-name>Explainable Entity Similarity, Knowledge Graphs, Knowledge Graph Embeddings</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Bosch Center for Artificial Intelligence</institution>
          ,
          <addr-line>Robert Bosch Campus 1, 71272 Renningen</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Department of Computer Science, KU Leuven</institution>
          ,
          <addr-line>Leuven BE 3000</addr-line>
          ,
          <country country="BE">Belgium</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Télécom Paris</institution>
          ,
          <addr-line>Palaiseau</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Knowledge Graph embedding (KGE) methods are concerned with mapping entities and relations in a KG into a low-dimensional vector space. KGEs have been efectively used for a variety of tasks such as link prediction, and entity classification or entity similarity. However, these methods are often considered as black boxes, providing users with no insights into the information captured by the embeddings and justifications for the computed outcome on a particular task. Recently, FeaBI, a framework for interpreting pre-computed entity embeddings relying on entity neighborhoods, has been proposed. In this paper we present a demo for this work. Our intuitive and interactive demo allows users to conveniently exploit the respective framework for computing embedding-based similarity between KG entities as well as generating and visualizing explanations for the respective similarity.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>CEUR
ceur-ws.org</p>
    </sec>
    <sec id="sec-2">
      <title>1. Introduction</title>
      <p>
        Knowledge Graph embeddings (KGEs) (see, e.g., [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]) represent entities and relations in a
lowdimensional vector space. They have been useful in a range of tasks, including link prediction
(e.g., [
        <xref ref-type="bibr" rid="ref2 ref3 ref4">2, 3, 4</xref>
        ]), entity classification (e.g., [
        <xref ref-type="bibr" rid="ref5 ref6">5, 6</xref>
        ]) or entity similarity. However, despite their success,
KG embeddings are often regarded as black boxes. Lack of transparency and interpretability of
KGEs limits users’ understanding of their inner mechanisms, and undermines the trust in these
models. E.g., given an entity, embedding-based suggestions regarding other entities similar to it
might be less convincing if the user cannot examine the reasons behind the similarities.
      </p>
      <p>
        Recently, a framework named FeaBI [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] has been proposed for explaining pre-computed
entity embeddings. More specifically, given a KG and its embedding model FeaBI employs
embedded feature selection techniques to extract from the KG propositional features in the form
of relations and entities that are important for a given KG embedding model. These features are
treated as KG embedding model explanations. FeaBI can be conveniently used for explaining
similarities between entities.
      </p>
      <p>Training Embedding Models
e.g., TransE, CompGCN, NodePiece,</p>
      <p>SNoRE entity embeddings</p>
      <p>KG Feature Generation
Initial feature vectors for entity embeddings</p>
      <p>Feature Selection
Recounssitnrugcitnioitnialoffeeantutirtye evmecbtoerdsdings</p>
    </sec>
    <sec id="sec-3">
      <title>2. Demo Overview</title>
      <p>Model</p>
      <p>TransE
CompGCN
NodePiece</p>
      <p>SNoRe</p>
      <p>TransE
CompGCN
NodePiece</p>
      <p>SNoRe</p>
      <p>
        FeaBI total runtime (s)
10.39 ± 0.37s
9.28 ± 0.31s
9.96 ± 0.19s
16.94 ± 0.20s
30.89 ± 2.02s
26.45 ± 1.0s
30.53 ± 0.82s
33.01 ± 0.68s
Feabi (Backend). For a given KG and its embedding model, FeaBI computes KG embedding
explanations defined as a list of KG features ranked based on their importance for the generation
of the KG embedding. The top most important features are then used to build interpretable
representations of the KG entity embeddings. The main components of FeaBI are KG embedding
training, feature construction and feature selection (see [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] for details). The training of the KG
embedding model is naturally the most time-consuming step, which typically takes up to 5
hours (e.g., for CompGCN on FB15K237 dataset). Therefore, in our demo we provide a number
of pre-trained embedding models. At the moment we support 4 popular embedding models:
TransE [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], CompGCN [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], NodePiece [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] and SNoRe [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], but other pretrained embeddings
can also be provided by users as illustrated in Figure 3.
      </p>
      <p>Table 1 shows the running time of the feature construction and feature selection steps of
FeaBI for two popular KGs and embedding models available in the demo.
Webservice. The webservice handles the communication of FeaBI with the frontend. In the
frontend, the KG and KG embedding models are first selected by the user, and then passed to
FeaBI via the webservice. Subsequently, FeaBI computes the results, which are then sent to the
webservice and presented to the user via the frontend.</p>
      <p>Frontend. The frontend allows users to conveniently explore the model explanations for a
given embedding model, entity embedding explanations, as well as explanations for similarities
between a pair of selected entities retrieved by the webservice.</p>
      <p>The workflow of the demo proceeds as follows. First, the user selects a KG and an
embedding model from the provided list (or uploads custom ones) via the visual interface. Then, a
model explanation (i.e., a list of symbolic features ranked by their importance) is automatically
generated and presented to the user (see Figure 4).</p>
      <p>Additionally, the demo ofers a possibility to compare entities in the KG in terms of their
similarity relying on the given embedding model. As shown in Figure 5, for a given entity
provided by the user, similar entities can be retrieved based on the distance metric in the
embedding space (cosine similarity and Euclidean distance are currently supported). The user
can select any pair of entities and use the system to generate explanations for their similarity, i.e.,
a list of selected KG features that the entities share along with their graph-based visualizations.</p>
    </sec>
    <sec id="sec-4">
      <title>3. Conclusion</title>
      <p>
        We presented a demo for FeaBI [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], which is a recently proposed framework for explaining
KG embedding models. While the work in [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] focuses on technical details of the method, our
demo system allows the users to easily analyse KG features captured by an embedding model
as well as reasons behind embedding-based entity similarities. Future directions include the
analysis of explanations for relation embeddings as well as the consideration of ontologies and
KG schemes within the studied framework.
      </p>
      <p>Acknowledgements. This work was partially funded by the grant ANR-20-CHIA-0012-01
(“NoRDF”) and the European project SMARTEDGE (grant number 101092908).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Q.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Mao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <article-title>Knowledge graph embedding: A survey of approaches and applications</article-title>
          ,
          <source>IEEE Transactions on Knowledge and Data Engineering</source>
          <volume>29</volume>
          (
          <year>2017</year>
          )
          <fpage>2724</fpage>
          -
          <lpage>2743</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A.</given-names>
            <surname>Bordes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Usunier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>García-Durán</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Weston</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Yakhnenko</surname>
          </string-name>
          ,
          <article-title>Translating embeddings for modeling multi-relational data</article-title>
          ,
          <source>in: NeurIPs</source>
          ,
          <year>2013</year>
          , pp.
          <fpage>2787</fpage>
          -
          <lpage>2795</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>T.</given-names>
            <surname>Trouillon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Welbl</surname>
          </string-name>
          , S. Riedel, É. Gaussier, G. Bouchard,
          <article-title>Complex embeddings for simple link prediction</article-title>
          , in: ICML,
          <year>2016</year>
          , pp.
          <fpage>2071</fpage>
          -
          <lpage>2080</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>B.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Yih</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <article-title>Embedding entities and relations for learning and inference in knowledge bases</article-title>
          ,
          <source>in: ICLR</source>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>P.</given-names>
            <surname>Ristoski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Rosati</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. D.</given-names>
            <surname>Noia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. D.</given-names>
            <surname>Leone</surname>
          </string-name>
          , H. Paulheim,
          <article-title>Rdf2vec: RDF graph embeddings and their applications</article-title>
          ,
          <source>Semantic Web</source>
          <volume>10</volume>
          (
          <year>2019</year>
          )
          <fpage>721</fpage>
          -
          <lpage>752</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>T. N.</given-names>
            <surname>Kipf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Welling</surname>
          </string-name>
          ,
          <article-title>Semi-supervised classification with graph convolutional networks</article-title>
          ,
          <source>in: ICLR</source>
          <year>2017</year>
          ,
          <article-title>OpenReview</article-title>
          .net,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Ismaeil</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Stepanova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. K.</given-names>
            <surname>Tran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Blockeel</surname>
          </string-name>
          ,
          <article-title>Feabi: A feature selection-based framework for interpreting kg embeddings</article-title>
          ,
          <source>ISWC</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>S.</given-names>
            <surname>Vashishth</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Sanyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Nitin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. P.</given-names>
            <surname>Talukdar</surname>
          </string-name>
          ,
          <article-title>Composition-based multi-relational graph convolutional networks</article-title>
          ,
          <source>in: ICLR</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>M.</given-names>
            <surname>Galkin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. G.</given-names>
            <surname>Denis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. L.</given-names>
            <surname>Hamilton</surname>
          </string-name>
          ,
          <article-title>Nodepiece: Compositional and parametereficient representations of large knowledge graphs</article-title>
          ,
          <source>in: ICLR</source>
          ,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>S.</given-names>
            <surname>Mežnar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Lavrač</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Škrlj</surname>
          </string-name>
          , Snore:
          <article-title>Scalable unsupervised learning of symbolic node representations</article-title>
          ,
          <source>IEEE Access 8</source>
          (
          <year>2020</year>
          )
          <fpage>212568</fpage>
          -
          <lpage>212588</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>