<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Towards Binarization of Knowledge Graph Embeddings for Node Classification</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Vitor Faria de Souza</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Heiko Paulheim</string-name>
          <email>heiko.paulheim@uni-mannheim.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Knowledge Graph Embedding</institution>
          ,
          <addr-line>Node Classification, Binarization, Compression</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>University of Mannheim, Data and Web Science Group</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>power. Knowledge Graph Embeddings (KGEs) are dense representations of entities and relations of a knowledge graph (KG) in a continuous vector space. In this paper, we present a preliminary analysis showing that KGEs can be substantially compressed by using only binary instead of continuous features, replacing costly floating point storage by a more space-eficient bitwise storage, while retaining the representational ∗Corresponding author.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>CEUR
ceur-ws.org</p>
    </sec>
    <sec id="sec-2">
      <title>1. Introduction</title>
      <p>
        Knowledge Graph Embeddings (KGEs) are dense representations of entities and relations in
knowledge graphs, where each entity and relation is represented by a vector in a continuous
vector space. They are used for various in knowledge graph tasks, such as link prediction,
entity classification, for helping in knowledge-graph related tasks such as entity linking and
disambiguation, but also as knowledge representations in other downstream tasks, such as
recommender systems. [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ]
      </p>
      <p>
        While the scalability of KGE methods has been identified as a challenge [
        <xref ref-type="bibr" rid="ref3 ref4">3, 4</xref>
        ], most papers
looking into scalability take the angle of scaling up the training process, i.e., being able to learn
a representation for a (large) knowledge graph with reasonable time and memory usage. What
has been overlooked so far is the size of the resulting embedding, which is also an important
consideration for downstream tasks. Using KGEs in downstream applications requires loading
and processing those within the application, which may be hindered by the sheer size of the
embedding, especially when computing resources are limited.
      </p>
      <p>
        To illustrate the issue of KGE size, we use the well-known knowledge graph DBpedia [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ],
whose latest release contain 7.62 million entities.1 Both the most common implementation of
word2vec [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] in the gensim library [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], which is used by RDF2vec [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], as well as the PyTorch
library [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], which underlies embedding frameworks like DGL-KE [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] and PyKEEN [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], use 32
http://www.heikopaulheim.com/ (H. Paulheim)
      </p>
      <p>© 2024 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
bit floating point values for representing embeddings by default. 2,3 Embedding DBpedia with
200 dimensions would thus yield a 6GB large KGE model.4</p>
      <p>In this paper, we propose to use binary instead of continuous embeddings, i.e., embeddings
that only use the values 0 and 1 for each dimension. By doing so, the size of embeddings can
be reduced by a factor of 32 (using only 1 bit per entity and dimension instead of 32). In the
example above, we could thus store a 200 dimensional binary embedding using less than 200MB.</p>
      <p>
        The proposed approach is not limited to a single KGE method. Instead, we propose to binarize
an already trained KGE. In experiments with the DLCC node classification benchmark, eleven
diferent embedding methods, and diferent binarization methods, we show that the loss in
accuracy of using binary embeddings is rather small, but allows for a drastic reduction of storage
size. Especially for embeddings that are already computed for popular large-scale knowledge
graphs [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], binarization can reduce the data volume and downstream processing load.
      </p>
    </sec>
    <sec id="sec-3">
      <title>2. Approach</title>
      <p>
        In this paper, we utilize two approaches for binarizing knowledge graph embeddings: a simple
dimension-based baseline, and an autoencoder-based approach adapted from [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ].
      </p>
      <sec id="sec-3-1">
        <title>2.1. Baseline – Dimension-wise Binarization</title>
        <p>Given that  ()</p>
        <p>is the embedding vector of an entity  , we first compute the average vector  
as
where  ()
vector</p>
        <p>is the i-th element of  ()
from  ()
as</p>
        <p>. With that average vector, we can obtain a binary
   =
 
 ()

,
all entities e
 
 () = {
0    () &lt;   
1    () ≥   
In other words: every floating point value lower than the average for that dimension is
represented by a 0, all others are represented by a 1.</p>
      </sec>
      <sec id="sec-3-2">
        <title>2.2. Autoencoder-based Binarization</title>
        <p>
          Autoencoders are, in their simplest form, three-layer neural networks trained for dimensionality
reduction [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]. Between an input and an output layer, they have a (smaller) code layer. They
are trained by presenting instances of a dataset both as input and output (i.e., the input and
output being identical), and thereby learning to minimize the reconstruction error.
        </p>
        <p>
          [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] propose to use a heavyside step function as an activation function   ∶ R → {0, 1},
which only outputs the values 0 and 1. Specifically, they use of the Heaviside step function ℎ,
        </p>
        <sec id="sec-3-2-1">
          <title>2https://groups.google.com/g/gensim/c/JSSenT7Hhlc/m/FTEynSwmAQAJ</title>
          <p>3https://pytorch.org/docs/stable/notes/numerical_accuracy.html
4200 dimensions × 7,620,000 entities × 32 Bit
(1)
(2)
With that, the output of () is a binary vector. For further optimizing computations, they
propose to align the dimensionality of the code layer to CPU register widths, such as 64, 128,
and 256 bits. In this paper, we use their implementation5 for binarizing knowledge graph
embeddings.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>3. Evaluation</title>
      <p>
        In order to evaluate the quality of binarized embeddings, we use the DBpedia portion of
DLCC [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. For our experiments, we use the DBpedia embeddings also used in the paper [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ],
which are available online6. As embedding methods, we use four flavors of RDF2Vec [ 15, 16]
and two flavors (Norms L1 and L2) of TransE [ 17] embeddings are exploited, as well as ComplEx
[18], DistMult [19], RESCAL [20], RotatE [21], and TransR [22].
      </p>
      <p>
        Table 1 shows the average file sizes achieved with the diferent binarization methods. The
standard storage method uses an uncompressed TXT variant (i.e., storing the value ”0” or ”1” in
one byte), which can yield a compression factor up to 14. The .VEC format, which is supported
by the autoencoder implementation used in this paper [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], actually stores each eight single
embedding values in one byte (i.e., multiple embedding dimensions are stored in one long
integer). With this method, a compression factor of almost 50 can be achieved, leading to an
embedding originally stored in 1.6 GB now only occupying 33 MB.
      </p>
      <p>DLCC has three sizes of problems for each task; here, we only present the results for largest
problem size (5,000 examples per task). The full results are available online7; the results for the
smaller groups (500 and 50 examples, resp.) are comparable to those on the largest task group.</p>
      <p>Table 2 shows the results on the DLCC entity classification dataset. The average losses in
accuracy are not very large, with the the baseline performing better than the smaller
autoencoders. This makes us assume that the original embeddings capture the required information
for the tasks at hand already at a very coarse level, i.e., it is suficient to know that a vector
value for an entity at a particulary dimension is high or low, but it is not necessary to know
how low it actually is.</p>
      <sec id="sec-4-1">
        <title>5https://github.com/tca19/near-lossless-binarization 6https://data.dws.informatik.uni-mannheim.de/kgvec2go/dbpedia/2021-09/ 7https://github.com/vitor-faria/kgembeddings-binarization</title>
        <p>Moreover, we observe the largest losses for ComplEx and DistMult, which are both tensor
factorization methods, whereas RDF2vec SG has the smallest loss, compared to the uncompressed
variant. Therefore, we conclude that binarization is not equally efective for each model, but
that there are diferences by model family.</p>
        <p>Furthermore, we can observe that the binarization has a smaller impact than the choice of
the embedding variant: even the binarized versions of RDF2vec SG OA yield superior results to
the original, i.e., non binarized embeddings of most other embedding methods.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>4. Conclusion and Outlook</title>
      <p>In this paper, we have shown two strategies for reducing the size of a KGE model, one simple
baseline using dimension-wise binarization, and an approach based on neural autoencoders.
A promising preliminary evaluation on the DLCC entity classification benchmark shows that
the size of embeddings can be drastically reduced at a comparatively low loss in downstream
classification performance.</p>
      <p>
        While the focus of this paper has been on optimizing the storage, we have not yet looked
into possible gains in computation time for downstream processing, i.e., at inference time. For
example, [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] have proposed processing binary vectors with logical bitwise operators, which
could also yield significant performance gains over floating point operations. This is a promising,
but yet underexplored area. Moreover, a comparison of binarizing classically trained embeddings
with directly trained binary embeddings would be intriguing, especially for cases where no
pre-trained embeddings exist.
      </p>
      <p>So far, we have only looked into tasks involving single entities, as defined in the gEval and
DLCC benchmarks. The transfer to another popular usage of knowledge graph embeddings,
i.e., link prediction and triple scoring, has not yet been investigated. Except for the B-CP
approach [23], there are, to the best of our knowledge, no methods for exploiting binary
knowledge graph embeddings for link prediction. In particular, since B-CP directly learns the
embeddings, there are no studies on binarizing existing embeddings for link prediction.</p>
      <p>In summary, the results have shown that binary knowledge graph embeddings can be an
appealing alternative to the widely used floating point representations, allowing a drastic
reduction of storage space while, at the same time, leading to only marginal loss in downstream
performance.
[cs].
[15] J. Portisch, H. Paulheim, Putting RDF2vec in Order, 2021. URL: http://arxiv.org/abs/2108.</p>
      <p>05280. doi:10.48550/arXiv.2108.05280, arXiv:2108.05280 [cs].
[16] P. Ristoski, J. Rosati, T. Di Noia, R. De Leone, H. Paulheim, RDF2Vec: RDF graph
embeddings and their applications, Semantic Web 10 (2019) 721–752. URL: https:
//www.medra.org/servlet/aliasResolver?alias=iospress&amp;doi=10.3233/SW-180317. doi:10.
3233/SW-180317.
[17] A. Bordes, N. Usunier, A. Garcia-Duran, J. Weston, O. Yakhnenko, Translating Embeddings
for Modeling Multi-relational Data 2013 (2013).
[18] T. Trouillon, J. Welbl, S. Riedel, É. Gaussier, G. Bouchard, Complex embeddings for
simple link prediction, in: International conference on machine learning, PMLR, 2016, pp.
2071–2080.
[19] B. Yang, W.-t. Yih, X. He, J. Gao, L. Deng, Embedding Entities and Relations for Learning
and Inference in Knowledge Bases, 2014. URL: https://arxiv.org/abs/1412.6575v4.
[20] M. Nickel, V. Tresp, P. Kröger, A Three-Way Model for Collective Learning on
Multi</p>
      <p>Relational Data., 2011, pp. 809–816.
[21] Z. Sun, Z.-H. Deng, J.-Y. Nie, J. Tang, RotatE: Knowledge Graph Embedding by Relational
Rotation in Complex Space, 2019. URL: http://arxiv.org/abs/1902.10197. doi:10.48550/
arXiv.1902.10197, arXiv:1902.10197 [cs, stat].
[22] Y. Lin, Z. Liu, M. Sun, Y. Liu, X. Zhu, Learning Entity and Relation Embeddings for
Knowledge Graph Completion, Proceedings of AAAI 29 (2015) 2181–2187. doi:10.1609/
aaai.v29i1.9491.
[23] K. Hayashi, K. Kishimoto, M. Shimbo, Binarized embeddings for fast, space-eficient
knowledge graph completion, IEEE Transactions on Knowledge and Data Engineering 35
(2021) 141–153.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J.</given-names>
            <surname>Portisch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Heist</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Paulheim</surname>
          </string-name>
          ,
          <article-title>Knowledge graph embedding for data mining vs. knowledge graph embedding for link prediction - two sides of the same coin?</article-title>
          ,
          <source>Semantic Web</source>
          <volume>13</volume>
          (
          <year>2022</year>
          )
          <fpage>1</fpage>
          -
          <lpage>24</lpage>
          . doi:
          <volume>10</volume>
          .3233/SW-212892.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Q.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Mao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <article-title>Knowledge graph embedding: A survey of approaches and applications</article-title>
          ,
          <source>IEEE transactions on knowledge and data engineering 29</source>
          (
          <year>2017</year>
          )
          <fpage>2724</fpage>
          -
          <lpage>2743</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>R.</given-names>
            <surname>Biswas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.-A.</given-names>
            <surname>Kafee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Cochez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Dumbrava</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. E.</given-names>
            <surname>Jendal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lissandrini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Lopez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. L.</given-names>
            <surname>Mencía</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Paulheim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Sack</surname>
          </string-name>
          , et al.,
          <article-title>Knowledge graph embeddings: open challenges and opportunities</article-title>
          ,
          <source>Transactions on Graph Data and Knowledge</source>
          <volume>1</volume>
          (
          <year>2023</year>
          )
          <fpage>4</fpage>
          -
          <lpage>1</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>D.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Song</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Ma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Tan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Ye</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Dong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Xiong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , G. Karypis, Dgl-ke:
          <article-title>Training knowledge graph embeddings at scale</article-title>
          ,
          <source>in: Proceedings of the 43rd international ACM SIGIR conference on research and development in information retrieval</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>739</fpage>
          -
          <lpage>748</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>J.</given-names>
            <surname>Lehmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Isele</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Jakob</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Jentzsch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kontokostas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. N.</given-names>
            <surname>Mendes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hellmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Morsey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Van Kleef</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Auer</surname>
          </string-name>
          , C. Bizer, DBpedia
          <article-title>- A large-scale, multilingual knowledge base extracted from Wikipedia, Semantic Web 6 (</article-title>
          <year>2015</year>
          )
          <fpage>167</fpage>
          -
          <lpage>195</lpage>
          . URL: https://www. medra.org/servlet/aliasResolver?alias=iospress&amp;doi=10.3233/SW-140134. doi:
          <volume>10</volume>
          .3233/ SW-140134.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>T.</given-names>
            <surname>Mikolov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Chen</surname>
          </string-name>
          , G. Corrado,
          <string-name>
            <given-names>J.</given-names>
            <surname>Dean</surname>
          </string-name>
          ,
          <source>Eficient Estimation of Word Representations in Vector Space</source>
          ,
          <year>2013</year>
          . URL: http://arxiv.org/abs/1301.3781. doi:
          <volume>10</volume>
          .48550/arXiv.1301.3781, arXiv:
          <fpage>1301</fpage>
          .3781 [cs].
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>R.</given-names>
            <surname>Řehůřek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Sojka</surname>
          </string-name>
          , et al.,
          <article-title>Gensim-statistical semantics in python, Retrieved from genism</article-title>
          .
          <source>org</source>
          (
          <year>2011</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>H.</given-names>
            <surname>Paulheim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Ristoski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Portisch</surname>
          </string-name>
          ,
          <source>Embedding Knowledge Graphs with RDF2vec</source>
          , Springer Nature,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>A.</given-names>
            <surname>Paszke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gross</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Massa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Lerer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bradbury</surname>
          </string-name>
          , G. Chanan,
          <string-name>
            <given-names>T.</given-names>
            <surname>Killeen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Gimelshein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Antiga</surname>
          </string-name>
          , et al.,
          <article-title>Pytorch: An imperative style, high-performance deep learning library</article-title>
          ,
          <source>Advances in neural information processing systems</source>
          <volume>32</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>M.</given-names>
            <surname>Ali</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Berrendorf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. T.</given-names>
            <surname>Hoyt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Vermue</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Sharifzadeh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Tresp</surname>
          </string-name>
          ,
          <source>J. Lehmann, Pykeen</source>
          <volume>1</volume>
          .
          <article-title>0: a python library for training and evaluating knowledge graph embeddings</article-title>
          ,
          <source>Journal of Machine Learning Research</source>
          <volume>22</volume>
          (
          <year>2021</year>
          )
          <fpage>1</fpage>
          -
          <lpage>6</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>J.</given-names>
            <surname>Portisch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hladik</surname>
          </string-name>
          , H. Paulheim,
          <fpage>KGvec2go</fpage>
          - Knowledge
          <source>Graph Embeddings as a Service</source>
          ,
          <year>2020</year>
          . URL: http://arxiv.org/abs/
          <year>2003</year>
          .05809. doi:
          <volume>10</volume>
          .48550/arXiv.
          <year>2003</year>
          .
          <volume>05809</volume>
          , arXiv:
          <year>2003</year>
          .05809 [cs].
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>J.</given-names>
            <surname>Tissier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Gravier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Habrard</surname>
          </string-name>
          ,
          <article-title>Near-lossless Binarization of Word Embeddings</article-title>
          ,
          <source>Proceedings of the AAAI Conference on Artificial Intelligence</source>
          <volume>33</volume>
          (
          <year>2019</year>
          )
          <fpage>7104</fpage>
          -
          <lpage>7111</lpage>
          . URL: http://arxiv.org/abs/
          <year>1803</year>
          .09065. doi:
          <volume>10</volume>
          .1609/aaai.v33i01.
          <volume>33017104</volume>
          , arXiv:
          <year>1803</year>
          .09065 [cs].
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Yao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <article-title>Auto-encoder based dimensionality reduction</article-title>
          ,
          <source>Neurocomputing</source>
          <volume>184</volume>
          (
          <year>2016</year>
          )
          <fpage>232</fpage>
          -
          <lpage>242</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>J.</given-names>
            <surname>Portisch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Paulheim</surname>
          </string-name>
          ,
          <article-title>The DLCC Node Classification Benchmark for Analyzing Knowledge Graph Embeddings</article-title>
          ,
          <year>2022</year>
          . URL: http://arxiv.org/abs/2207.06014, arXiv:
          <fpage>2207</fpage>
          .
          <fpage>06014</fpage>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>