<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>disentanglement, with applications to the ΔVAE⋆</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Mahefa Ratsisetraina Ravelonanosy</string-name>
          <email>m.r.ravelonanosy@tue.nl</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Vlado Menkovski</string-name>
          <email>v.menkovski@tue.nl</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jacobus W. Portegies</string-name>
          <email>j.w.portegies@tue.nl</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Mathematics and Computer Science, Eindhoven University of Technology</institution>
          ,
          <addr-line>5612 AZ Eindhoven</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>EAISI, Eindhoven University of Technology</institution>
          ,
          <addr-line>5612 AZ Eindhoven</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2024</year>
      </pub-date>
      <abstract>
        <p>We investigate the ability of Difusion Variational Autoencoder ( ΔVAE) with unit sphere  2 as latent space to capture topological and geometrical structure and disentangle latent factors in datasets. For this, we introduce a new diagnostic of disentanglement: namely the topological degree of the encoder, which is a map from the data manifold to the latent space. We derive and implement an algorithm that computes this degree, and we use it to compute the degree of the encoder of models that result from the training procedure. Our experimental results show that the ΔVAE achieves relatively small LSBD scores, and that regardless of the degree after initialization, the degree of the encoder after training becomes −1 or +1, which implies that the resulting encoder is at least homotopic to a homeomorphism.</p>
      </abstract>
      <kwd-group>
        <kwd>Disentangled representation</kwd>
        <kwd>Homeomorphic autoencoding</kwd>
        <kwd>Topological degree</kwd>
        <kwd />
        <kwd>Variational Autoencoder</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>CEUR</p>
      <p>
        ceur-ws.org
1. Introduction
homeomorphism.
mismatch problem [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
manifold as latent space.
structure.
measures do exist, such as Linear Symmetry Based Disentanglement (LSBD) [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] and the LSBD score [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
      </p>
      <p>
        Besides formal definitions of disentanglement, there are desired characteristics for disentangled latent
factors, for instance that nearby points in the dataspace should correspond to nearby points in the latent
space representation. This could lead to a requirement that the encoder should be a homeomorphism
[
        <xref ref-type="bibr" rid="ref4 ref5">4, 5</xref>
        ].
      </p>
      <p>Given that an agreed definition of disentanglement does not yet exist, we consider it desirable to
develop a wide range of diagnostics that are somehow related to the intuitive concept of disentanglement.
Moreover, in practice it can be dificult to test whether a given encoder is a homeomorphism. Therefore,
we introduce the topological degree as a discrete diagnostic for disentanglement. A homeomorphic
encoder always has degree +1 or −1, whereas an encoder with degree ±1 is at least homotopic to a</p>
      <p>To achieve a homeomorphic encoder, or to get an encoder with degree ±1, one needs to choose
a latent space that matches the topology of the dataset, otherwise one will encounter the manifold</p>
      <p>
        In order to have a wider range of latent spaces and solve the manifold mismatch problem [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], Pérez Rey
et al [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] developed the Difusion Variational Autoencoder ( ΔVAE) that allows for any closed Riemannian
      </p>
      <p>We immediately apply the degree as a diagnostic for disentanglement in an evaluation of the ΔVAE,
in which we test the ΔVAE with a spherical latent space on data which naturally has a spherical latent
⋆This work was supported by NWO GROOT project UNRAVEL, OCENW.GROOT.2019.044.</p>
      <p>
        It can be challenging to train a ΔVAE, and we wondered whether this was due to the initialization and
topological obstructions, see also [
        <xref ref-type="bibr" rid="ref4 ref8">8, 4</xref>
        ]. Indeed, if training corresponds to a continuous deformation of
the encoder and decoder, if the degree would not be initialized at 1 or −1, the encoder would have no
chance to reach suitable disentanglement.
      </p>
      <p>
        Our experiments show that regardless of the initial model weight, the topological degree of the
encoder can change to become eventually constant equal to +1 or −1, after some epochs of the training
process. We perform the same experiments for the  -VAE [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] and compare the results. The code that
we used in the experiments can be found at https://gitlab.tue.nl/diffusion-vae/degree.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Related work</title>
      <p>
        The VAE [
        <xref ref-type="bibr" rid="ref10 ref9">9, 10</xref>
        ] and its extensions are among the most used models when it comes to learning
disentangled representations [
        <xref ref-type="bibr" rid="ref11 ref12 ref13">11, 12, 13</xref>
        ]. Some VAE extensions propose the use of more complex prior
distribution other than the Gaussian in order to better match the distribution of the latent code [
        <xref ref-type="bibr" rid="ref14 ref15 ref16 ref17">14, 15, 16, 17</xref>
        ].
Some extensions propose independence of each latent dimension by modifying the VAE loss function
[
        <xref ref-type="bibr" rid="ref11 ref13">11, 13</xref>
        ]. Other extensions use more geometric approaches to make the latent space itself match the
geometry of the dataset [
        <xref ref-type="bibr" rid="ref18 ref19 ref4 ref6 ref7">18, 6, 19, 4, 20, 7</xref>
        ].
      </p>
      <p>
        Intuitions and some aspects of disentangled representation are presented in [
        <xref ref-type="bibr" rid="ref1">1, 21, 22</xref>
        ], while overviews
of several disentanglement metrics are given in [23] and [24]. Disentanglement is originally assessed
with visual inspections and performance on downstream tasks [23]. Eforts have been devoted to propose
metrics to evaluate diferent aspects of disentanglement [
        <xref ref-type="bibr" rid="ref3 ref8">8, 25, 26, 27, 3</xref>
        ]. The disentanglement metrics
derived in these works do not check geometric aspects of disentanglement such as homeomorphism
and topological degree according to the original mathematical definitions of these aspects. The degree
was mentioned as a topological obstruction to homeomorphic autoencoding in [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Topological degree as a diagnostic for disentanglement</title>
      <p>The encoder of a VAE can be considered as a continuous map from a dataspace  ⊆ ℝ  to the latent
space  . Intuitively speaking, its topological degree is the number of times that the encoder wraps the
data manifold around the latent space, counted in such a way that positive cancels negative orientation
(cf. [28, Page 134] and [29, Page 27]). Encoder degree unequal to 1 or −1 indicates that the encoder
cannot be a homeomorphism [29, Page 51].</p>
      <p>Computing the topological degree of the encoder Although general methods exist [30], we
developed and implemented a basic algorithm targeted to the case at hand of computing the degree of a
map between spheres. We triangulate the two spheres, and create a “rounding” of the original map that
maps vertices to vertices, edges to collections of edges and faces to collections of faces. We finally count
how many times the faces in the target sphere are covered, taking into account orientation. Using tools
from homology theory we can prove that the algorithm gives the correct result cf. [31].</p>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments</title>
      <p>
        We train the ΔVAE with  2 as latent space, using a second-order expansion of the heat kernel [32]. We
use a dataset of spherical harmonics as a proxy for a more natural dataset of axisymmetric pictures on the
unit sphere, which naturally has the topology of  2 [33, Page 88] [34, 35]. We include a semisupervised
LSBD-loss as in [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] and a semisupervised LSBD loss for the decoder. Also, we evaluate the LSBD score
outlined in [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] with the group (3) . The representation of the data given by the models is then good if
the corresponding LSBD score is small. Furthermore, we compute the distance distortion (DD) metric
as given in [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], and the log-likelihood estimate as in [36]; for further details see also [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. We compare
the result with  -VAE. The numerical results of the experiments are presented in Table 1.
      </p>
      <p>
        Evolution of the degree during the training We conducted more experiments for spherical
harmonics of degree  = 7, 5, 3 with ΔVAE in order to get insight into the evolution of the degree during
the training. We performed 5 experiments where the absolute value of the degree before training was
not 1, whereas the absolute value of the degree after all training was 1. In particular, even though
we share the opinion that topological obstructions might hamper training [
        <xref ref-type="bibr" rid="ref4 ref8">8, 4</xref>
        ], for the ΔVAE the
obstruction to the degree can be overcome.
      </p>
    </sec>
    <sec id="sec-5">
      <title>5. Discussion</title>
      <p>We derive a second order expansion of the heat kernel on the unit sphere  2 by using the theoretical
result of [32], and use it as approximation in the ΔVAE loss function. Though the efect of such higher
order approximation in the performance of ΔVAE is not studied yet.</p>
      <p>Our algorithm for degree computation could be generalized to higher dimensional sphere   with
 &gt; 2 , but due to the curse of dimensionality, practical computation is most likely only feasible in very
low dimensions: for a  -dimensional manifold and a discretization length  , the number of faces needed
in the triangulation scales as  − .</p>
      <p>The amount of semisupervision is relatively high in our experiments. For lower degree spherical
harmonics ( = 1, 3, 5 ), the amount of semisupervision can be reduced drastically, although we have
not yet performed a systematic study.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion</title>
      <p>We evaluate to what extent the ΔVAE can capture topological properties or disentangle generating
factors, as measured by the LSBD score, and as expressed by a new discrete diagnostic for
disentanglement: the degree of the encoder. We use the encoder degree as a means to gain more insight in the
training behavior.</p>
      <p>First, we obtain relatively small LSBD scores, which expresses that the ΔVAE indeed can capture or
disentangle the latent rotational factor relatively well. In comparison with the  -VAE, we find that the
 -VAE typically obtains better log-likelihood scores, while the reconstruction error and LSBD score are
a bit better for the ΔVAE.</p>
      <p>Secondly, we implemented an algorithm for computing the topological degree of the encoder and
ifnd that even though the encoder is typically initialized with degree 0, this degree can change and after
training the encoder indeed has degree of ±1, which means that the encoder is at least homotopic to a
homeomorphism and that the learned spherical representation preserves the topological structure of
the dataset at least up to a homotopy. In particular, we find that the sphere in latent space is completely
covered by the image of the data manifold.
disentanglement learning, in: Proceedings of the IEEE/CVF conference on computer vision and
pattern recognition, 2020, pp. 7920–7929.
[20] I. Huh, J. M. Choe, Y. KIM, D. Kim, et al., Isometric quotient variational auto-encoders for
structurepreserving representation learning, Advances in Neural Information Processing Systems 36
(2024).
[21] K. Do, T. Tran, Theory and evaluation metrics for learning disentangled representations,
International Conference on Learning Representations, ICLR 2020 (2020).
[22] S. Van Steenkiste, F. Locatello, J. Schmidhuber, O. Bachem, Are disentangled representations
helpful for abstract visual reasoning?, Advances in neural information processing systems 32
(2019).
[23] M.-A. Carbonneau, J. Zaidi, J. Boilard, G. Gagnon, Measuring disentanglement: A review of
metrics, IEEE transactions on neural networks and learning systems (2022).
[24] A. Sepliarskaia, J. Kiseleva, M. de Rijke, How not to measure disentanglement, in: ICML Workshop
on Theoretic Foundation, Criticism, and Application Trend of Explainable AI, 2021.
[25] I. Higgins, L. Matthey, A. Pal, C. P. Burgess, X. Glorot, M. M. Botvinick, S. Mohamed, A. Lerchner,
beta-VAE: Learning basic visual concepts with a constrained variational framework., ICLR (Poster)
3 (2017).
[26] F. Locatello, S. Bauer, M. Lucic, G. Raetsch, S. Gelly, B. Schölkopf, O. Bachem, Challenging common
assumptions in the unsupervised learning of disentangled representations, in: international
conference on machine learning, PMLR, 2019, pp. 4114–4124.
[27] R. Suter, D. Miladinovic, B. Schölkopf, S. Bauer, Robustly disentangled causal mechanisms:
Validating deep representations for interventional robustness, in: International Conference on
Machine Learning, PMLR, 2019, pp. 6056–6065.
[28] A. Hatcher, Algebraic topology, Cambridge University Press, 2005.
[29] J. Milnor, Topology from the diferentiable viewpoint, univ, Press of Virginia, Charlottesville 1990
(1965).
[30] T. Kaczynski, K. M. Mischaikow, M. Mrozek, Computational homology, volume 157, Springer, 2004.
[31] M. R. Ravelonanosy, V. Menkovski, J. W. Portegies, Topological degree as a discrete diagnostic for
disentanglement, with applications to the ΔVAE, https://arxiv.org/abs/2409.01303 (2024).
[32] V. Menkovski, J. W. Portegies, M. R. Ravelonanosy, Small time asymptotics of the entropy of the
heat kernel on a riemannian manifold, Applied and Computational Harmonic Analysis 71 (2024).</p>
      <p>URL: https://www.sciencedirect.com/science/article/pii/S1063520324000198.
[33] T. Bröcker, T. tom Dieck, Representations of compact lie groups, Graduate Texts in Mathematics
(1985).
[34] M. A. Blanco, M. Flórez, M. Bermejo, Evaluation of the rotation matrices in the basis of real
spherical harmonics, Journal of Molecular structure: THEOCHEM 419 (1997) 19–27.
[35] S. Harmonics, Claus mulier, Lecture Notes in Mathematics (LNM)) 17 (1966).
[36] Y. Burda, R. B. Grosse, R. Salakhutdinov, Importance weighted autoencoders, in: Y. Bengio, Y. LeCun
(Eds.), 4th International Conference on Learning Representations, ICLR 2016, San Juan, Puerto
Rico, May 2-4, 2016, Conference Track Proceedings, 2016. URL: http://arxiv.org/abs/1509.00519.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bengio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Courville</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Vincent</surname>
          </string-name>
          ,
          <article-title>Representation learning: A review and new perspectives</article-title>
          ,
          <source>IEEE transactions on pattern analysis and machine intelligence</source>
          <volume>35</volume>
          (
          <year>2013</year>
          )
          <fpage>1798</fpage>
          -
          <lpage>1828</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>I.</given-names>
            <surname>Higgins</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Amos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Pfau</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Racaniere</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Matthey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Rezende</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Lerchner</surname>
          </string-name>
          ,
          <article-title>Towards a definition of disentangled representations</article-title>
          , arXiv preprint arXiv:
          <year>1812</year>
          .
          <volume>02230</volume>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>L.</given-names>
            <surname>Tonnaer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. A. P.</given-names>
            <surname>Rey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Menkovski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Holenderski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Portegies</surname>
          </string-name>
          ,
          <article-title>Quantifying and learning linear symmetry-based disentanglement</article-title>
          ,
          <source>in: International Conference on Machine Learning, PMLR</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>21584</fpage>
          -
          <lpage>21608</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>L.</given-names>
            <surname>Falorsi</surname>
          </string-name>
          , P. De Haan,
          <string-name>
            <given-names>T. R.</given-names>
            <surname>Davidson</surname>
          </string-name>
          , N. De Cao,
          <string-name>
            <given-names>M.</given-names>
            <surname>Weiler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Forré</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. S.</given-names>
            <surname>Cohen</surname>
          </string-name>
          ,
          <article-title>Explorations in homeomorphic variational auto-encoding</article-title>
          ,
          <source>ICML18 Workshop on Theoretical Foundations and Applications of Deep Generative Models</source>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>P. de Haan</surname>
          </string-name>
          , L. Falorsi,
          <article-title>Topological constraints on homeomorphic auto-encoding</article-title>
          ,
          <source>NeurIPS 2018 workshop on Integration of Deep Learning Theories</source>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>T. R.</given-names>
            <surname>Davidson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Falorsi</surname>
          </string-name>
          , N. De Cao,
          <string-name>
            <given-names>T.</given-names>
            <surname>Kipf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Tomczak</surname>
          </string-name>
          ,
          <source>Hyperspherical variational autoencoders, 34th Conference on Uncertainty in Artificial Intelligence (UAI-18)</source>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>L. A.</given-names>
            <surname>Perez Rey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Menkovski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Portegies</surname>
          </string-name>
          ,
          <article-title>Difusion variational autoencoders</article-title>
          ,
          <source>in: Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, IJCAI-20</source>
          ,
          <year>2020</year>
          . doi:
          <volume>10</volume>
          .24963/ijcai.
          <year>2020</year>
          /375.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>B.</given-names>
            <surname>Esmaeili</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Walters</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Zimmermann</surname>
          </string-name>
          , J.-W. van de Meent,
          <article-title>Topological obstructions and how to avoid them</article-title>
          ,
          <source>Advances in Neural Information Processing Systems</source>
          <volume>36</volume>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>D. P.</given-names>
            <surname>Kingma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Welling</surname>
          </string-name>
          ,
          <article-title>Auto-encoding Variational Bayes</article-title>
          , in: Y. Bengio, Y. LeCun (Eds.),
          <source>2nd International Conference on Learning Representations, ICLR</source>
          <year>2014</year>
          ,
          <article-title>Banf</article-title>
          ,
          <string-name>
            <surname>AB</surname>
          </string-name>
          , Canada,
          <source>April 14-16</source>
          ,
          <year>2014</year>
          , Conference Track Proceedings,
          <year>2014</year>
          . URL: http://arxiv.org/abs/1312.6114.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>D. J.</given-names>
            <surname>Rezende</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mohamed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Wierstra</surname>
          </string-name>
          ,
          <article-title>Stochastic backpropagation and approximate inference in deep generative models</article-title>
          ,
          <source>in: International conference on machine learning, PMLR</source>
          ,
          <year>2014</year>
          , pp.
          <fpage>1278</fpage>
          -
          <lpage>1286</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>C. P.</given-names>
            <surname>Burgess</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Higgins</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Pal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Matthey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Watters</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Desjardins</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Lerchner</surname>
          </string-name>
          ,
          <article-title>Understanding disentangling in beta-VAE, Learning Disentangled Representations: from Perception to</article-title>
          Control Workshop,
          <year>2017</year>
          (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>J.</given-names>
            <surname>Cha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Thiyagalingam</surname>
          </string-name>
          ,
          <article-title>Orthogonality-enforced latent space in autoencoders: an approach to learning disentangled representations</article-title>
          ,
          <source>in: International Conference on Machine Learning, PMLR</source>
          ,
          <year>2023</year>
          , pp.
          <fpage>3913</fpage>
          -
          <lpage>3948</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>H.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mnih</surname>
          </string-name>
          ,
          <article-title>Disentangling by factorising</article-title>
          ,
          <source>in: International conference on machine learning, PMLR</source>
          ,
          <year>2018</year>
          , pp.
          <fpage>2649</fpage>
          -
          <lpage>2658</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>M. D. Hofman</surname>
            ,
            <given-names>M. J.</given-names>
          </string-name>
          <string-name>
            <surname>Johnson</surname>
          </string-name>
          ,
          <article-title>Elbo surgery: yet another way to carve up the variational evidence lower bound</article-title>
          , in: Workshop in Advances in Approximate Bayesian Inference,
          <string-name>
            <surname>NIPS</surname>
          </string-name>
          , volume
          <volume>1</volume>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>A.</given-names>
            <surname>Klushyn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Kurle</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Cseke</surname>
          </string-name>
          , P. van der Smagt,
          <article-title>Learning hierarchical priors in vaes</article-title>
          ,
          <source>Advances in neural information processing systems</source>
          <volume>32</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>J.</given-names>
            <surname>Tomczak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Welling</surname>
          </string-name>
          ,
          <article-title>Vae with a vampprior</article-title>
          ,
          <source>in: International conference on artificial intelligence and statistics</source>
          , PMLR,
          <year>2018</year>
          , pp.
          <fpage>1214</fpage>
          -
          <lpage>1223</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>C. K. Sønderby</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Raiko</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Maaløe</surname>
            ,
            <given-names>S. K.</given-names>
          </string-name>
          <string-name>
            <surname>Sønderby</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          <string-name>
            <surname>Winther</surname>
          </string-name>
          , Ladder variational autoencoders,
          <source>Advances in neural information processing systems</source>
          <volume>29</volume>
          (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>C.</given-names>
            <surname>Chadebec</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Thibeau-Sutre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Burgos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Allassonnière</surname>
          </string-name>
          ,
          <article-title>Data augmentation in high dimensional low sample size setting using a geometry-based variational autoencoder</article-title>
          ,
          <source>IEEE Transactions on Pattern Analysis and Machine Intelligence</source>
          <volume>45</volume>
          (
          <year>2022</year>
          )
          <fpage>2879</fpage>
          -
          <lpage>2896</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Ding</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Parmar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Welling</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Tu</surname>
          </string-name>
          , Guided variational autoencoder for
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>