<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>April</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Serge Dolgikh</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>National Aviation University</institution>
          ,
          <addr-line>1, Lubomira Huzara, Kyiv, 03058</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Solana Networks</institution>
          ,
          <addr-line>301 Moodie Dr., Ottawa</addr-line>
          ,
          <country country="CA">Canada</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2021</year>
      </pub-date>
      <volume>23</volume>
      <issue>2021</issue>
      <fpage>0000</fpage>
      <lpage>0001</lpage>
      <abstract>
        <p>In this work latent representations of image data were investigated with neural network models of generative self-learning. A convolutional autoencoder with strong redundancy reduction was used to create latent representations of images of basic geometric shapes in the process of unsupervised generative learning and the characteristics of distributions of concept regions in the latent space of models investigated. It was demonstrated that conceptual representations with good separation of latent regions can be produced with generative models of limited complexity and that characteristic types of data, or “concepts” form well-defined, continuous and connected regions in the latent space. Geometric structure of the latent representations was described in detail confirming connected and continuous topology of latent concept regions clearly associated with characteristic types of observable data, such as shape, size and contrast. The results indicate that conceptual representations created in the process of unsupervised generative learning can form a natural basis for the emergence of abstract concepts in intelligent systems.</p>
      </abstract>
      <kwd-group>
        <kwd>1 Machine learning</kwd>
        <kwd>unsupervised learning</kwd>
        <kwd>concept learning</kwd>
        <kwd>representations</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Representation learning with the objective to identify the informative structure in general real-world
data has a well-established record in the field of Machine Learning. Hierarchical representations of
different types of data were obtained with Restricted Boltzmann Machines (RBM), Deep Belief
Networks (DBN) [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ][
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], different flavors of autoencoders [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ][
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] and other models and architectures
allowed to improve accuracy of supervised learning. Different types, architectures and flavors of
generative models were investigated since including autoencoder neural networks, Generative
Adversarial Networks (GAN) and others [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ][
        <xref ref-type="bibr" rid="ref6">6</xref>
        ][
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
      <p>
        A number of interesting experimental results were obtained in concept learning with artificial
systems, such as the “cat experiment”, that demonstrated spontaneous emergence of concept sensitivity
on a single neuron level in unsupervised deep learning with image data [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. Disentangled
representations were produced and studied with deep variational autoencoder models and different
types of image data [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] pointing at the possibility of a general nature of the effect. Concept-associated
structure was observed in latent representations of diverse real-world data such Internet in large public
networks, aerial surveillance images [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ][
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] and other results [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ][
        <xref ref-type="bibr" rid="ref13">13</xref>
        ].
      </p>
      <p>
        The relations between learning and statistical thermodynamics was studied in the theory of learning
systems [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ][
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] leading to understanding of a deep connection between learning processes in artificial
systems and principles of information theory and statistical thermodynamics.
      </p>
      <p>
        These results demonstrated that training of artificial learning models under certain constraints such
as minimization of predictive error can lead to emergence of structure associated and correlated with
characteristic patterns in the observable data. This approach, known as representation learning, was
developed in a number of studies [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. Of interest in this work is unsupervised conceptual representation
learning with models of generative learning that does not use explicitly labeled data but rather attempts
to minimize the error of reproducing the observed distribution from a representation in the latent space
created in the process of generative self-learning. As some of the earlier results have indicated, such
entirely unsupervised process can produce non-trivial structure in the latent representations that is
correlated with characteristic patterns in the observable data and can be used as a foundation for learning
methods and processes based on such structured low-dimensional representations.
      </p>
      <p>
        Interestingly, these observations in unsupervised learning of artificial systems were paralleled very
recently by several results in the studies of biologic sensory networks [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ][
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] that demonstrated
commonality of low-dimensional representations in processing sensory information by mammals,
including humans.
      </p>
      <p>Based on and inspired by these results in conceptual representation learning, the questions
investigated in this work are the following: what properties characterize latent representations of
successful generative models? Is there a stable association between characteristic patterns in the
observable data and the structure that emerges in the latent distributions in the process of generative
self-learning? What structure can be identified in the latent representations with entirely unsupervised
methods, without prior knowledge of conceptual content of the observable data?</p>
      <p>
        These objectives were approached with artificial neural network models of deep convolutional
autoencoder that showed effectiveness in producing informative representations [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] and a dataset of
images of basic geometric shapes that are described in the following sections.
      </p>
      <p>In conclusion, a brief clarification of terminology used throughout this work. We will refer to the
characteristic types or patterns in the dataset that are labeled with known types or classes as “external”
concepts, that signifies that the type or label of the input is defined outside of the model based on some
external or prior knowledge about the observable data. An example of an external concept for an image
with a geometric shape can be its general type, “a triangle”. In contrast, a characteristic structure in the
latent representation that can be identified reliably by unsupervised means without any external or prior
information, will be referred to as “internal” or “native” concept. Thus, a question of the relation
between the external and native concepts can be of interest as well.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Methods</title>
      <p>In this section the models and data used in the study are described. A dataset of basic shapes such
as circles, triangles and greyscale backgrounds, was used to produce low-dimensional latent
representations that were studied with several methods as described in the following sections.
2.1.</p>
    </sec>
    <sec id="sec-3">
      <title>Deep Convolutional Autoencoder</title>
      <p>
        We used artificial neural network models with the architecture of convolutional autoencoder [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] with
added strong dimensionality reduction in the representation layer to produce three-dimensional latent
representations of image datasets of basic geometric shapes. The advantages of the autoencoder
architecture are that it allows to learn essential characteristics of the input data in an unsupervised
process without labeled samples, while virtually unlimited depths and complexity allows it to be used
with complex real-world data, demonstrating successful learning in many applications [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]-[
        <xref ref-type="bibr" rid="ref11">11</xref>
        ],[
        <xref ref-type="bibr" rid="ref19">19</xref>
        ].
      </p>
      <p>The architecture diagram of the models used in the study is presented in Figure 1.
approximately 40,000 trainable parameters.</p>
      <p>Unsupervised training was performed with standard methods such as Stochastic Gradient Descent
[20] to minimize average distance between the input and generated output with binary cross-entropy
cost function. The models were implemented with Keras / Tensorflow [21] and several standard
machine learning libraries were used in the analysis of latent representations.</p>
      <p>The latent representation was produced by activations of the neurons in the central, encoding layer
of the model (Figure 1). The activations of neurons were interpreted as coordinates in a
threedimensional latent representation space.</p>
    </sec>
    <sec id="sec-4">
      <title>Data</title>
      <p>pixels.</p>
      <p>Three datasets of geometric shape images with different variance and complexity were used to
investigate the structure of the latent representations. The images were greyscale, of the size 64 × 64</p>
      <p>The first, generated dataset, Shapes-1 consisted of 600 greyscale images of circles, triangles and
greyscale backgrounds with two representative samples per type with difference in the size and contrast
of fore / background.</p>
      <p>The second generated dataset, Shapes-2 contained the total of 600 – 1,000 greyscale images of
circles, triangles and backgrounds with variation in size in the range 0.3 – 1.0 of the image size (that is,
0.3 × 64 pixels), with variation of contrast of fore- vs. background for each size.</p>
      <p>In the generated datasets the images were centered, symmetrical and had no rotation. Only darker
foregrounds relative to the background were used.</p>
      <p>The third dataset, Shapes-3 contained 1,500 images of randomly generated colored geometric shapes
and backgrounds varying in: size; position relative to the center of the image; rotation; and the color of
the fore- and background [22]. The images in this dataset were converted to greyscale.
2.3.</p>
    </sec>
    <sec id="sec-5">
      <title>Latent Representations</title>
      <p>Following the process of unsupervised training in which significant reduction in the value of cost
function was observed, the models were able to perform two essential transformations of data:</p>
      <p>The encoding transformation, from the input data space, that is, images to the three-dimensional
latent representation, with coordinates represented by activations of neurons in the central, “encoding”
layer of the model (Figure 1):

 =  ( ) = 
  =  ( ) = 
( )
( )
The generative transformation operates in the opposite direction, i.e. from the latent representation into
the observable (image) space and is performed by the generative part of the autoencoder:
where y is a latent position, i.e. a point in the latent space represented by latent coordinates (y1, y2, y3).</p>
      <p>Transformations of encoding and generation allow to study the structure in the latent representation
by measuring and visualization of the distributions of samples encoded to the latent space, as well as
generative properties of the models via producing images of latent positions and regions of interest.
(1)
(2)</p>
    </sec>
    <sec id="sec-6">
      <title>3. Results</title>
      <p>The subject of this investigation was topological structure in the latent representations created by
generative models in the process of unsupervised learning. To verify the hypothesis of the study,
methods of analysis of latent distributions and generative ability of models were applied as described
in this section. It is essential to note that the process of production of latent representations via
generative learning was entirely unsupervised and no labeled data was used in configuration and / or
training of the models.</p>
    </sec>
    <sec id="sec-7">
      <title>Structure of Latent Representations</title>
      <p>Latent representations of generative models displayed well structured, connected and continuous
distribution of data by characteristic type as illustrated in Figure 2. The visualization data, a subset of
Shapes-1 and Shapes-2 datasets described in Section 2.2 represented a selection of images with
variation of size and greyscale contrast, color-coded by characteristic type (i.e., circle, triangle or
greyscale background).</p>
      <p>Visualization of latent distributions by characteristic type showed that the shapes present in the
training dataset were assigned specific well-defined regions in the latent space, with latent coordinates
encoding essential characteristics of shapes, such as size and contrast.</p>
    </sec>
    <sec id="sec-8">
      <title>3.1.1. Connectedness of Latent Representations</title>
      <p>As discussed above, latent representations created by generative models in unsupervised training
displayed connected and continuous topology allowing to conclude that there exist an association
between well-defined latent regions and characteristic types, or general concepts in the observable data.
This conclusion was further substantiated with two experiments.</p>
      <p>In the first experiment, a set of images S of the same type was selected, with variation in
characteristics, such as size and contrast. The images were transformed to the latent space as in (1) and
the mean of the resulting set of latent coordinates calculated as:


= 
( ( ))
(3)
The mean latent representation rmean of the input set was then propagated to the observed space with
generative transformation (2) and the resulting image produced. In the experiments, the generated image
of the input set was of the same type (Table 1).</p>
      <sec id="sec-8-1">
        <title>Generated mean of input set of shape type</title>
      </sec>
      <sec id="sec-8-2">
        <title>Circles-2</title>
      </sec>
      <sec id="sec-8-3">
        <title>Circles-3</title>
      </sec>
      <sec id="sec-8-4">
        <title>Triangles-2</title>
      </sec>
      <sec id="sec-8-5">
        <title>Input set</title>
      </sec>
      <sec id="sec-8-6">
        <title>Encoded mean (example)</title>
        <p>Triangles-3 (1.437, 0.073, -1.358) Triangle*
Background-3 (3.974, 0.011, -0.176) Background
* An exception was observed in the boundary region (Section 3.1.2), where the mean of a certain
input set produced mixed result.</p>
        <p>In the second group of experiments, a set of points in a close neighborhood of a selected latent
position of a given type was selected, and images corresponding to these positions generated with (2).
The resulting images were observed to be of the same type as that of the position of origin (Figure 3).</p>
      </sec>
    </sec>
    <sec id="sec-9">
      <title>3.1.2. Boundary Regions</title>
      <p>Well defined character of latent regions associated with characteristic types of shapes was confirmed
by observation of boundary areas between the regions of different types, that produced generated output
of mixed form with features from both types. The examples of such boundary generative regions are
shown in Figure 4.</p>
    </sec>
    <sec id="sec-10">
      <title>Semantics of Latent Coordinates</title>
      <p>Generative transformation (2) allows to investigate semantical meaning of the latent coordinates
defined as activations of neurons in the encoding layer of the generative model. A method of latent
probing was developed that allowed to produce and evaluate generated image of a specified position in
the latent space, after conceptual latent representation was produced in the unsupervised learning phase.
By applying probing to latent representations of generative models the structure that formed in the
process of generative learning under the constraints discussed earlier can be studied in practically any
level of detail.</p>
      <p>In the input to the method, a set of coordinates of a latent position of interest was provided. The
position was propagated via generative part of the autoencoder with generative transformation (2)
producing an output as an observable image. Evaluation of generated images associated with a set of
latent positions, such as a neighborhood, a line, a surface and so on, in a latent region of interest allowed
to obtain essential observations on the semantical meaning of the latent coordinates. The examples of
probing along the latent axes are shown in Figure 5.</p>
    </sec>
    <sec id="sec-11">
      <title>4. Conclusion</title>
      <p>In this work latent representations of images representing basic geometric shapes were analyzed
with methods of unsupervised machine learning. The analysis produced a number or essential results
and observations.</p>
      <p>It was demonstrated that successful generative learning with stable association of characteristic
observable inputs to a structure in the low-dimensional latent representations can be achieved with
models of limited complexity, both in size and architecture, well within a range of simple biologic
systems. The complexity of models in this work was roughly equivalent to a nervous system of a
jellyfish that has a comparable number of neurons and synapses [24].</p>
      <p>By applying unsupervised methods that did not use any externally labeled data in the analysis of
latent representations it was demonstrated that generative learning can lead to emergence of
welldefined latent regions associated with essentially different types of observable data. These results points
to possible origin of higher-level abstract concepts in unsupervised generative learning under the
constraints of generative accuracy and redundancy reduction, both having clear evolutionary advantage
for the learner.</p>
      <p>The structure of latent regions associated with characteristic patterns, native concepts in the
observable data was investigated and described, confirming connectedness and continuous topology of
the latent regions, with distinct boundaries between different concept regions. An analysis of generative
capacity of models with the developed method of latent probing allowed to make essential observations
on the semantics of the latent coordinates in the representations produced by the models, showing an
association with essential characteristics of the data such as type, size and contrast. However, as the
analysis of the semantics of latent coordinates has shown (Section 3.2) there may not be a fixed global
semantics to the latent coordinates and their significance is determined by the locality, i.e. the position
and the region, again confirming manifold-like topology of latent representations emergent in
generative self-learning.</p>
      <p>Overall, the results of this work demonstrated that conceptual representations can be a natural result
of unsupervised observation of the environment under the constraints that are both natural and common
for a learning system [25], of artificial or biologic nature.</p>
    </sec>
    <sec id="sec-12">
      <title>5. Future Work</title>
      <p>
        Following the direction outlined in this work, further studies can focus on investigation of latent
representations of more complex visual data with greater variation of conceptual content in the
observable environment. It is likely that advances in this direction would require models of considerably
greater complexity, in size, depth and architecture [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. For example, investigation of neural network
models with sparsity constraints can be a promising direction due to apparent success of this architecture
in the sensory input processing neural networks of biologic systems as reported in recently published
results [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ].
      </p>
    </sec>
    <sec id="sec-13">
      <title>6. References</title>
      <p>[20] Spall, J. C., Introduction to stochastic search and optimization: estimation, simulation, and control,</p>
      <p>Wiley Hoboken, NJ, ISBN 0-471-33052-3 2003.
[21] Keras: Python deep learning library, https://keras.io/, last accessed: 2020/11/21.
[22] El Korchi, A., 2D geometric shapes dataset, Mendeley Data V1, 2020,</p>
      <p>doi:10.17632/wzr2yv7r53.1.
[23] Zhou X., Belkin M., Semi-supervised learning. In: Academic Press Library in Signal Processing,</p>
      <p>Elsevier 1, 2014, pp. 1239–1269.
[24] Garm, A., Poussart, Y., Parkefelt, L., Ekström, P., Nilsson, D-E., The ring nerve of the box jellyfish</p>
      <p>Tripedalia cystophora, Cell Tissue Research 329 (1) (2007) 147–157.
[25] Dolgikh, S., Why good generative models categorize? International Journal of Modern Education
and Computer Science (2021). To appear.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Hinton</surname>
            ,
            <given-names>G.E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Osindero</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Teh</surname>
            <given-names>Y.W.,</given-names>
          </string-name>
          <article-title>A fast learning algorithm for deep belief nets</article-title>
          ,
          <source>Neural Computation</source>
          <volume>18</volume>
          (
          <issue>7</issue>
          ) (
          <year>2006</year>
          )
          <fpage>1527</fpage>
          -
          <lpage>1554</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Fischer</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Igel</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <article-title>Training restricted Boltzmann machines: an introduction</article-title>
          ,
          <source>Pattern Recognition</source>
          <volume>47</volume>
          (
          <year>2014</year>
          )
          <fpage>25</fpage>
          -
          <lpage>39</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Bengio</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <article-title>Learning deep architectures for AI, Foundations and Trends in Machine Learning 2 (1) (</article-title>
          <year>2009</year>
          )
          <fpage>1</fpage>
          -
          <lpage>127</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Coates</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ng</surname>
            ,
            <given-names>A.Y.</given-names>
          </string-name>
          ,
          <article-title>An analysis of single-layer networks in unsupervised feature learning</article-title>
          ,
          <source>in: Proceedings of 14th International Conference on Artificial Intelligence and Statistics 15</source>
          ,
          <year>2011</year>
          , pp.
          <fpage>215</fpage>
          -
          <lpage>223</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Welling</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Kingma</surname>
            ,
            <given-names>D.P.</given-names>
          </string-name>
          ,
          <article-title>An introduction to variational autoencoders</article-title>
          ,
          <source>Foundations and Trends in Machine Learning</source>
          <volume>12</volume>
          (
          <issue>4</issue>
          ) (
          <year>2019</year>
          )
          <fpage>307</fpage>
          -
          <lpage>392</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Creswell</surname>
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>White</surname>
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dumoulin</surname>
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Arulkumaran</surname>
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sengupta</surname>
            <given-names>B.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Bharath</surname>
            <given-names>A.A.</given-names>
          </string-name>
          ,
          <article-title>Generative adversarial networks: an overview</article-title>
          ,
          <source>IEEE Signal Processing Magazine</source>
          <volume>35</volume>
          (
          <issue>1</issue>
          ) (
          <year>2018</year>
          )
          <fpage>53</fpage>
          -
          <lpage>65</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Partaourides</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chatzis</surname>
            ,
            <given-names>S.P.</given-names>
          </string-name>
          ,
          <source>Asymmetric deep generative models, Neurocomputing</source>
          <volume>241</volume>
          (
          <year>2017</year>
          )
          <fpage>90</fpage>
          -
          <lpage>96</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Le</surname>
            ,
            <given-names>Q.V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ransato</surname>
            ,
            <given-names>M. A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Monga</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          et al.,
          <article-title>Building high level features using large scale unsupervised learning</article-title>
          .
          <source>arXiv 1112</source>
          .
          <article-title>6209 [cs</article-title>
          .LG]
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Higgins</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Matthey</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Glorot</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pal</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          et al.,
          <article-title>Early visual concept learning with unsupervised deep learning</article-title>
          .
          <source>arXiv 1606</source>
          .
          <article-title>05579 [cs</article-title>
          .LG]
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Dolgikh</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <article-title>Categorized representations and general learning</article-title>
          ,
          <source>in: Advances in Intelligent Systems and Computing</source>
          , volume
          <volume>1095</volume>
          , Springer,
          <year>Cham 2019</year>
          , pp.
          <fpage>93</fpage>
          -
          <lpage>100</lpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>030</fpage>
          -35249- 3_
          <fpage>11</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Prystavka</surname>
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cholyshkina</surname>
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dolgikh</surname>
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Karpenko</surname>
            <given-names>D.</given-names>
          </string-name>
          ,
          <article-title>Automated object recognition system based on aerial photography</article-title>
          ,
          <source>in: Proceedings of 10th International Conference on Advanced Computer Information Technologies, ACIT-2020</source>
          , Deggendorf, Germany,
          <year>2020</year>
          , pp.
          <fpage>830</fpage>
          -
          <lpage>833</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Rodriguez</surname>
            ,
            <given-names>R.C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Alaniz</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Akata</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <article-title>Modeling conceptual understanding in image reference games</article-title>
          ,
          <source>in: Proceedings of Advances in Neural Information Processing Systems</source>
          , Vancouver,
          <year>2019</year>
          , pp.
          <fpage>13155</fpage>
          -
          <lpage>13165</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Young</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Harwood</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Ong</surname>
            ,
            <given-names>C. S.</given-names>
          </string-name>
          ,
          <article-title>Discriminative concept learning network: reveal high-level differential concepts from shallow architecture</article-title>
          ,
          <source>in: Proceedings of 2015 International Joint Conference on Neural Networks, IJCNN</source>
          ,
          <year>2015</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>9</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Hinton</surname>
            ,
            <given-names>G. E.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Zemel</surname>
          </string-name>
          , R.S.:
          <article-title>Autoencoders, minimum description length and Helmholtz free energy</article-title>
          ,
          <source>Advances in Neural Information Processing Systems</source>
          ,
          <volume>6</volume>
          (
          <year>1994</year>
          )
          <fpage>3</fpage>
          -
          <lpage>10</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Ranzato</surname>
            ,
            <given-names>M.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Boureau</surname>
            <given-names>Y.-L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chopra</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          , LeCun, Y.:
          <article-title>A unified energy-based framework for unsupervised learning</article-title>
          ,
          <source>in: Proceedings of 11th International Conference on Artificial Intelligence and Statistics 2</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>371</fpage>
          -
          <lpage>379</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Bengio</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Courville</surname>
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vincent</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <article-title>Representation Learning: a review and new perspectives</article-title>
          ,
          <source>arXiv:1206</source>
          .5538 [cs.LG]
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Yoshida</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ohki</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <article-title>Natural images are reliably represented by sparse and variable populations of neurons in visual cortex</article-title>
          ,
          <source>Nature Communications</source>
          <volume>11</volume>
          (
          <year>2020</year>
          )
          <fpage>872</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Bao</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gjorgiea</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shanahan</surname>
            ,
            <given-names>L.K.</given-names>
          </string-name>
          et al.,
          <article-title>Grid-like neural representations support olfactory navigation of a two-dimensional odor space</article-title>
          ,
          <source>Neuron</source>
          <volume>102</volume>
          (
          <issue>5</issue>
          ) (
          <year>2019</year>
          )
          <fpage>1066</fpage>
          -
          <lpage>1075</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Holden</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Saito</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Komura</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Joyce</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <article-title>Learning motion manifolds with convolutional autoencoders, in: Proceeding of Asia 2015 Technical Briefs</article-title>
          ,
          <string-name>
            <surname>SA</surname>
          </string-name>
          '15 SIGGRAPH,
          <year>2015</year>
          , p.
          <fpage>18</fpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>