=Paper= {{Paper |id=Vol-2915/paper17 |storemode=property |title=Topology of Conceptual Representations in Unsupervised Generative Models |pdfUrl=https://ceur-ws.org/Vol-2915/paper17.pdf |volume=Vol-2915 |authors=Serge Dolgikh |dblpUrl=https://dblp.org/rec/conf/ivus/Dolgikh21 }} ==Topology of Conceptual Representations in Unsupervised Generative Models== https://ceur-ws.org/Vol-2915/paper17.pdf
Topology of Conceptual Representations in Unsupervised Generative Models

Serge Dolgikha,b
a
    National Aviation University, 1, Lubomira Huzara, Kyiv, 03058, Ukraine
b
    Solana Networks, 301 Moodie Dr., Ottawa, Canada


                 Abstract
                 In this work latent representations of image data were investigated with neural network models
                 of generative self-learning. A convolutional autoencoder with strong redundancy reduction
                 was used to create latent representations of images of basic geometric shapes in the process of
                 unsupervised generative learning and the characteristics of distributions of concept regions in
                 the latent space of models investigated. It was demonstrated that conceptual representations
                 with good separation of latent regions can be produced with generative models of limited
                 complexity and that characteristic types of data, or “concepts” form well-defined, continuous
                 and connected regions in the latent space. Geometric structure of the latent representations was
                 described in detail confirming connected and continuous topology of latent concept regions
                 clearly associated with characteristic types of observable data, such as shape, size and contrast.
                 The results indicate that conceptual representations created in the process of unsupervised
                 generative learning can form a natural basis for the emergence of abstract concepts in
                 intelligent systems.

                 Keywords 1
                 Machine learning, unsupervised learning, concept learning, representations

1. Introduction
    Representation learning with the objective to identify the informative structure in general real-world
data has a well-established record in the field of Machine Learning. Hierarchical representations of
different types of data were obtained with Restricted Boltzmann Machines (RBM), Deep Belief
Networks (DBN) [1][2], different flavors of autoencoders [3][4] and other models and architectures
allowed to improve accuracy of supervised learning. Different types, architectures and flavors of
generative models were investigated since including autoencoder neural networks, Generative
Adversarial Networks (GAN) and others [5][6][7].
    A number of interesting experimental results were obtained in concept learning with artificial
systems, such as the “cat experiment”, that demonstrated spontaneous emergence of concept sensitivity
on a single neuron level in unsupervised deep learning with image data [8]. Disentangled
representations were produced and studied with deep variational autoencoder models and different
types of image data [9] pointing at the possibility of a general nature of the effect. Concept-associated
structure was observed in latent representations of diverse real-world data such Internet in large public
networks, aerial surveillance images [10][11] and other results [12][13].
    The relations between learning and statistical thermodynamics was studied in the theory of learning
systems [14][15] leading to understanding of a deep connection between learning processes in artificial
systems and principles of information theory and statistical thermodynamics.
    These results demonstrated that training of artificial learning models under certain constraints such
as minimization of predictive error can lead to emergence of structure associated and correlated with
characteristic patterns in the observable data. This approach, known as representation learning, was
developed in a number of studies [16]. Of interest in this work is unsupervised conceptual representation

IVUS2021: Information Society and University Studies 2021, April 23, 2021, Kaunas, Lithuania
EMAIL: sdolgikh@nau.edu.ua (A. 1)
ORCID: 0000-0001-5929-8954 (A. 1)
              ©️ Copyright 2021 for this paper by its authors.
              Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
              CEUR Workshop Proceedings (CEUR-WS.org)
learning with models of generative learning that does not use explicitly labeled data but rather attempts
to minimize the error of reproducing the observed distribution from a representation in the latent space
created in the process of generative self-learning. As some of the earlier results have indicated, such
entirely unsupervised process can produce non-trivial structure in the latent representations that is
correlated with characteristic patterns in the observable data and can be used as a foundation for learning
methods and processes based on such structured low-dimensional representations.
    Interestingly, these observations in unsupervised learning of artificial systems were paralleled very
recently by several results in the studies of biologic sensory networks [17][18] that demonstrated
commonality of low-dimensional representations in processing sensory information by mammals,
including humans.
    Based on and inspired by these results in conceptual representation learning, the questions
investigated in this work are the following: what properties characterize latent representations of
successful generative models? Is there a stable association between characteristic patterns in the
observable data and the structure that emerges in the latent distributions in the process of generative
self-learning? What structure can be identified in the latent representations with entirely unsupervised
methods, without prior knowledge of conceptual content of the observable data?
    These objectives were approached with artificial neural network models of deep convolutional
autoencoder that showed effectiveness in producing informative representations [19] and a dataset of
images of basic geometric shapes that are described in the following sections.
    In conclusion, a brief clarification of terminology used throughout this work. We will refer to the
characteristic types or patterns in the dataset that are labeled with known types or classes as “external”
concepts, that signifies that the type or label of the input is defined outside of the model based on some
external or prior knowledge about the observable data. An example of an external concept for an image
with a geometric shape can be its general type, “a triangle”. In contrast, a characteristic structure in the
latent representation that can be identified reliably by unsupervised means without any external or prior
information, will be referred to as “internal” or “native” concept. Thus, a question of the relation
between the external and native concepts can be of interest as well.

2. Methods
   In this section the models and data used in the study are described. A dataset of basic shapes such
as circles, triangles and greyscale backgrounds, was used to produce low-dimensional latent
representations that were studied with several methods as described in the following sections.

2.1.    Deep Convolutional Autoencoder
   We used artificial neural network models with the architecture of convolutional autoencoder [3] with
added strong dimensionality reduction in the representation layer to produce three-dimensional latent
representations of image datasets of basic geometric shapes. The advantages of the autoencoder
architecture are that it allows to learn essential characteristics of the input data in an unsupervised
process without labeled samples, while virtually unlimited depths and complexity allows it to be used
with complex real-world data, demonstrating successful learning in many applications [8]-[11],[19].
   The architecture diagram of the models used in the study is presented in Figure 1.
Figure 1: Convolutional autoencoder with deep dimensionality reduction
   The models had three stages of convolution-pooling layers followed by several layers of
dimensionality reduction in a symmetrical layout. The total number of layers was 21, with
approximately 40,000 trainable parameters.
   Unsupervised training was performed with standard methods such as Stochastic Gradient Descent
[20] to minimize average distance between the input and generated output with binary cross-entropy
cost function. The models were implemented with Keras / Tensorflow [21] and several standard
machine learning libraries were used in the analysis of latent representations.
   The latent representation was produced by activations of the neurons in the central, encoding layer
of the model (Figure 1). The activations of neurons were interpreted as coordinates in a three-
dimensional latent representation space.

2.2.    Data
    Three datasets of geometric shape images with different variance and complexity were used to
investigate the structure of the latent representations. The images were greyscale, of the size 64 × 64
pixels.
    The first, generated dataset, Shapes-1 consisted of 600 greyscale images of circles, triangles and
greyscale backgrounds with two representative samples per type with difference in the size and contrast
of fore / background.
    The second generated dataset, Shapes-2 contained the total of 600 – 1,000 greyscale images of
circles, triangles and backgrounds with variation in size in the range 0.3 – 1.0 of the image size (that is,
0.3 × 64 pixels), with variation of contrast of fore- vs. background for each size.
    In the generated datasets the images were centered, symmetrical and had no rotation. Only darker
foregrounds relative to the background were used.
    The third dataset, Shapes-3 contained 1,500 images of randomly generated colored geometric shapes
and backgrounds varying in: size; position relative to the center of the image; rotation; and the color of
the fore- and background [22]. The images in this dataset were converted to greyscale.

2.3.    Latent Representations
    Following the process of unsupervised training in which significant reduction in the value of cost
function was observed, the models were able to perform two essential transformations of data:
    The encoding transformation, from the input data space, that is, images to the three-dimensional
latent representation, with coordinates represented by activations of neurons in the central, “encoding”
layer of the model (Figure 1):
                                      𝑟𝑥 = 𝐸(𝑥) = 𝑒𝑛𝑐𝑜𝑑𝑒(𝑥)                                             (1)
The generative transformation operates in the opposite direction, i.e. from the latent representation into
the observable (image) space and is performed by the generative part of the autoencoder:
                                    𝑔𝑦 = 𝐺(𝑦) = 𝑔𝑒𝑛𝑒𝑟𝑎𝑡𝑒(𝑦)                                             (2)
where y is a latent position, i.e. a point in the latent space represented by latent coordinates (y1, y2, y3).
     Transformations of encoding and generation allow to study the structure in the latent representation
by measuring and visualization of the distributions of samples encoded to the latent space, as well as
generative properties of the models via producing images of latent positions and regions of interest.

3. Results
   The subject of this investigation was topological structure in the latent representations created by
generative models in the process of unsupervised learning. To verify the hypothesis of the study,
methods of analysis of latent distributions and generative ability of models were applied as described
in this section. It is essential to note that the process of production of latent representations via
generative learning was entirely unsupervised and no labeled data was used in configuration and / or
training of the models.

3.1.    Structure of Latent Representations
    Latent representations of generative models displayed well structured, connected and continuous
distribution of data by characteristic type as illustrated in Figure 2. The visualization data, a subset of
Shapes-1 and Shapes-2 datasets described in Section 2.2 represented a selection of images with
variation of size and greyscale contrast, color-coded by characteristic type (i.e., circle, triangle or
greyscale background).
    Visualization of latent distributions by characteristic type showed that the shapes present in the
training dataset were assigned specific well-defined regions in the latent space, with latent coordinates
encoding essential characteristics of shapes, such as size and contrast.




Figure 2: Latent distributions by shape type. Models: Shapes-1 (left), Shapes-2 (right). Legend:
magenta: circles; green: triangles; blue/red: greyscale background.
    The observed pattern of latent distributions was consistent between different individual models
training with the same dataset, that was verified with several instances of trained models. These findings
indicate that the emergent concept-associated structure in the latent representations of generative
models reflects essential general characteristics of the data in the training data and is to a large extent
invariant to individual model.

3.1.1. Connectedness of Latent Representations
    As discussed above, latent representations created by generative models in unsupervised training
displayed connected and continuous topology allowing to conclude that there exist an association
between well-defined latent regions and characteristic types, or general concepts in the observable data.
This conclusion was further substantiated with two experiments.
    In the first experiment, a set of images S of the same type was selected, with variation in
characteristics, such as size and contrast. The images were transformed to the latent space as in (1) and
the mean of the resulting set of latent coordinates calculated as:
                                      𝑟𝑚𝑒𝑎𝑛 = 𝑚𝑒𝑎𝑛(𝐸(𝑆))                                            (3)
The mean latent representation rmean of the input set was then propagated to the observed space with
generative transformation (2) and the resulting image produced. In the experiments, the generated image
of the input set was of the same type (Table 1).
Table 1
Generated mean of input set of shape type
             Input set                  Encoded mean (example)                      Result
 Circles-2                                                                           Circle
 Circles-3                                 (-1.979, -1.635, -0.817)                  Circle
 Triangles-2                                                                       Triangle
 Triangles-3                              (1.437, 0.073, -1.358)                 Triangle*
 Background-3                             (3.974, 0.011, -0.176)                Background
   *
      An exception was observed in the boundary region (Section 3.1.2), where the mean of a certain
input set produced mixed result.
     In the second group of experiments, a set of points in a close neighborhood of a selected latent
position of a given type was selected, and images corresponding to these positions generated with (2).
The resulting images were observed to be of the same type as that of the position of origin (Figure 3).




Figure 3: Latent neighborhood experiment. Position of origin: leftmost.
    These experiments confirmed the connected and continuous topology of latent representations
produced by generative models.

3.1.2. Boundary Regions
   Well defined character of latent regions associated with characteristic types of shapes was confirmed
by observation of boundary areas between the regions of different types, that produced generated output
of mixed form with features from both types. The examples of such boundary generative regions are
shown in Figure 4.




Figure 4: Latent boundary areas. Left to right: circle/triangle; circle/background; triangle/background.
   Observation of the boundary areas between the regions associated with characteristic shapes
supports the hypothesis of a continuous and connected topology of the latent space.

3.2.    Semantics of Latent Coordinates
    Generative transformation (2) allows to investigate semantical meaning of the latent coordinates
defined as activations of neurons in the encoding layer of the generative model. A method of latent
probing was developed that allowed to produce and evaluate generated image of a specified position in
the latent space, after conceptual latent representation was produced in the unsupervised learning phase.
By applying probing to latent representations of generative models the structure that formed in the
process of generative learning under the constraints discussed earlier can be studied in practically any
level of detail.
    In the input to the method, a set of coordinates of a latent position of interest was provided. The
position was propagated via generative part of the autoencoder with generative transformation (2)
producing an output as an observable image. Evaluation of generated images associated with a set of
latent positions, such as a neighborhood, a line, a surface and so on, in a latent region of interest allowed
to obtain essential observations on the semantical meaning of the latent coordinates. The examples of
probing along the latent axes are shown in Figure 5.
Figure 5: Generated output along latent axes, 0 – 2.
   The results of generative probing in the latent space of generative models in this section support the
manifold assumption [23] and confirm that latent coordinates in low-dimensional representations can
be associated with essential characteristics of input data, such as in this case, the size, contrast and the
type of the shape. It can be seen also that the semantics of latent coordinates are local, rather than global:
the same axis, at different latent position can signify either of these characteristics. Consequently,
investigation of topological structure of latent representations, including with more complex data merits
further attention.

4. Conclusion
    In this work latent representations of images representing basic geometric shapes were analyzed
with methods of unsupervised machine learning. The analysis produced a number or essential results
and observations.
    It was demonstrated that successful generative learning with stable association of characteristic
observable inputs to a structure in the low-dimensional latent representations can be achieved with
models of limited complexity, both in size and architecture, well within a range of simple biologic
systems. The complexity of models in this work was roughly equivalent to a nervous system of a
jellyfish that has a comparable number of neurons and synapses [24].
    By applying unsupervised methods that did not use any externally labeled data in the analysis of
latent representations it was demonstrated that generative learning can lead to emergence of well-
defined latent regions associated with essentially different types of observable data. These results points
to possible origin of higher-level abstract concepts in unsupervised generative learning under the
constraints of generative accuracy and redundancy reduction, both having clear evolutionary advantage
for the learner.
    The structure of latent regions associated with characteristic patterns, native concepts in the
observable data was investigated and described, confirming connectedness and continuous topology of
the latent regions, with distinct boundaries between different concept regions. An analysis of generative
capacity of models with the developed method of latent probing allowed to make essential observations
on the semantics of the latent coordinates in the representations produced by the models, showing an
association with essential characteristics of the data such as type, size and contrast. However, as the
analysis of the semantics of latent coordinates has shown (Section 3.2) there may not be a fixed global
semantics to the latent coordinates and their significance is determined by the locality, i.e. the position
and the region, again confirming manifold-like topology of latent representations emergent in
generative self-learning.
    Overall, the results of this work demonstrated that conceptual representations can be a natural result
of unsupervised observation of the environment under the constraints that are both natural and common
for a learning system [25], of artificial or biologic nature.

5. Future Work
   Following the direction outlined in this work, further studies can focus on investigation of latent
representations of more complex visual data with greater variation of conceptual content in the
observable environment. It is likely that advances in this direction would require models of considerably
greater complexity, in size, depth and architecture [8]. For example, investigation of neural network
models with sparsity constraints can be a promising direction due to apparent success of this architecture
in the sensory input processing neural networks of biologic systems as reported in recently published
results [17].


6. References

[1] Hinton, G.E., Osindero, S., Teh Y.W., A fast learning algorithm for deep belief nets, Neural
     Computation 18 (7) (2006) 1527–1554.
[2] Fischer, A., Igel, C., Training restricted Boltzmann machines: an introduction, Pattern Recognition
     47 (2014) 25–39.
[3] Bengio, Y., Learning deep architectures for AI, Foundations and Trends in Machine Learning 2
     (1) (2009) 1–127.
[4] Coates, A., Lee, H., Ng, A.Y., An analysis of single-layer networks in unsupervised feature
     learning, in: Proceedings of 14th International Conference on Artificial Intelligence and Statistics
     15, 2011, pp. 215–223.
[5] Welling, M. and Kingma, D.P., An introduction to variational autoencoders, Foundations and
     Trends in Machine Learning 12 (4) (2019) 307–392.
[6] Creswell A., White T., Dumoulin V., Arulkumaran K., Sengupta B. and Bharath A.A., Generative
     adversarial networks: an overview, IEEE Signal Processing Magazine 35 (1) (2018) 53–65.
[7] Partaourides, H., Chatzis, S.P., Asymmetric deep generative models, Neurocomputing 241 (2017)
     90 – 96.
[8] Le, Q.V., Ransato, M. A., Monga, R. et al., Building high level features using large scale
     unsupervised learning. arXiv 1112.6209 [cs.LG] 2012.
[9] Higgins, I., Matthey, L., Glorot, X., Pal, A. et al., Early visual concept learning with unsupervised
     deep learning. arXiv 1606.05579 [cs.LG] 2016.
[10] Dolgikh, S., Categorized representations and general learning, in: Advances in Intelligent Systems
     and Computing, volume 1095, Springer, Cham 2019, pp. 93 – 100. doi:10.1007/978-3-030-35249-
     3_11.
[11] Prystavka P., Cholyshkina O., Dolgikh S., Karpenko D., Automated object recognition system
     based on aerial photography, in: Proceedings of 10th International Conference on Advanced
     Computer Information Technologies, ACIT-2020, Deggendorf, Germany, 2020, pp. 830 – 833.
[12] Rodriguez, R.C., Alaniz, S., and Akata, Z., Modeling conceptual understanding in image reference
     games, in: Proceedings of Advances in Neural Information Processing Systems, Vancouver, 2019,
     pp. 13155–13165.
[13] Wang, Q., Young, S., Harwood, A., and Ong, C. S., Discriminative concept learning network:
     reveal high-level differential concepts from shallow architecture, in: Proceedings of 2015
     International Joint Conference on Neural Networks, IJCNN, 2015, pp. 1–9.
[14] Hinton, G. E. and Zemel, R.S.: Autoencoders, minimum description length and Helmholtz free
     energy, Advances in Neural Information Processing Systems, 6 (1994) 3 – 10.
[15] Ranzato, M.A., Boureau Y.-L., Chopra, S., LeCun, Y.: A unified energy-based framework for
     unsupervised learning, in: Proceedings of 11th International Conference on Artificial Intelligence
     and Statistics 2, 2017, pp. 371–379.
[16] Bengio, Y., Courville A., Vincent, P., Representation Learning: a review and new perspectives,
     arXiv:1206.5538 [cs.LG] 2014.
[17] Yoshida, T., Ohki, K., Natural images are reliably represented by sparse and variable populations
     of neurons in visual cortex, Nature Communications 11 (2020) 872.
[18] Bao, X., Gjorgiea, E., Shanahan, L.K. et al., Grid-like neural representations support olfactory
     navigation of a two-dimensional odor space, Neuron 102 (5) (2019) 1066–1075.
[19] Holden, D., Saito, J., Komura, T. and Joyce, T., Learning motion manifolds with convolutional
     autoencoders, in: Proceeding of Asia 2015 Technical Briefs, SA '15 SIGGRAPH, 2015, p.18.
[20] Spall, J. C., Introduction to stochastic search and optimization: estimation, simulation, and control,
     Wiley Hoboken, NJ, ISBN 0-471-33052-3 2003.
[21] Keras: Python deep learning library, https://keras.io/, last accessed: 2020/11/21.
[22] El Korchi, A., 2D geometric shapes dataset, Mendeley Data V1, 2020,
     doi:10.17632/wzr2yv7r53.1.
[23] Zhou X., Belkin M., Semi-supervised learning. In: Academic Press Library in Signal Processing,
     Elsevier 1, 2014, pp. 1239–1269.
[24] Garm, A., Poussart, Y., Parkefelt, L., Ekström, P., Nilsson, D-E., The ring nerve of the box jellyfish
     Tripedalia cystophora, Cell Tissue Research 329 (1) (2007) 147–157.
[25] Dolgikh, S., Why good generative models categorize? International Journal of Modern Education
     and Computer Science (2021). To appear.