=Paper=
{{Paper
|id=Vol-2254/invited-2
|storemode=property
|title=Opportunities
and challenges in deep generative models
|pdfUrl=https://ceur-ws.org/Vol-2254/invited-2.pdf
|volume=Vol-2254
|authors=Evgeny I. Nikolaev
}}
==Opportunities
and challenges in deep generative models==
Opportunities and challenges in deep generative models Evgeny I. Nikolaev Institute of Information Technologies and Telecommunications North-Caucasus Federal University, Stavropol, Russia notdeveloper@gmail.com Abstract A Generative Model is a powerful way of learning any kind of data dis- tribution using unsupervised learning and it has achieved tremendous success in just few years. Though there are several approaches to design information systems for generating synthetic data, wich are referred to as Deep Generative Model (DGM). Since then, DGM has become a trending topic both in academic literature and industrial applications. It is also receiving increasing attention in machine learning competi- tions. This paper aims to provide an overview of the current progress towards DGM, as well as discussing its various applications and open problems for future research. Moreover, we discuss some research we conducted during last years that may extend the existing state of the art approaches in synthetic data generation or improving existing deep models. 1 Introduction Generative models have a long history and recent methods have combined the generality of probabilistic reasoning with the scalability of deep learning to develop learning algorithms that have been applied to a wide variety of problems giving state-of-the-art results in image generation, text-to-speech synthesis, image captioning and data augmentation, amongst many others. Advances in deep generative models are at the forefront of deep learning research because of the promise they offer for allowing data-efficient learning, and for model-based reinforcement learning. The latest advances in generative modeling include following types of models: Markov models, latent variable models and implicit models. These models can be scaled to high-dimensional data. From the historical perspective it is important to mention research linked with Boltzmann machine [Hinton84, Hinton86], Restricted Boltsmann Machine [Smol86, Long10], Deep Belief Network [Hinton06, Hinton08], Deep Boltzmann Machine [Salakh09], Convolutional Boltzmann Machine [Desj08]. In recent years, there has been resurgence of interest in deep generative models. Emerging approaches such as Variational Autoencoders [Kingma13, Rezend14], Generative Adversarial Networks [Goodfellow16], auto-regressive networks (pixelRNNs [Oord16], RNN language models [Zarem14]), and many of their variants and extensions have led to impressive results in different applications. Researchers are making great progress in generating realistic high-resolution images [Woo17], manipulating and changing text, learning interpretable data representations, automatically augmenting data for training the models. Copyright c by the paper’s authors. Copying permitted for private and academic purposes. In: Marco Schaerf, Massimo Mecella, Drozdova Viktoria Igorevna, Kalmykov Igor Anatolievich (eds.): Proceedings of REMS 2018 – Russian Federation & Europe Multidisciplinary Symposium on Computer Science and ICT, Stavropol – Dombay, Russia, 15–20 October 2018, published at http://ceur-ws.org 2 Generative models All types of generative models aim at learning the true data distribution of the training set so as to generate new data points with some variations. But it is not always possible to learn the exact distribution of our data either implicitly or explicitly and so we try to model a distribution which is as similar as possible to the true data distribution. For this, we can leverage the power of neural networks to learn a function which can approximate the model distribution to the true distribution. Two of the most commonly used and efficient approaches are Variational Autoencoders (VAE) and Generative Adversarial Networks (GAN). VAE aims at maximizing the lower bound of the data log-likelihood and GAN aims at achieving an equilibrium between Generator and Discriminator. Many researches attempt to compile a unified view: new formulation of GANs and VAEs, and linked back to the classic variational inference algorithm and the wake-sleep algorothm. 2.1 Variational Autoencoder The variational autoencoder [Kingma13, Rezend14] is a directed model that uses learned approximate inference and can be trained purely with gradient-based methods. An autoencoder can be used to encode an input image to a much smaller dimensional representation which can store latent information about the input data distribution. The variational autoencoder approach is theoretically pleasing and simple to implement. It also obtains excellent results and is among the state-of-the-art approaches to generative modeling. One of the main features is that samples from VAEs trained on images tend to be somewhat blurry. The causes of this phenomenon are not yet known. The key idea of VAE are shown in Fig. 1. z ~ N(0, I) q X Figure 1: Mapping latent vector to data distribution using parameter The primary objective is to model the data X with some parameters which maximizes the likelihood of training data X. In short, we are assuming that a low-dimensional latent vector has generated our data x(x ∈ X) and we can map this latent vector to data x using a deterministic function f (z; θ) parametrized by theta θ which we need to evaluate. During generative process, our aim is to maximize the probability of each data in X. 2.2 Generative Adversarial Network Generative adversarial networks, or GANs [Goodfellow16], are another generative modeling approach based on differentiable generator networks. Adversarial training has completely changed the way we train the artificial neural networks. GAN dont link with any explicit density estimation like VAE. GAN is based on game theory approach with an objective to find equilibrium between the two networks: generator and discriminator. The aim is to sample from a simple distribution and then learn to transform this noise to data distribution using approximators such as neural networks. This approach is shown in Fig 2 Input Generator noise z Train images Generate fake (Real) images G(z) Discriminator Real / Fake Figure 2: Training Generative Adversarial Network We can formulate learning in GAN as a zero-sum game, in which a function v(θ(g) , θ(d) ) determines the payoff of the discriminator. During learning, each player attempts to maximize its own payo, so that at convergence g ∗ = arg ming maxd v(g, d). One of the earliest model on GAN employing Convolutional Neural Network (CNN) is Deep Convolutional Generative Adversarial Networks (DCGAN) [Radf17]. 3 Conclusions and Future Work Unsupervised learning is a next frontier in artificial intelligence. One of the main advantages of Generative Mod- els is a possibility of the training in semi-supervised manner. Such models can be applied for solving complex problems: text-to-image translation, synthetic images generation, solving problems with multimodal data dis- tribution, drug discovery, visual marks retrieval from images. DGM is a way to improve existing discriminative models. GANs help to solve the one of the main challenges in deep learning: huge amount of labelled data. These models help in building a better future for machine learning. References [Hinton84] G. E. Hinton, T. J. Sejnowski and D. H. Ackley Boltzmann Machines: Constraint satisfaction networks that learn. Technical Report CMU-CS-84-119, Carnegie-Mellon University. [Hinton86] G. E. Hinton, T. J. Sejnowski Learning and relearning in Boltzmann machines. In Rumelhart, D. E. and McClelland, J. L., editors, Parallel Distributed Processing: Explorations in the Microstructure of Cognition. Vol. 1: Foundations, MIT Press, Cambridge, MA. pp 282-317 [Smol86] P. Smolensky Information processing in dynamical systems: Foundations of harmony theory. Parallel distributed processing: Explorations in the microstructure of cognition, MIT Press, Cambridge, MA, (1986). [Long10] P. Long, R. Servedio. Restricted Boltzmann Machines are Hard to Approximately Evaluate or Simulate. 27th International Conference on Machine Learning (ICML). 2010. [Hinton06] G. E. Hinton, S. Osindero and Y. Teh A fast learning algorithm for deep belief nets. Neural Computation. Vol. 18, 2006, pp. 1527-1554. [Hinton08] G. E. Hinton, R. Salakhutdinov Using Deep Belief Nets to Learn Covariance Kernels for Gaussian Processes. Advances in Neural Information Processing Systems. Vol. 20, 2008. [Salakh09] G. E. Hinton, R. Salakhutdinov Deep Boltzmann Machines. To appear in Artificial Intelligence and Statistics. 2009. [Desj08] G. Desjardins, Y. Bengio Empirical evaluation of convolutional RBMs for vision. Technical Report 1327, Dpartement dInformatique et de Recherche Opra-tionnelle, Universit de Montral. 2008. [Kingma13] D. Kingma, M. Welling Auto-Encoding Variational Bayes. Proc. of the 2nd Int. Conf. on Learning Representations (ICLR), arXiv: 1312.6114 [stat. ML]. [Rezend14] D. J. Rezende, S. Mohamed, D. Wierstra Stochastic Backpropagation and Approximate Inference in Deep Generative Models. Proc. of the 31st Int. Conf. on Machine Learning (ICML), vol. 32, 2014. [Goodfellow16] Goodfellow Nips 2016 tutorial: Generative adversarial networks. NIPS 2016, arXiv:1701.00160, 2016. [Oord16] Aron van den Oord, N. Kalchbrenner, K. Kavukcuoglu Pixel Recurrent Neural Networks. International Conference on Machine Learning (ICML). 2016. [Zarem14] W. Zaremba, I. Sutskever and O. Vinyals Recurrent Neural Network Regularization. arXiv:1409.2329 [cs.NE]. 2014. [Woo17] B. Wu, H. Duan, Zh. Liu, G. Sun SRPGAN: Perceptual Generative Adversarial Network for Single Image Super Resolution. arXiv:1712.05927 [cs.CV]. 2017. [Radf17] A. Radford, L. Metz, S. Chintala Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks. Int. Conf. on Learning Representations (ICLR). 2016.