<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Using WGAN for Improving Imbalanced Classi cation Performance</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Snehal Bhatia</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Rozenn Dahyot</string-name>
          <email>dahyotrg@tcd.ie</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>School of Computer Science and Statistics Trinity College Dublin</institution>
          ,
          <country country="IE">Ireland</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>This paper investigates data synthesis with a Generative Adversarial Network (GAN) for augmenting the amount of data used for training classi ers (in supervised learning) to compensate for class imbalance (when the classes are not represented equally by the same number of training samples). Our data synthesis approach with GAN is compared with data augmentation in the context of image classi cation. Our experimental results show encouraging results in comparison to standard data augmentation schemes based on image transforms.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Image classi cation is a standard process in image processing and many
machine (deep) learning techniques are routinely evaluated [14] using labelled
images datasets (e.g. FMNIST [16], CIFAR-10 [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]). The occurrence of imbalance
in datasets collected from real-life domains is sometimes unavoidable due to the
cost of collecting and labelling data, or due to privacy issue, or simply due to
rare event scenarios. This imbalance of number of training examples per class
often has a detrimental e ect on the performance of classi ers [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. In this
research we explore the potential of data synthesis with GANs to address this
issue, more speci cally we compare data synthesis with the Wasserstein
Generative Adversarial Networks (WGANs) against the traditional data augmentation
approach traditionally used for training deep learning architectures. GANs and
WGANs are rst introduced in Section 2, and our approach is introduced next
(Sec. 3). We show experimentally that WGAN data augmentation outperforms
traditional data augmentation with image transforms on both FMNIST and
CIFAR-10 datasets (Sec. 4).
The Generative Adversarial Network (GANs) framework was rst introduced
for arti cially generating realistic images from scratch [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Since then, GANs
have been employed for a variety of image processing and computer vision tasks,
such as generating high resolution images from low resolution input images [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ],
texture synthesis in images [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] and human face synthesis [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
      <p>
        This generative capacity of GANs makes them suitable for the purpose of data
augmentation, and many recent studies have shown how to tackle the problem
of imbalanced datasets in classi cation by using variations of the GAN
architecture. Mariani et al [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] introduced a new architecture called "BAGAN" or
Balancing GAN, and achieved a signi cantly better classi cation performance
in comparison to ACGAN [11] and simple GAN [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], when tested on various
arti cially imbalanced distributions of the image datasets of MNIST, CIFAR-10,
Flowers and GTSRB datasets. Mariani et al [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]'s original work with GAN
(introduced Paragraph 2.1) is extended in our paper by evaluating the performance of
Wasserstein GAN (WGAN and WGAN-GP [
        <xref ref-type="bibr" rid="ref2 ref4">2, 4</xref>
        ], explained in Paragraph 2.2)
with the same methodology (Sec. 3 and 4).
2.1
      </p>
      <sec id="sec-1-1">
        <title>Generative Adversarial Networks</title>
        <p>
          The rst and original GAN architecture introduced by Ian Goodfellow et al [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ]
consists of two sub-networks, which are competing Arti cial Neural Networks
(ANNs), namely the `Generator' (G ) and the `Discriminator' (D). The Generator
learns to transfer the distribution of input data (e.g. distribution of noise images)
into a target distribution (e.g. distribution of images of horses). At the same time,
the Discriminator, which is essentially a binary classi er, is trained to distinguish
between the real data points (e.g. real images of horses) and arti cially generated
data points by the generator (e.g. synthesised images of horses created by the
generator transfer function).
        </p>
        <p>
          A GAN [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ] is often de ned as a two-player minimax game in which Generator
G wants to minimize the cost function whereas the discriminator D aims to
maximize it:
min max Ex pr [log D(x )] + Ex~ pg [log(1
G D
        </p>
        <p>
          D(x~)]
(1)
where x~ = G (z ) with z p(z ) (input of the generator sampled from a simple
noise distribution such as Uniform or Normal), and pr is the (real) data
distribution [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]. The GAN value function (Equation (1)) is essentially the
JensenShanon Divergence between the real data and the arti cially generated data.
The training process of a GAN can be viewed as a double feedback-loop where
the discriminator is in a feedback loop with the real images and the generator
uses the feedback from the discriminator to learn how to produce images that
are realistic enough to fool the discriminator. The feedback process is repeated
throughout the training phase, until Nash Equilibrium is achieved1. This process
is called Adversarial Training.
        </p>
        <p>
          Although Vanilla GANs have achieved state of the art results in many domains,
they su er from certain drawbacks:
1 In Game Theory, when multiple interacting, non-cooperating participants are
involved, Nash Equilibrium is the state of stability achieved when no participant can
bene t solely from changing its own strategy or actions if the other players' strategies
remain constant.
{ Nash Equilibrium is Hard to Achieve: The training process of GAN
is based on gradient descent. The two models, generator and discriminator,
are simultaneously trained to nd a Nash Equilibrium. However, since both
models update their loss functions concurrently and independently, there is
no guarantee of convergence
{ Vanishing Gradient Problem: Training a GAN loss function poses a
dilemma. If the discriminator is trained perfectly (especially early on in the
training process), then D(xreal)=1 and D(xreal)=0. From Equation (1)
it can be observed that in this case, the value of the loss function would
become 0, and there would be no gradient left to update during the training
iterations, hence leading to the vanishing gradient problem. However, if the
discriminator function is not trained to perfection then the generator would
not receive relevant feedback, meaning that the learned loss function would
not be good enough to generate realistic images
{ Mode Collapse: This is a common failure observed in GANs where the
generator achieves a state where it always produces the same images as
outputs. This may in some cases be enough to fool the discriminator, but
the low variety of images generated is not representative of the complexities
observed in real-world data distribution [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]
Therefore, various modi cations of the original GAN have been proposed, one
of which is introduced below.
2.2
        </p>
      </sec>
      <sec id="sec-1-2">
        <title>Wasserstein GAN (WGAN) and WGAN-GP</title>
        <p>
          The WGAN architecture proposed by Arjovsky et al [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ] replaces the
JensenShannon divergence from the original GAN architecture [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ] with the Wasserstein
Distance function. The Wasserstein Distance is also called the "Earth Mover
Distance", as it can be informally interpreted as the minimal cost of moving and
transforming some quantity of mass (say, a pile of dirt) from the shape of one
probability distribution P to that of another probability distribution Q. The
cost of moving in this scenario is calculated as the product of the amount of
mass moved and the distance by which it has been moved. W (P, Q) which is
a measure of distance between the points in probability distributions P and Q.
The WGAN value function corresponds to:
min max Ex pr [D(x )]
G D2D
        </p>
        <p>
          Ex~ pg [(D(x~)]
(2)
In Equation (2), D denotes the set of 1-Lipschitz functions2, meaning that the
discriminator loss should follow the Lipschitz constraint. Gulrajani et al [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]
extended WGAN to WGAN-GP (gradient penalty) to improve the training of
WGANs. The Weight-Clipping method used to enforce the Lipschitz Constraint
introduces numerous problems such as vanishing gradient (when the clipping
2 A Lipschitz function is a function f such that |f(x)-f(y)|
x and y, where K is a constant independent of x and y.
        </p>
        <p>
          K|x-y| for all
window is too large), slow convergence (when the clipping window is too small)
and it is not very suitable for very complex data [
          <xref ref-type="bibr" rid="ref2 ref4">2, 4</xref>
          ]. One solution is to impose
a Gradient Penalty instead of weight clipping as a means to enforce Lipschitz
constraint [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]. The value function for WGAN-GP can be observed in Equation
(3). Here, L represents the loss function, x' represents a sample from fake or
generated data, and x^ represents randomly sampled data. Note that a "soft penalty"
is imposed (i.e. only on the randomly sampled data) to prevent tractability
issues. The last term in the equation is the penalty term, with being the penalty
coe cient.
        </p>
        <p>min max Ex~ pg [D(x~)]
G D2D</p>
        <p>Ex pr [D(x )] +</p>
        <p>
          Ex^ px^ [(krx^D(x^)k2
(3)
px^ is de ned for sampling uniformally on straight edges (see [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] for details ).
Generally, for a 1-Lipschitz function, the maximum gradient norm should be 1.
Therefore, instead of applying weight-clipping, WGAN-GP loss function imposes
a penalty if the gradient norm moves away from its maximum target norm value
of 1.
3
        </p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>Method</title>
      <p>
        In this study, we compared the traditional image data augmentation technique
of using geometric and photometric transformations [13], with our proposed
technique of using Generative Adversarial Networks (speci cally, WGAN-GP [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ])
for the same purpose. We rst arti cially introduce imbalance in two benchmark
balanced datasets of FMNIST [16] and CIFAR-10 [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], and the class distribution
can be observed in Figure 1.
      </p>
      <p>FMNIST
CIFAR-10
To e ectively study the e ects of dataset imbalance on classi cation
performance, we use a Hybrid CNN-SVM architecture. The overall work ow of the
study can be observed in Figure 2. The classi er is separately trained on each
of these datasets and these trained models are used for evaluation. However, no
change is made to the test set of these datasets.
Some of the most common geometric and photo-metric transformations applied
on images to generate additional data while preserving the context of the image
are rotation, translation, shearing, scaling, ipping, zooming, blurring, whitening
etc [13]. This has been achieved using ImageDataGenerator class of the image
pre-processing module of Keras, and a sample of images generated using this
methodology can be observed in Figure 3.
3.2</p>
      <sec id="sec-2-1">
        <title>Oversampling using WGAN-GP</title>
        <p>
          WGAN-GP [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ] is a superior GAN architecture which is known to achieve
convergence without facing the issues of vanishing or exploding gradients. The training
process is very stable for this architecture, and it has also proven to generate
highly diverse data samples with low noise. It can achieve high quality results
with almost no hyperparameter tuning, making it a suitable choice for a wide
variety of applications and datasets. Therefore, we have adopted the WGAN-GP
architecture for our study.
        </p>
        <p>The general methods for GAN architecture construction and the choices made
in our model are described below:
{ Use of Leaky ReLU: The Recti ed Linear Unit (ReLU) activation
function is a widely adopted, e cient activation function that returns the input
directly as the output, or returns 0 when the input is 0.0 or less. However,
the best practice for GANs is to use a variation called LeakyReLU, which
allows some values lesser than 0, and learns the optimum cut-o for each
node. In our architecture, we have used LeakyRelu in both the generator
as well as discriminator, with slope values being of the order of the default
value, 0.2.
{ Use of Batch Normalization: Batch normalization is a technique used to
improve the speed, performance and stability of neural networks by
normalizing the input layer by adjusting and scaling the activations. We employ
batch normalization after the convolution layers. In the case of GANs, it
helps avoid vanishing and exploding gradients, as well as mode collapse.
{ Using Gaussian Weight Initialization: Before starting the training
process, the weights (parameters) of the neural network must be initialized with
small random variables to prevent the activation layer from producing
vanishing or exploding outputs, which would cause very small or very large
gradient updates, giving rise to convergence problems. It is considered to be
a good practice to initialize all weights using a zero-centred Gaussian
distribution, with mean value as 0 and variance value as 1/N, where N speci es
the number of input neurons. Therefore, we have used Xavier initialization
in our architecture, which is based on the same principle.
{ Not using Max Pooling: A max-pooling layer is often used in CNNs
to after each convolution layer to downsample the input and feature maps.
However, we would not be using this approach as in the case of Generative
Adversarial Networks, it has been shown that having all convolutional layers
allows the network to learn its own spatial down-sampling, which leads to
an increase in performance.</p>
        <p>
          Qualitative Evaluation Metrics for WGAN-GP Since our goal is to use
the WGAN-GP model to generate additional minority class images in order to
augment an imbalanced dataset and balance it, we aim to ful l the following
criterion for our generated images through visual inspection [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ]:
{ Generated images should be similar to the other images of the class in
question. If this target is not met, it would mean that the generator is not trained
enough to produce quality, realistic images.
{ Generated images must not all be the same, or repetitive. This will ensure
that the generator does not su er from mode collapse problem.
{ Generated images should be di erent from the images which are already
present in the training set. If not, it would mean that we have simply trained
our generative model to repeat the training data.
        </p>
        <p>In this study, the visual inspection has been done by the researchers themselves,
by randomly mixing real and generated samples and testing if the fake samples
can be spotted in that mix. The generated images sampled at various epochs of
WGAN-GP training phase for the class horse can be observed in Figure 5.
3.3</p>
      </sec>
      <sec id="sec-2-2">
        <title>Classi cation Architecture</title>
        <p>We adopted the hybrid CNN-SVM architecture as the classi er as proposed by
Niu and Suen [10]. In this hybrid architecture, the original CNN (with the output
layer) is trained on the input dataset until convergence is achieved. Then, the
output layer is replaced with the Radial Bias Function (RBF) of SVM. The
output from the CNN hidden layer is taken as a feature vector for training the
SVM. Once trained, this SVM is able to perform the classi cation task on unseen
data.
4</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Experimental Results</title>
      <p>New data samples generated by WGAN-GP (Figure 4) result in a superior
augmented (balanced) dataset as compared to the dataset obtained by
augmentation with images generated using geometric transforms (Figure 3). This can be
attributed to the following qualities of WGAN-generated images:
{ Realistic looking, and impossible to tell apart from the original dataset
images of the same class by a human observer
{ Preservation of context or semantic information of the class in question
{ Samples are not repetitive, signifying that there is very less possibility of
over- tting
{ Samples generated are variable or diverse in nature, therefore resulting in
an e ciently augmented dataset which contains samples representing many
possibilities</p>
      <p>Original Image</p>
      <p>New images generated using Image
Transformations
The results of classi cation on the CNN-SVM architecture are reported in Table
1 when using FMNIST dataset. We note that imbalance causes the classi er's
(a) Original Images
(b) Synthesised Images Generated using WGAN-GP
performance to decrease (e.g. Accuracy drops by 5.25% , and similarly for the
F1-Score). While using traditional image transforms results in an improvement
of approximately 4% in testing Accuracy, and of around 4.3% in the F1-Score, it
can be observed that WGAN-GP does a better job and increases the Accuracy
and F1-Score by 5% and 4.5% in comparison to the imbalanced dataset.
The results of classi cation on the CNN-SVM architecture are reported in Table
2 for the dataset CIFAR-10. While the class imbalance creates a 8% drop in
accuracy, data augmentation approaches both manage to restore the performance
with an increase of 5% for Image transforms and 6% for WGAN.</p>
    </sec>
    <sec id="sec-4">
      <title>Conclusion</title>
      <p>We have shown the potential bene t of using WGAN for improving the
performance of a classi er when the training dataset su ers from class imbalance.
For further testing, it would not only be vital to observe the performance of the
classi er with di erent distributions of imbalanced classes, but to also take into
account the impact of noise, dataset shift problem (where the train and test sets
follow di erent distributions) and the cases where there is possibility of class
overlapping. It would also be useful to introduce quantitative metrics to assess
the images generated by GANs, such as SSIM (Structural Similarity Index) [15],
Inception Score [12], GAN Quality Index [17], which would reduce or eliminate
the need for human visual inspection.</p>
      <p>Acknowledgements. This work has emanated from research conducted as part
of the MSc in Computer Science programme (Future Networked Systems) at
Trinity College Dublin, Ireland. This research is partly supported by the ADAPT
Centre for Digital Content Technology funded under the SFI Research Centres
Programme (Grant 13/RC/2106) and co-funded under the European Regional
Development Fund.
10. Niu, X.X., Suen, C.Y.: A novel hybrid cnn{svm classi er for recognizing
handwritten digits. Pattern Recognition 45(4), 1318{1325 (2012)
11. Odena, A., Olah, C., Shlens, J.: Conditional image synthesis with auxiliary classi er
GANs. In: Proceedings of the 34th International Conference on Machine Learning.
vol. 70, pp. 2642{2651. PMLR (06{11 Aug 2017)
12. Salimans, T., Goodfellow, I., Zaremba, W., Cheung, V., Radford, A., Chen, X.,
Chen, X.: Improved techniques for training gans. In: Advances in Neural
Information Processing Systems 29. pp. 2234{2242 (2016)
13. Shorten, C., Khoshgoftaar, T.M.: A survey on image data augmentation for deep
learning. Journal of Big Data 6(1), 60 (Jul 2019), https://doi.org/10.1186/
s40537-019-0197-0
14. Ulicny, M., Krylov, V., Dahyot, R.: Harmonic networks for image classi cation. In:</p>
      <p>British Machine Vision Conference (BMVC). Cardi UK (9-12 September 2019)
15. Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P., et al.: Image quality
assessment: from error visibility to structural similarity. IEEE transactions on image
processing 13(4), 600{612 (2004)
16. Xiao, H., Rasul, K., Vollgraf, R.: Fashion-mnist: a novel image dataset for
benchmarking machine learning algorithms. CoRR abs/1708.07747 (2017), http:
//arxiv.org/abs/1708.07747
17. Ye, Y., Wang, L., Wu, Y., Chen, Y., Tian, Y., Liu, Z., Zhang, Z.: Gan quality
index (gqi) by gan-induced classi er (2018), https://openreview.net/forum?id=
S1CIev1vM
(d) Epoch 54
(e) Epoch 71
(h) Epoch 260</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Arjovsky</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bottou</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Towards principled methods for training generative adversarial networks</article-title>
          .
          <source>In: International Conference on Learning Representation</source>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Arjovsky</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chintala</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bottou</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Wasserstein generative adversarial networks</article-title>
          .
          <source>In: Proceedings of the 34th International Conference on Machine Learning. Proceedings of Machine Learning Research</source>
          , vol.
          <volume>70</volume>
          , pp.
          <volume>214</volume>
          {
          <issue>223</issue>
          (
          <issue>06</issue>
          {
          <issue>11</issue>
          <year>Aug 2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Goodfellow</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pouget-Abadie</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mirza</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Xu</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Warde-Farley</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ozair</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Courville</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bengio</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          :
          <article-title>Generative adversarial nets</article-title>
          .
          <source>In: Advances in Neural Information Processing Systems</source>
          <volume>27</volume>
          . pp.
          <volume>2672</volume>
          {
          <issue>2680</issue>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Gulrajani</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ahmed</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Arjovsky</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dumoulin</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Courville</surname>
            ,
            <given-names>A.C.</given-names>
          </string-name>
          :
          <article-title>Improved training of wasserstein gans</article-title>
          .
          <source>In: Advances in neural information processing systems</source>
          . pp.
          <volume>5767</volume>
          {
          <issue>5777</issue>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          , Zhang,
          <string-name>
            <given-names>S.</given-names>
            ,
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            ,
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <surname>R.</surname>
          </string-name>
          :
          <article-title>Beyond face rotation: Global and local perception gan for photorealistic and identity preserving frontal view synthesis</article-title>
          .
          <source>2017 IEEE International Conference on Computer Vision</source>
          (ICCV) pp.
          <volume>2458</volume>
          {
          <issue>2467</issue>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Krizhevsky</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Learning multiple layers of features from tiny images</article-title>
          .
          <source>Tech. rep., CIFAR-10 (Canadian Institute for Advanced Research)</source>
          (
          <year>2009</year>
          ), http://www.cs. toronto.edu/~kriz/cifar.html
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Ledig</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Theis</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Huszar</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Caballero</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Aitken</surname>
            ,
            <given-names>A.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tejani</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Totz</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shi</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          :
          <article-title>Photo-realistic single image super-resolution using a generative adversarial network</article-title>
          .
          <source>2017 IEEE Conference on Computer Vision</source>
          and Pattern
          <string-name>
            <surname>Recognition</surname>
          </string-name>
          (CVPR) pp.
          <volume>105</volume>
          {
          <issue>114</issue>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wand</surname>
            ,
            <given-names>M.:</given-names>
          </string-name>
          <article-title>Precomputed real-time texture synthesis with markovian generative adversarial networks</article-title>
          .
          <source>CoRR abs/1604</source>
          .04382 (
          <year>2016</year>
          ), http://arxiv.org/abs/ 1604.04382
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Mariani</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Scheidegger</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Istrate</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bekasa</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Malossi</surname>
            ,
            <given-names>A.C.I.</given-names>
          </string-name>
          :
          <article-title>Bagan: Data augmentation with balancing gan</article-title>
          .
          <source>In: ICML Workshop on Theoretical Foundations and Applications of Deep Generative Models</source>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>