<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Generating Artificial Data for Private Deep Learning</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>Aleksei Triastcyn and Boi Faltings Artificial Intelligence Laboratory Ecole Polytechnique Fe ́de ́rale de Lausanne Lausanne</institution>
          ,
          <country country="CH">Switzerland</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>In this paper, we propose generating artificial data that retain statistical properties of real data as the means of providing privacy for the original dataset. We use generative adversarial networks to draw privacy-preserving artificial data samples and derive an empirical method to assess the risk of information disclosure in a differential-privacy-like way. Our experiments show that we are able to generate labelled data of high quality and use it to successfully train and validate supervised models. Finally, we demonstrate that our approach significantly reduces vulnerability of such models to model inversion attacks.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Following recent advancements in deep learning, more and
more people and companies get interested in putting their
data in use and employ machine learning (ML) to generate
a wide range of benefits that span financial, social,
medical, security, and other aspects. At the same time, however,
such models are able to capture a fine level of detail in
training data, potentially compromising privacy of
individuals whose features sharply differ from others. Recent
research
        <xref ref-type="bibr" rid="ref12 ref17 ref30 ref31">(Fredrikson, Jha, and Ristenpart 2015)</xref>
        suggests that
even without access to internal model parameters it is
possible to recover (up to a certain degree) individual examples,
e.g. faces, from the training set.
      </p>
      <p>The latter result is especially disturbing knowing that deep
learning models are becoming an integral part of our lives,
making its way to phones, smart watches, cars, and
appliances. And since these models are often trained on
customers’ data, such training set recovery techniques endanger
privacy even without access to the manufacturer’s servers
where these models are being trained.</p>
      <p>
        One direction to tackle this problem is enforcing privacy
during training
        <xref ref-type="bibr" rid="ref1 ref25 ref25">(Abadi et al. 2016; Papernot et al. 2016;
2018)</xref>
        . We will refer to these techniques as model release
methods. While these approaches perform well in ML tasks
and provide strong privacy guarantees, they are often
restrictive. First and foremost, releasing a single trained model
does not provide much flexibility in the future. For
instance, it would significantly reduce possibilities for
combining models trained on data from different sources.
Evaluating a variety of such models and picking the best one is
also complicated by the need of adjusting private training for
      </p>
      <p>Labels
Noise</p>
      <p>Critic
Generator</p>
      <p>Real</p>
      <p>Fake
Artificial Data</p>
      <p>ML
each of them. Moreover, most of these methods assume
(implicitly or explicitly) access to public data of similar nature,
which may not be possible in areas like medicine.</p>
      <p>In contrast, we study the task of privacy-preserving data
release, which has many immediate advantages. First, any
ML model could be trained on released data without
additional assumptions. Second, data from different sources
could be easily pooled to build stronger models. Third,
released data could be traded on data markets1, where
anonymisation and protection of sensitive information is one
of the biggest obstacles. Finally, data publishing would
facilitate transparency and reproducibility of research studies.</p>
      <p>In particular, we are interested in solving two problems.
First, how to preserve high utility of data for ML algorithms
while protecting sensitive information in the dataset.
Second, how to quantify the risk of recovering private
information from the published dataset, and thus, the trained model.</p>
      <p>
        The main idea of our approach is to use generative
adversarial networks (GANs) (Goodfellow et al. 2014) to create
artificial datasets to be used in place of real data for
training. This method has a number of advantages over the
earlier work
        <xref ref-type="bibr" rid="ref1 ref25 ref25 ref32 ref7 ref8">(Abadi et al. 2016; Papernot et al. 2016; 2018;
Bindschaedler, Shokri, and Gunter 2017)</xref>
        . First of all, our
solution allows releasing entire datasets, thereby possessing all
the benefits of private data release as opposed to model
release. Second, it achieves high accuracy without pre-training
      </p>
      <sec id="sec-1-1">
        <title>1https://www.datamakespossible.com/</title>
        <p>value-of-data-2018/dawn-of-data-marketplace
on similar public data. Third, it is more intuitive and flexible,
e.g. it does not require a complex distributed architecture.</p>
        <p>To estimate potential privacy risks, we design an ex post
analysis framework for generated data. We use KL
divergence estimation and Chebyshev’s inequality to find
statistical bounds on expected privacy loss for a dataset in question.</p>
        <p>Our contributions in this paper are the following:
we propose a novel, yet simple, approach for private data
release, and to the best of our knowledge, this is the first
practical solution for complex real-world data;
we introduce a new framework for statistical estimation
of potential privacy loss of the released data;
we show that our method achieves learning performance
of model release methods and is resilient to model
inversion attacks.</p>
        <p>The rest of the paper is structured as follows. In Section 2,
we give an overview of related work. Section 3 contains
some preliminaries. In Section 4, we describe our approach
and privacy estimation framework, and discuss its
limitations. Experimental results and implementation details are
presented in Section 5, and Section 6 concludes the paper.
2</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <p>
        In recent years, as machine learning applications become a
commonplace, a body of work on security of these methods
grows at a rapid pace. Several important vulnerabilities and
corresponding attacks on ML models have been discovered,
raising the need of devising suitable defences. Among the
attacks that compromise privacy of training data, model
inversion
        <xref ref-type="bibr" rid="ref12 ref17 ref30 ref31">(Fredrikson, Jha, and Ristenpart 2015)</xref>
        and
membership inference
        <xref ref-type="bibr" rid="ref32 ref7">(Shokri et al. 2017)</xref>
        received high attention.
      </p>
      <p>
        Model inversion
        <xref ref-type="bibr" rid="ref12 ref17 ref30 ref31">(Fredrikson, Jha, and Ristenpart 2015)</xref>
        is based on observing the output probabilities of the target
model for a given class and performing gradient descent on
an input reconstruction. Membership inference
        <xref ref-type="bibr" rid="ref32 ref7">(Shokri et al.
2017)</xref>
        assumes an attacker with access to similar data, which
is used to train a ”shadow” model, mimicking the target, and
an attack model. The latter predicts if a certain example has
already been seen during training based on its output
probabilities. Note that both attacks can be performed in a
blackbox setting, without access to the model internal parameters.
      </p>
      <p>
        To protect privacy while still benefiting from the use of
statistics and ML, many techniques have been developed
over the years, including k-anonymity
        <xref ref-type="bibr" rid="ref34">(Sweeney 2002)</xref>
        ,
ldiversity
        <xref ref-type="bibr" rid="ref23">(Machanavajjhala et al. 2007)</xref>
        , t-closeness
        <xref ref-type="bibr" rid="ref21">(Li,
Li, and Venkatasubramanian 2007)</xref>
        , and differential privacy
(DP)
        <xref ref-type="bibr" rid="ref11">(Dwork 2006)</xref>
        . The latter has been recognised as a
rigorous standard and is widely accepted by the research
community. Its generic formulation, however, makes it hard to
achieve and to quantify potential privacy loss of the already
trained model. To overcome this, we build upon notions of
empirical DP
        <xref ref-type="bibr" rid="ref2">(Abowd, Schneider, and Vilhuber 2013)</xref>
        and
on-average KL privacy
        <xref ref-type="bibr" rid="ref10 ref36">(Wang, Lei, and Fienberg 2016)</xref>
        .
      </p>
      <p>
        Most of the ML-specific literature in the area concentrates
on the task of privacy-preserving model release. One take on
the problem is to distribute training and use disjoint datasets.
For example,
        <xref ref-type="bibr" rid="ref31">Shokri and Shmatikov (2015)</xref>
        propose to train
a model in a distributed manner by communicating
sanitised updates from participants to a central authority. Such
a method, however, yields high privacy losses
        <xref ref-type="bibr" rid="ref1 ref25 ref25">(Abadi et al.
2016; Papernot et al. 2016)</xref>
        . An alternative technique
suggested by
        <xref ref-type="bibr" rid="ref25">Papernot et al. (2016)</xref>
        , also uses disjoint training
sets and builds an ensemble of independently trained teacher
models to transfer knowledge to a student model by labelling
public data. This result has been extended in
        <xref ref-type="bibr" rid="ref26">(Papernot et al.
2018)</xref>
        to achieve state-of-the-art image classification results
in a private setting (with single-digit DP bounds). A
different approach is taken by
        <xref ref-type="bibr" rid="ref1">Abadi et al. (2016)</xref>
        . They suggest
using differentially private stochastic gradient descent
(DPSGD) to train deep learning models in a private manner. This
approach achieves high accuracy while maintaining low DP
bounds, but may also require pre-training on public data.
      </p>
      <p>
        A more recent line of research focuses on private data
release and providing privacy via generating synthetic
data
        <xref ref-type="bibr" rid="ref16 ref32 ref4 ref7 ref8">(Bindschaedler, Shokri, and Gunter 2017; Huang et al.
2017; Beaulieu-Jones et al. 2017)</xref>
        . In this scenario, DP is
hard to guarantee, and thus, such models either relax the
DP requirements or remain limited to simple data. In
        <xref ref-type="bibr" rid="ref32 ref7 ref8">(Bindschaedler, Shokri, and Gunter 2017)</xref>
        , authors use a
graphical probabilistic model to learn an underlying data
distribution and transform real data points (seeds) into synthetic
data points, which are then filtered by a privacy test based
on a plausible deniability criterion. This procedure would be
rather expensive for complex data, such as images.
        <xref ref-type="bibr" rid="ref16">Huang et
al. (2017)</xref>
        introduce the notion of generative adversarial
privacy and use GANs to obfuscate real data points w.r.t.
predefined private attributes, enabling privacy for more realistic
datasets. Finally, a natural approach to try is training GANs
using DP-SGD
        <xref ref-type="bibr" rid="ref37 ref39 ref4">(Beaulieu-Jones et al. 2017; Xie et al. 2018;
Zhang, Ji, and Wang 2018)</xref>
        . However, it proved extremely
difficult to stabilise training with the necessary amount of
noise, which scales as pm w.r.t. the number of model
parameters m. It makes these methods inapplicable to more
complex datasets without resorting to unrealistic (at least for
some areas) assumptions, like access to public data from the
same distribution.
      </p>
      <p>Similarly, our approach uses GANs, but data is generated
without real seeds or applying noise to gradients. Instead, we
verify experimentally that out-of-the-box GAN samples can
be sufficiently different from real data, and expected privacy
loss is empirically bounded by single-digit numbers.
3</p>
    </sec>
    <sec id="sec-3">
      <title>Preliminaries</title>
      <p>This section provides necessary definitions and background.
Let us commence with approximate differential privacy.
Definition 1. A randomised function (mechanism) M :
D ! R with domain D and range R satisfies (";
)differential privacy if for any two adjacent inputs d; d0 2 D
and for any outcome o 2 R the following holds:
Pr [M(d) = o]
e" Pr [M(d0) = o] + :
(1)
Definition 2. Privacy loss of a randomised mechanism M :
D ! R for inputs d; d0 2 D and outcome o 2 R takes the
following form:</p>
      <p>L(M(d)kM(d0)) = log</p>
      <p>
        For more details on differential privacy and the Gaussian
mechanism, we refer the reader to
        <xref ref-type="bibr" rid="ref9">(Dwork and Roth 2014)</xref>
        .
      </p>
      <p>In our privacy estimation framework, we also use some
classical notions from probability and information theory.
Definition 4. The Kullback–Leibler (KL) divergence
between two continuous probability distributions P and Q with
corresponding densities p, q is given by:</p>
      <p>DKL(P kQ) =</p>
      <p>Z +1
1
p(x) log
p(x)
q(x)
dx:</p>
      <p>Note that KL divergence between the distributions of
M(d) and M(d0) is nothing but the expectation of the
privacy loss random variable E[L(M(d)kM(d0))].</p>
      <p>
        Finally, Chebyshev’s inequality is used to obtain tail
bounds. In particular, as we expect the distribution to be
asymmetric, we use the version with semi-variances
        <xref ref-type="bibr" rid="ref6">(Berck
and Hihn 1982)</xref>
        to get a sharper bound:
      </p>
      <p>Pr(x</p>
      <p>E[x] + k )
k12 +22 ;
where
variance.</p>
      <p>+2 = RE+[x1] p(x)(x</p>
      <p>E[x])2dx is the upper
semi4</p>
    </sec>
    <sec id="sec-4">
      <title>Our Approach</title>
      <p>
        In this section, we describe our solution, its further
improvements, and provide details of the privacy estimation
framework. We then discuss limitations of the method. More
background on privacy can be found in
        <xref ref-type="bibr" rid="ref9">(Dwork and Roth 2014)</xref>
        .
      </p>
      <p>
        The main idea of our approach is to use artificial data for
learning and publishing instead of real (see Figure 1 for a
general workflow). The intuition behind it is the following.
Since it is possible to recover training examples from ML
models
        <xref ref-type="bibr" rid="ref12 ref17 ref30 ref31">(Fredrikson, Jha, and Ristenpart 2015)</xref>
        , we need to
limit the exposure of real data during training. While this can
be achieved by DP training (e.g. DP-SGD), it would have the
limitations mentioned earlier. Moreover, certain attacks can
still be successful if DP bounds are loose
        <xref ref-type="bibr" rid="ref15 ref8">(Hitaj, Ateniese,
and Pe´rez-Cruz 2017)</xref>
        . Removing real data from the
training process altogether would add another layer of protection
and limit the information leakage to artificial samples. What
remains to show is that artificial data is sufficiently different
from real.
4.1
      </p>
      <sec id="sec-4-1">
        <title>Differentially Private Critic</title>
        <p>
          Despite the fact that the generator does not have access
to real data in the training process, one cannot guarantee
(2)
(3)
(4)
(5)
that generated samples will not repeat the input. To
alleviate this problem, we propose to enforce differential
privacy on the output of the discriminator (critic). This is done
by employing the Gaussian noise mechanism
          <xref ref-type="bibr" rid="ref9">(Dwork and
Roth 2014)</xref>
          at the second-to-last layer: clipping the L2 norm
of the input and adding Gaussian noise. To be more
specific, activations a(x) of the second-to-last layer become
a~(x) = a(x)= max(ka(x)k2; 1) + N (0; 2). We refer to this
version of the critic as DP critic.
        </p>
        <p>
          Note that if the chosen GAN loss function was directly
differentiable w.r.t. generator output, i.e. if critic could be
treated as a black box, this modification would enforce the
same DP guarantees on generator parameters, and
consequently, all generated samples. Unfortunately, this is not the
case for practically all existing versions of GANs, including
WGAN-GP
          <xref ref-type="bibr" rid="ref14">(Gulrajani et al. 2017)</xref>
          used in our experiments.
        </p>
        <p>As our evaluation shows, this modification has a number
of advantages. First, it improves diversity of samples and
decreases similarity with real data. Second, it allows to prolong
training, and hence, obtain higher quality samples. Finally,
in our experiments, it significantly improves the ability of
GANs to generate samples conditionally.
4.2</p>
      </sec>
      <sec id="sec-4-2">
        <title>Privacy Estimation Framework</title>
        <p>
          Our framework builds upon ideas of empirical DP
(EDP)
          <xref ref-type="bibr" rid="ref17 ref2 ref30 ref31">(Abowd, Schneider, and Vilhuber 2013; Schneider
and Abowd 2015)</xref>
          and on-average KL privacy
          <xref ref-type="bibr" rid="ref10 ref36">(Wang, Lei,
and Fienberg 2016)</xref>
          . The first can be viewed as a measure of
sensitivity on posterior distributions of outcomes
          <xref ref-type="bibr" rid="ref8">(Charest
and Hou 2017)</xref>
          (in our case, generated data distributions),
while the second relaxes DP notion to the case of an average
user.
        </p>
        <p>As we don’t have access to exact posterior distributions, a
straightforward EDP procedure in our scenario would be the
following: (1) train GAN on the original dataset D; (2)
remove a random sample from D; (3) re-train GAN on the
updated set; (4) estimate probabilities of all outcomes and the
maximum privacy loss value; (5) repeat (1)–(4) sufficiently
many times to approximate ", .</p>
        <p>If the generative model is simple, this procedure can
be used without modification. Otherwise, for models like
GANs, it becomes prohibitively expensive due to repetitive
re-training (steps (1)–(3)). Another obstacle is estimating the
maximum privacy loss value (step (4)). To overcome these
two issues, we propose the following.</p>
        <p>First, to avoid re-training, we imitate the removal of
examples directly on the generated set De . We define a
similarity metric sim(x; y) between two data points x and y that
reflects important characteristics of data (see Section 5 for
details). For every randomly selected real example i, we
remove k nearest artificial neighbours to simulate absence of
this example in the training set and obtain De i. Our
intuition behind this operation is the following. Removing a real
example would result in a lower probability density in the
corresponding region of space. If this change is picked up
by a GAN, which we assume is properly trained (e.g. there
is no mode collapse), the density of this region in the
generated examples space should also decrease. The number of
neighbours k is a hyper-parameter. In our experiments, it is
chosen heuristically by computing KL divergence between
the real and artificial data distributions and assuming that all
the difference comes from one point.</p>
        <p>
          Second, we propose to relax the worst-case privacy loss
bound in step (4) by the expected-case bound, in the same
manner as on-average KL privacy. This relaxation allows us
to use a high-dimensional KL divergence estimator
          <xref ref-type="bibr" rid="ref27">(Pe´rezCruz 2008)</xref>
          to obtain the expected privacy loss for every pair
of adjacent datasets (De and De i). There are two major
advantages of this estimator: it converges almost surely to the
true value of KL divergence; and it does not require
intermediate density estimates to converge to the true
probability measures. Also since this estimator uses nearest
neighbours to approximate KL divergence, our heuristic described
above is naturally linked to the estimation method.
        </p>
        <p>
          Finally, after obtaining sufficiently many samples of
different pairs (De ; De i), we use Chebyshev’s inequality to
bound the probability = Pr(E[L(M(D)kM(D0))] ) of
the expected privacy loss
          <xref ref-type="bibr" rid="ref10">(Dwork and Rothblum 2016)</xref>
          exceeding a predefined threshold . To deal with the problem
of insufficiently many samples, one could use a sample
version of inequality
          <xref ref-type="bibr" rid="ref29">(Saw, Yang, and Mo 1984)</xref>
          at the cost of
looser bounds.
4.3
        </p>
      </sec>
      <sec id="sec-4-3">
        <title>Limitations</title>
        <p>Our empirical privacy estimator could be improved in a
number of ways. For instance, providing worst-case privacy
loss bounds would be largely beneficial. Furthermore,
simulating the removal of training examples currently depends
on heuristics and the chosen similarity metric, which may
not lead to representative samples and therefore, poor
guarantees.</p>
        <p>
          We provide bounds on expected privacy loss based on ex
post analysis of the artificial dataset, which is not
equivalent to the traditional formulation of DP and has certain
limitations
          <xref ref-type="bibr" rid="ref8">(Charest and Hou 2017)</xref>
          (e.g. it only concerns
a given dataset). Nevertheless, it may be useful in the
situations where strict privacy guarantees are not required or
cannot be achieved by existing methods, or when one wants
to get a better idea about expected privacy loss rather than
the highly unlikely worst-case.
        </p>
        <p>Lastly, all existing limitations of GANs (or generative
models in general), such as training instability or mode
collapse, will apply to this method. Hence, at the current state
of the field, our approach may be difficult to adapt to inputs
other than image data. Yet, there is still a number of
privacysensitive applications, e.g. medical imaging or facial
analysis, that could benefit from our technique. And as generative
methods progress, new uses will be possible.</p>
        <p>5</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Evaluation</title>
      <p>
        In this section, we describe the experimental setup and
implementation, and evaluate our method on MNIST
        <xref ref-type="bibr" rid="ref19">(LeCun
et al. 1998)</xref>
        , SVHN
        <xref ref-type="bibr" rid="ref24">(Netzer et al. 2011)</xref>
        , and CelebA
        <xref ref-type="bibr" rid="ref22">(Liu et
al. 2015)</xref>
        datasets.
We evaluate our method in two major ways. First, we show
that not only is it feasible to train ML models purely on
generated data, but it is also possible to achieve high learning
performance (Section 5.3). Second, we compute empirical
bounds on expected privacy loss and evaluate the
effectiveness of artificial data against model inversion attacks
(Section 5.4).
      </p>
      <p>Learning performance experiments are set up as follows:
1. Train a generative model (teacher) on the original dataset
using only the training split.
2. Generate an artificial dataset by the obtained model and
use it to train ML models (students).
3. Evaluate students on a held-out test set.</p>
      <p>Note that there is no dependency between teacher and
student models. Moreover, student models are not constrained
to neural networks and can be implemented as any type of
machine learning algorithm.</p>
      <p>We choose three commonly used image datasets for our
experiments: MNIST, SVHN, and CelebA. MNIST is a
handwritten digit recognition dataset consisting of 60000
training examples and 10000 test examples, each example
is a 28x28 size greyscale image. SVHN is also a digit
recognition task, with 73257 images for training and 26032 for
testing. The examples are coloured 32x32 pixel images of
house numbers from Google Street View. CelebA is a
facial attributes dataset with 202599 images, each of which
we crop to 128x128 and then downscale to 48x48.
5.2</p>
      <sec id="sec-5-1">
        <title>Implementation Details</title>
        <p>
          For our experiments, we use Python and Pytorch
framework.2 We implement, with some minor modifications, a
Wasserstein GAN with gradient penalty (WGAN-GP) by
          <xref ref-type="bibr" rid="ref14">Gulrajani et al. (2017)</xref>
          . More specifically, the critic consists
        </p>
        <sec id="sec-5-1-1">
          <title>2http://pytorch.org</title>
          <p>
            of four convolutional layers with SELU
            <xref ref-type="bibr" rid="ref18">(Klambauer et al.
2017)</xref>
            activations (instead of ReLU) followed by a fully
connected linear layer which outputs a d-dimensional feature
vector (d = 64). For the DP critic, we implement the
Gaussian noise mechanism
            <xref ref-type="bibr" rid="ref9">(Dwork and Roth 2014)</xref>
            by clipping
the L2-norm of this feature vector to C = 1 and adding
Gaussian noise with = 1:5 (we refer to it as DP layer).
Finally, it is passed through a linear classification layer. The
generator starts with a fully connected linear layer that
transforms noise and labels into a 4096-dimensional feature
vector which is then passed through a SELU activation and three
deconvolution layers with SELU activations. The output of
the third deconvolution layer is downsampled by max
pooling and normalised with a tanh activation function.
          </p>
          <p>Similarly to the original paper, we use a classical WGAN
value function with the gradient penalty that enforces
Lipschitz constraint on a critic. We also set the penalty parameter
= 10 and the number of critic iterations ncritic = 5.
Furthermore, we modify the architecture to allow for
conditioning WGAN on class labels. Binarised labels are appended to
the input of the generator and to the linear layer of the critic
after convolutions. Therefore, the generator can be used to
create labelled datasets for supervised learning.</p>
          <p>
            Both networks are trained using Adam
            <xref ref-type="bibr" rid="ref17 ref30 ref31">(Kingma and Ba
2015)</xref>
            with learning rate 10 4, 1 = 0, 2 = 0:9, and a
batch size of 64.
          </p>
          <p>The student network is constructed of two
convolutional layers with ReLU activations, batch normalisation and
max pooling, followed by two fully connected layers with
ReLU, and a softmax output layer. Note that this network
does not achieve state-of-the-art performance on the used
datasets, but we are primarily interested in evaluating the
relative performance drop compared to a non-private model.</p>
          <p>
            To estimate privacy loss, we carry out the procedure
presented in Section 4. Specifically, based on recent ideas
in image qualitative evaluation, e.g. FID and Inception
Score, we compute image features by the Inception V3
network
            <xref ref-type="bibr" rid="ref35">(Szegedy et al. 2016)</xref>
            and use inverse distances
between features as sim function. We implement the KL
divergence estimator
            <xref ref-type="bibr" rid="ref27">(Pe´rez-Cruz 2008)</xref>
            and use k-d trees
            <xref ref-type="bibr" rid="ref5">(Bentley 1975)</xref>
            for fast nearest neighbour searches. For privacy
evaluation, we implement the model inversion attack.
5.3
          </p>
        </sec>
      </sec>
      <sec id="sec-5-2">
        <title>Learning Performance</title>
        <p>First, we evaluate the generalisation ability of a student
model trained on artificial data. More specifically, we train
a student model on generated data and report test
classification accuracy on a held-out real set.</p>
        <p>
          As noted above, most of the work on privacy-preserving
ML focuses on model release methods and assumes
(explicitly or implicitly) access to similar ”public” data in one form
or another
          <xref ref-type="bibr" rid="ref1 ref25 ref25 ref39">(Abadi et al. 2016; Papernot et al. 2016; 2018;
Zhang, Ji, and Wang 2018)</xref>
          . On the other hand, existing data
release solutions struggle with high-dimensional data
          <xref ref-type="bibr" rid="ref40">(Zhu
et al. 2017)</xref>
          . It limits the choice of methods for comparison.
        </p>
        <p>
          We chose to compare learning performance with the
current state-of-the-art model release technique, PATE by
          <xref ref-type="bibr" rid="ref26">Papernot et al. (2018)</xref>
          , which uses a relatively small set of
unlabelled ”public” data. Since our approach does not require
any ”public” data, in order to make the evaluation more
appropriate, we pick the results of PATE corresponding to the
least number of labelling queries.
        </p>
        <p>Table 1 shows test accuracy for the non-private
baseline model (trained on the real training set), PATE, and our
method. We observe that artificial data allows us to achieve
98:3% accuracy on MNIST and 87:7% accuracy on SVHN,
which is comparable or better than corresponding results of
PATE. These results demonstrate that our approach does not
compromise learning performance, and may even improve
it, while enabling the full flexibility of data release methods.</p>
        <p>Additionally, we train a simple logistic regression model
on artificial MNIST samples, and obtain 91:69% accuracy,
(a) Generated
(b) Real
(a) Generated
(b) Real
compared to 92:58% on the original data, confirming that
student models are not restricted to a specific type.</p>
        <p>Furthermore, we observe that one could use artificial data
for validation and hyper-parameter tuning. In our
experiments, correlation coefficients between real and artificial
validation losses range from 0.7197 to 0.9972 for MNIST
and from 0.8047 to 0.9810 for SVHN.
5.4</p>
      </sec>
      <sec id="sec-5-3">
        <title>Privacy Analysis</title>
        <p>
          Using the privacy estimation framework (see Section 4), we
fix the probability of exceeding the expected privacy loss
bound in all experiments to 10 5 and compute the
corresponding for each dataset and two versions of WGAN-GP
(vanilla and with DP critic). Table 2 summarises our
findings. It is worth noting, that our should not be viewed as
an empirical estimation of " of DP, since the former bounds
expected privacy loss while the latter–maximum. These two
quantities, however, in our experiments turn out to be similar
to deep learning DP bounds found in recent literature
          <xref ref-type="bibr" rid="ref1 ref25 ref26">(Abadi
et al. 2016; Papernot et al. 2018)</xref>
          . This may be explained by
tight concentration of privacy loss random variable
          <xref ref-type="bibr" rid="ref10">(Dwork
and Rothblum 2016)</xref>
          or loose estimation. Additionally, DP
critic helps to bring down values in all cases.
        </p>
        <p>
          The lack of theoretical privacy guarantees for our method
neccesitates assessing the strength of provided protection.
We perform this evaluation by running the model inversion
attack
          <xref ref-type="bibr" rid="ref12 ref17 ref30 ref31">(Fredrikson, Jha, and Ristenpart 2015)</xref>
          on a student
model. Note that we also experimented with another
wellknown attack on machine learning models, the membership
inference
          <xref ref-type="bibr" rid="ref32 ref7">(Shokri et al. 2017)</xref>
          . However, we did not include
it in the final evaluation, because of the poor attacker’s
performance in our setting (nearly random guess accuracy for
given datasets and models even without any protection).
        </p>
        <p>In order to run the attack, we train a student model (a
simple multi-layer perceptron with two hidden layers of 1000
and 300 neurons) in three settings: real data, artificial data
generated by GAN (with DP critic), and real data with
differential privacy (using DP-SGD with a small " &lt; 1). As facial
recognition is a more privacy-sensitive application, and
provides a better visualisation of the attack, we picked CelebA
attribute prediction task to run this experiment.</p>
        <p>Figure 2 shows the results of the model inversion attack.
The top row presents the real target images. The following
rows depict reconstructed images from a non-private model,
a model trained on GAN samples, and DP model,
correspondingly. One can observe a clear information loss in
reconstructed images going from non-private model, to
artificial data, to DP. The latter is superior in decoupling the
model and the training data, and is a preferred choice in
the model release setting and/or if public data is accessible
for pre-training. The non-private model, albeit trained with
abundant data ( 200K images) reveals facial features, such
as skin and hair colour, expression, etc. Our method,
despite failing to conceal general shapes in training images (i.e.
faces), seems to achieve a trade-off, hiding most of the
specific features. The obtained reconstructions are either very
noisy (columns 1, 2, 6, 8), much like DP, or converge to
some average feature-less faces (columns 4, 5, 7).</p>
        <p>
          We also analyse real and reconstructed image pairs using
OpenFace
          <xref ref-type="bibr" rid="ref3">(Amos et al. 2016)</xref>
          (see Table 3). It confirms our
initial findings: in images reconstructed from a non-private
model, faces were detected (recognised) 63:6% (11%) of
the time, while for our method, detection succeeded only
in 1:3% of cases and recognition rate was 0:3%, well within
state-of-the-art error margins. For DP both rates were at 0%.
        </p>
        <p>To evaluate our privacy estimation method, we look at
how the privacy loss bound correlates with the success
of the attack. Figure 3 depicts the privacy-accuracy trade-off
curve for an MLP (64-32-10) trained on artificial data. In this
setting, we use a stacked denoising autoencoder to compress
images to 64-dimensional feature vectors and facilitate the
attack performance. Along the curve, we plot examples of
the model inversion reconstruction at corresponding points.
We see that with growing , meaning lower privacy, both
model accuracy and reconstruction quality increase.</p>
        <p>Finally, as an additional measure, we perform visual
inspection of generated examples and corresponding nearest
neighbours in real data. Figures 4 and 5 depict generated
and the corresponding most similar real images from SVHN
and CelebA datasets. We observe that, despite general
visual similarity, generated images differ from real examples
in details, which is normally more important for privacy. For
SVHN, digits vary either in shape, colour or surroundings.
A lot of pairs come from different classes. For CelebA, the
pose and lighting may be similar, but such details as gender,
skin colour, facial features are usually significantly different.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Conclusions</title>
      <p>We investigate the problem of private data release for
complex high-dimensional data. In contrast to commonly
studied model release setting, this approach enables important
advantages and applications, such as data pooling from
multiple sources, simpler development process, and data trading.</p>
      <p>We employ generative adversarial networks to produce
artificial privacy-preserving datasets. The choice of GANs as
a generative model ensures scalability and makes the
technique suitable for real-world data with complex structure.
Unlike many prior approaches, our method does not
assume access to similar publicly available data. In our
experiments, we show that student models trained on artificial data
can achieve high accuracy on MNIST and SVHN datasets.
Moreover, models can also be validated on artificial data.</p>
      <p>We propose a novel technique for estimating privacy of
released data by empirical bounds on expected privacy loss.
We compute privacy bounds for samples from WGAN-GP
on MNIST, SVHN, and CelebA, and demonstrate that
expected privacy loss is bounded by single-digit values. To
evaluate provided protection, we run a model inversion
attack and show that training with GAN reduces information
leakage (e.g. face detection drops from 63:6% to 1:3%) and
that attack success correlates with estimated privacy bounds.</p>
      <p>Additionally, we introduce a simple modification to the
critic: differential privacy layer. Not only does it improve
privacy loss bounds and ensures DP guarantees for the critic
output, but it also acts as a regulariser, improving stability of
training, and quality and diversity of generated images.</p>
      <p>Considering the rising importance of privacy research and
the lack of good solutions for private data publishing, there is
a lot of potential future work. In particular, a major direction
of advancing current work would be achieving differential
privacy guarantees for generative models while still
preserving high utility of generated data. A step in another direction
would be to improve the privacy estimation framework, e.g.
by bounding maximum privacy loss, or finding a more
principled way of sampling from outcome distributions.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <surname>Abadi</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Chu</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Goodfellow</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>McMahan</surname>
            ,
            <given-names>H. B.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Mironov</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ; Talwar,
          <string-name>
            <surname>K.</surname>
          </string-name>
          ; and Zhang,
          <string-name>
            <surname>L.</surname>
          </string-name>
          <year>2016</year>
          .
          <article-title>Deep learning with differential privacy</article-title>
          .
          <source>In Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security</source>
          ,
          <fpage>308</fpage>
          -
          <lpage>318</lpage>
          . ACM.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <surname>Abowd</surname>
            ,
            <given-names>J. M.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Schneider</surname>
            ,
            <given-names>M. J.;</given-names>
          </string-name>
          and
          <string-name>
            <surname>Vilhuber</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <year>2013</year>
          .
          <article-title>Differential privacy applications to bayesian and linear mixed model estimation</article-title>
          .
          <source>Journal of Privacy and Confidentiality</source>
          <volume>5</volume>
          (
          <issue>1</issue>
          ):
          <fpage>4</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <surname>Amos</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Ludwiczuk</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Satyanarayanan</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ; et al.
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <surname>Beaulieu-Jones</surname>
            ,
            <given-names>B. K.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>Z. S.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Williams</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Greene</surname>
            ,
            <given-names>C. S.</given-names>
          </string-name>
          <year>2017</year>
          .
          <article-title>Privacy-preserving generative deep neural networks support clinical data sharing</article-title>
          .
          <source>bioRxiv 159756.</source>
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <surname>Bentley</surname>
            ,
            <given-names>J. L.</given-names>
          </string-name>
          <year>1975</year>
          .
          <article-title>Multidimensional binary search trees used for associative searching</article-title>
          .
          <source>Communications of the ACM</source>
          <volume>18</volume>
          (
          <issue>9</issue>
          ):
          <fpage>509</fpage>
          -
          <lpage>517</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <surname>Berck</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Hihn</surname>
            ,
            <given-names>J. M.</given-names>
          </string-name>
          <year>1982</year>
          .
          <article-title>Using the semivariance to estimate safety-first rules</article-title>
          .
          <source>American Journal of Agricultural Economics</source>
          <volume>64</volume>
          (
          <issue>2</issue>
          ):
          <fpage>298</fpage>
          -
          <lpage>300</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <surname>Bindschaedler</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Shokri</surname>
          </string-name>
          , R.; and
          <string-name>
            <surname>Gunter</surname>
            ,
            <given-names>C. A.</given-names>
          </string-name>
          <year>2017</year>
          .
          <article-title>Plausible deniability for privacy-preserving data synthesis</article-title>
          .
          <source>Proceedings of the VLDB Endowment</source>
          <volume>10</volume>
          (
          <issue>5</issue>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <surname>Charest</surname>
            ,
            <given-names>A.-S.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Hou</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <year>2017</year>
          .
          <article-title>On the meaning and limits of empirical differential privacy</article-title>
          .
          <source>Journal of Privacy and Confidentiality</source>
          <volume>7</volume>
          (
          <issue>3</issue>
          ):
          <fpage>3</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <surname>Dwork</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Roth</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <year>2014</year>
          .
          <article-title>The algorithmic foundations of differential privacy</article-title>
          .
          <source>Foundations and Trends R in Theoretical Computer Science</source>
          <volume>9</volume>
          (
          <issue>3</issue>
          -4):
          <fpage>211</fpage>
          -
          <lpage>407</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <surname>Dwork</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Rothblum</surname>
            ,
            <given-names>G. N.</given-names>
          </string-name>
          <year>2016</year>
          .
          <article-title>Concentrated differential privacy</article-title>
          .
          <source>arXiv preprint arXiv:1603</source>
          .
          <year>01887</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <surname>Dwork</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <year>2006</year>
          .
          <article-title>Differential privacy</article-title>
          .
          <source>In 33rd International Colloquium on Automata, Languages and Programming</source>
          ,
          <string-name>
            <surname>part II</surname>
          </string-name>
          (ICALP
          <year>2006</year>
          ), volume
          <volume>4052</volume>
          ,
          <fpage>1</fpage>
          -
          <lpage>12</lpage>
          . Venice, Italy: Springer Verlag.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <surname>Fredrikson</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Jha</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ; and Ristenpart,
          <string-name>
            <surname>T.</surname>
          </string-name>
          <year>2015</year>
          .
          <article-title>Model inversion attacks that exploit confidence information and basic countermeasures</article-title>
          .
          <source>In Proceedings of the 22nd ACM SIGSAC Conference on Computer and Communications Security</source>
          ,
          <fpage>1322</fpage>
          -
          <lpage>1333</lpage>
          . ACM.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          2014.
          <article-title>Generative adversarial nets</article-title>
          .
          <source>In Advances in Neural Information Processing Systems</source>
          ,
          <volume>2672</volume>
          -
          <fpage>2680</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <string-name>
            <surname>Gulrajani</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Ahmed</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Arjovsky</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Dumoulin</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Courville</surname>
            ,
            <given-names>A. C.</given-names>
          </string-name>
          <year>2017</year>
          .
          <article-title>Improved training of wasserstein gans</article-title>
          .
          <source>In Advances in Neural Information Processing Systems</source>
          ,
          <volume>5769</volume>
          -
          <fpage>5779</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <string-name>
            <surname>Hitaj</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Ateniese</surname>
          </string-name>
          , G.;
          <article-title>and Pe´rez-</article-title>
          <string-name>
            <surname>Cruz</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <year>2017</year>
          .
          <article-title>Deep models under the gan: information leakage from collaborative deep learning</article-title>
          .
          <source>In Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security</source>
          ,
          <fpage>603</fpage>
          -
          <lpage>618</lpage>
          . ACM.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Kairouz</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Sankar</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ; and Rajagopal,
          <string-name>
            <surname>R.</surname>
          </string-name>
          <year>2017</year>
          .
          <article-title>Context-aware generative adversarial privacy</article-title>
          .
          <source>Entropy</source>
          <volume>19</volume>
          (
          <issue>12</issue>
          ):
          <fpage>656</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <string-name>
            <surname>Kingma</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Ba</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <year>2015</year>
          .
          <article-title>Adam: A method for stochastic optimization</article-title>
          .
          <source>In Proceedings of the 3rd International Conference for Learning Representations.</source>
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          <string-name>
            <surname>Klambauer</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ; Unterthiner,
          <string-name>
            <given-names>T.</given-names>
            ;
            <surname>Mayr</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            ; and
            <surname>Hochreiter</surname>
          </string-name>
          ,
          <string-name>
            <surname>S.</surname>
          </string-name>
          <year>2017</year>
          .
          <article-title>Self-normalizing neural networks</article-title>
          .
          <source>In Advances in Neural Information Processing Systems</source>
          ,
          <volume>972</volume>
          -
          <fpage>981</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          <string-name>
            <surname>LeCun</surname>
          </string-name>
          , Y.;
          <string-name>
            <surname>Bottou</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Bengio</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ; and Haffner,
          <string-name>
            <surname>P.</surname>
          </string-name>
          <year>1998</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          <source>Proceedings of the IEEE</source>
          <volume>86</volume>
          (
          <issue>11</issue>
          ):
          <fpage>2278</fpage>
          -
          <lpage>2324</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Venkatasubramanian</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <year>2007</year>
          .
          <article-title>tcloseness: Privacy beyond k-anonymity and l-diversity</article-title>
          .
          <source>In Data Engineering</source>
          ,
          <year>2007</year>
          .
          <article-title>ICDE 2007</article-title>
          . IEEE 23rd International Conference on,
          <fpage>106</fpage>
          -
          <lpage>115</lpage>
          . IEEE.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Luo</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Tang</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          <year>2015</year>
          .
          <article-title>Deep learning face attributes in the wild</article-title>
          .
          <source>In Proceedings of International Conference on Computer Vision</source>
          (ICCV).
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          <string-name>
            <surname>Machanavajjhala</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Kifer</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Gehrke</surname>
            , J.; and Venkitasubramaniam,
            <given-names>M.</given-names>
          </string-name>
          <year>2007</year>
          .
          <article-title>l-diversity: Privacy beyond kanonymity. ACM Transactions on Knowledge Discovery from Data (TKDD) 1(1):3</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          <string-name>
            <surname>Netzer</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Coates</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Bissacco</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Ng</surname>
            ,
            <given-names>A. Y.</given-names>
          </string-name>
          <year>2011</year>
          .
          <article-title>Reading digits in natural images with unsupervised feature learning</article-title>
          .
          <source>In NIPS workshop on deep learning and unsupervised feature learning</source>
          , volume
          <year>2011</year>
          ,
          <volume>5</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          <string-name>
            <surname>Papernot</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Abadi</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Erlingsson</surname>
            ,
            <given-names>U</given-names>
          </string-name>
          ´ .;
          <string-name>
            <surname>Goodfellow</surname>
            ,
            <given-names>I.;</given-names>
          </string-name>
          and
          <string-name>
            <surname>Talwar</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <year>2016</year>
          .
          <article-title>Semi-supervised knowledge transfer for deep learning from private training data</article-title>
          .
          <source>arXiv preprint arXiv:1610</source>
          .
          <fpage>05755</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          <string-name>
            <surname>Papernot</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Song</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Mironov</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Raghunathan</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Talwar</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ; and Erlingsson,
          <string-name>
            <surname>U</surname>
          </string-name>
          ´ .
          <year>2018</year>
          .
          <article-title>Scalable private learning with pate</article-title>
          .
          <source>arXiv preprint arXiv:1802</source>
          .08908.
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          <article-title>Pe´rez-</article-title>
          <string-name>
            <surname>Cruz</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <year>2008</year>
          .
          <article-title>Kullback-leibler divergence estimation of continuous distributions</article-title>
          .
          <source>In Information Theory</source>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          <string-name>
            <surname>ISIT</surname>
          </string-name>
          <year>2008</year>
          . IEEE International Symposium on,
          <fpage>1666</fpage>
          -
          <lpage>1670</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          <string-name>
            <surname>Saw</surname>
            ,
            <given-names>J. G.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>M. C.</given-names>
          </string-name>
          ; and Mo,
          <string-name>
            <surname>T. C.</surname>
          </string-name>
          <year>1984</year>
          .
          <article-title>Chebyshev inequality with estimated mean and variance</article-title>
          .
          <source>The American Statistician</source>
          <volume>38</volume>
          (
          <issue>2</issue>
          ):
          <fpage>130</fpage>
          -
          <lpage>132</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          <string-name>
            <surname>Schneider</surname>
            ,
            <given-names>M. J.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Abowd</surname>
            ,
            <given-names>J. M.</given-names>
          </string-name>
          <year>2015</year>
          .
          <article-title>A new method for protecting interrelated time series with bayesian prior distributions and synthetic data</article-title>
          .
          <source>Journal of the Royal Statistical Society: Series A (Statistics in Society)</source>
          <volume>178</volume>
          (
          <issue>4</issue>
          ):
          <fpage>963</fpage>
          -
          <lpage>975</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          <string-name>
            <surname>Shokri</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Shmatikov</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <year>2015</year>
          .
          <article-title>Privacy-preserving deep learning</article-title>
          .
          <source>In Proceedings of the 22nd ACM SIGSAC conference on computer and communications security</source>
          ,
          <fpage>1310</fpage>
          -
          <lpage>1321</lpage>
          . ACM.
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          <string-name>
            <surname>Shokri</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ; Stronati,
          <string-name>
            <given-names>M.</given-names>
            ;
            <surname>Song</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            ; and
            <surname>Shmatikov</surname>
          </string-name>
          ,
          <string-name>
            <surname>V.</surname>
          </string-name>
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          <article-title>Membership inference attacks against machine learning models</article-title>
          .
          <source>In Security and Privacy (SP)</source>
          ,
          <source>2017 IEEE Symposium on</source>
          ,
          <fpage>3</fpage>
          -
          <lpage>18</lpage>
          . IEEE.
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          <string-name>
            <surname>Sweeney</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <year>2002</year>
          .
          <article-title>K-anonymity: A model for protecting privacy</article-title>
          .
          <source>Int. J. Uncertain. Fuzziness Knowl.-Based Syst.</source>
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          <string-name>
            <surname>Szegedy</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Vanhoucke</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Ioffe</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Shlens</surname>
          </string-name>
          , J.; and
          <string-name>
            <surname>Wojna</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <year>2016</year>
          .
          <article-title>Rethinking the inception architecture for computer vision</article-title>
          .
          <source>In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition</source>
          ,
          <fpage>2818</fpage>
          -
          <lpage>2826</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>Y.-X.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Lei</surname>
          </string-name>
          , J.; and
          <string-name>
            <surname>Fienberg</surname>
            ,
            <given-names>S. E.</given-names>
          </string-name>
          <year>2016</year>
          .
          <article-title>On-average kl-privacy and its equivalence to generalization for maxentropy mechanisms</article-title>
          .
          <source>In International Conference on Privacy in Statistical Databases</source>
          ,
          <fpage>121</fpage>
          -
          <lpage>134</lpage>
          . Springer.
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          <string-name>
            <surname>Xie</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Lin</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Zhou</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          <article-title>Differentially private generative adversarial network</article-title>
          . arXiv preprint arXiv:
          <year>1802</year>
          .06739.
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Ji</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <year>2018</year>
          .
          <article-title>Differentially private releasing via deep generative model</article-title>
          . arXiv preprint arXiv:
          <year>1801</year>
          .01594.
        </mixed-citation>
      </ref>
      <ref id="ref40">
        <mixed-citation>
          <string-name>
            <surname>Zhu</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Zhou</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Philip</surname>
            ,
            <given-names>S. Y.</given-names>
          </string-name>
          <year>2017</year>
          .
          <article-title>Differentially private data publishing and analysis: a survey.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref41">
        <mixed-citation>
          <source>IEEE Transactions on Knowledge and Data Engineering</source>
          <volume>29</volume>
          (
          <issue>8</issue>
          ):
          <fpage>1619</fpage>
          -
          <lpage>1638</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>