<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>San Diego,
California, USA, August</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Knowledge Intensive Learning of Generative Adversarial Networks</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Devendra Singh Dhami</string-name>
          <email>devendra.dhami@utdallas.edu</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mayukh Das</string-name>
          <email>mayukh.das@samsung.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Sriraam Natarajan</string-name>
          <email>sriraam.natarajan@utdallas.edu</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Samsung Research</institution>
          <country country="IN">India</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>The University of Texas at Dallas</institution>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2020</year>
      </pub-date>
      <volume>24</volume>
      <issue>2020</issue>
      <abstract>
        <p>While Generative Adversarial Networks (GANs) have accelerated the use of generative modelling within the machine learning community, most of the applications of GANs are restricted to images. The use of GANs to generate clinical data has been rare due to the inability of GANs to faithfully capture the intrinsic relationships between features. We hypothesize and verify that this challenge can be mitigated by incorporating domain knowledge in the generative process. Specifically, we propose human-allied GANs that using correlation advice from humans to create synthetic clinical data. Our empirical evaluation demonstrates the superiority of our approach over other GAN models.</p>
      </abstract>
      <kwd-group>
        <kwd>generative adversarial networks</kwd>
        <kwd>human in the loop</kwd>
        <kwd>healthcare</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>CCS CONCEPTS</title>
    </sec>
    <sec id="sec-2">
      <title>INTRODUCTION</title>
      <p>
        Deep learning models have reshaped the machine learning landscape
over the past decade [
        <xref ref-type="bibr" rid="ref16 ref29">16, 29</xref>
        ]. Specifically, Generative
Adversarial Networks (GANs) [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] have found tremendous success in
generating examples for images [
        <xref ref-type="bibr" rid="ref34 ref37 ref45">34, 37, 45</xref>
        ], photographs of human
faces [
        <xref ref-type="bibr" rid="ref1 ref25 ref52">1, 25, 52</xref>
        ], image to image translation [
        <xref ref-type="bibr" rid="ref30 ref33 ref55">30, 33, 55</xref>
        ] and 3D
object generation [
        <xref ref-type="bibr" rid="ref44 ref51 ref53">44, 51, 53</xref>
        ] to name a few. Despite such success,
there are several key factors that limit the widespread adoption of
GANs, for a broader range of tasks, including, widely acknowledged
data hungry nature of such methods, potential access issues of real
medical data and finally, their restricted usage, mainly in the
context of images. These factors have limited the use of these arguably
successful techniques in medical (or similar) domains. However,
recently, synthetic data generation has become a centerpiece of
research in medical AI due to the diverse difcfiulties in collection,
persistence, sharing and analysis of real clinical data.
      </p>
      <p>
        We aim to address the above limitations. Inspired by Mitchell’s
argument of “The Need for Biases in Learning Generalizations” [
        <xref ref-type="bibr" rid="ref38">38</xref>
        ],
we mitigate the challenges of existing data hungry methods via
inductive bias while learning GANs. We show that effective inductive
bias can be provided by humans in the form of domain
knowledge [
        <xref ref-type="bibr" rid="ref14 ref27 ref41 ref50">14, 27, 41, 50</xref>
        ]. Rich human advice can effectively balance
the impact of quality (sparsity) of training data. Data quality also
contributes to, the well studied, modal instability of GANs. This
problem is especially critical in domains such as medical/clinical
analytics that does not typically exhibit ‘spatial homophily’ [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ],
unlike images, and are prone to distributional diversity among feature
clusters as well. Our human-guided framework proposes a robust
strategy to address this challenge. Note that in our setting the human
is an ally and not an adversary.
      </p>
      <p>
        The second limitation of access is crucial for medical data
generation. Access to existing medical databases [
        <xref ref-type="bibr" rid="ref10 ref18">10, 18</xref>
        ] is hard due to
cost and access concerns and thus synthetic data generation holds
tremendous promise [
        <xref ref-type="bibr" rid="ref13 ref19 ref35 ref48 ref6">6, 13, 19, 35, 48</xref>
        ]. While previous methods
generated synthetic images, we go beyond images and generate
clinical data. Building on this body of work, we present a synthetic data
generation framework that effectively exploits domain expertise to
handle data quality.
      </p>
      <p>We make a few key contributions:
(1) We demonstrate how effective human advice can be provided
to a GAN as an inductive bias.
(2) We present a method for generating data given this advice.
(3) Finally, we demonstrate the effectiveness and efficacy of our
approach on 2 de-identified clinical data sets. Our method
is generalizable to multiple modalities of data and is not
necessarily restricted to images.
(4) Yet another feature of this approach is that training occurs
from very few data samples (&lt; 50 in one domain) thus
providing human guidance as a data generation alternative.
2</p>
    </sec>
    <sec id="sec-3">
      <title>RELATED WORK</title>
      <p>
        The key principle behind GANs [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] is a zero-sum game [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ] from
game theory, a mathematical representation where each participant’s
gain or loss is exactly balanced by the losses or gains of the other
participants and is generally solved by a minimax algorithm. The
generator distribution  () over the given data  is learned by
sampling  from a random distribution  () (initially uniform was
proposed but Gaussians have been proven superior [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]). While GANs
have proven to be a powerful framework for estimating generative
distributions, convergence dynamics of naive mini-max algorithm
has been shown to be unstable. Some recent approaches, among
many others, augment learning either via statistical relationships
between true and learned generative distributions such as Wasserstein-1
distance [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], MMD [
        <xref ref-type="bibr" rid="ref32">32</xref>
        ] or via spectral normalization of the
parameter space of the generator [
        <xref ref-type="bibr" rid="ref39">39</xref>
        ] which controls the generator
distribution from drifting too far. Although these approaches have improved
the GAN learning in some cases, there is room for improvement.
      </p>
      <p>
        Guidance via human knowledge is a provably effective way to
control learning in presence of systematic noise (which leads to
instability). One typical strategy to incorporate such guidance is
by providing rules over training examples and features. Some of
the earliest approaches are explanation-based learning (EBL-NN,
[
        <xref ref-type="bibr" rid="ref49">49</xref>
        ]) or ANNs augmented with symbolic rules (KBANN, [
        <xref ref-type="bibr" rid="ref50">50</xref>
        ]).
Various widely-studied techniques of leveraging domain knowledge
for optimal model generalization include polyhedral constraints in
case of knowledge-based SVMs, [
        <xref ref-type="bibr" rid="ref14 ref28 ref47 ref9">9, 14, 28, 47</xref>
        ]), preferences rules
[
        <xref ref-type="bibr" rid="ref27 ref41 ref42 ref5">5, 27, 41, 42</xref>
        ] or qualitative constraints (ex: monotonicities /
synergies [
        <xref ref-type="bibr" rid="ref54">54</xref>
        ] or quantitative relationships [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]). Notably, whereas these
models exhibit considerable improvement with the incorporation of
human knowledge, there is only limited use of such knowledge in
training GANs. Our approach resembles the qualitative constraints
framework in spirit.
      </p>
      <p>
        While widely successful in building optimally generalized models
in presence of systematic noise (or sample biases), knowledge-based
approaches have mostly been explored in the context of
discriminative modeling. In the generative setting, a recent work extends
the principle of posterior regularization from Bayesian modeling to
deep generative models in order to incorporate structured domain
knowledge [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ]. Traditionally, knowledge based generative learning
has been studied as a part of learning probabilistic graphical models
with structure/parameter priors [
        <xref ref-type="bibr" rid="ref36">36</xref>
        ]. We aim to extend the use of
knowledge to the generative model setting.
3
      </p>
    </sec>
    <sec id="sec-4">
      <title>KNOWLEDGE INTENSIVE LEARNING OF</title>
    </sec>
    <sec id="sec-5">
      <title>GENERATIVE ADVERSARIAL NETWORKS</title>
      <p>
        A notable disadvantage of adversarial training formulation is that
the training is slow and unstable, leading to mode collapse [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] where
the generator starts generating data of only a single modality. This
has resulted in GANs not being exploited to their full potential in
generating synthetic non-image clinical data. Human advice can
encourage exploration in diverse areas of the feature space and helps
learn more stable models [
        <xref ref-type="bibr" rid="ref43">43</xref>
        ]. Hence, we propose a human-allied
GAN architecture (HA-GAN) (figure 1). The architecture
incorporates human advice in form of feature correlations. Such intrinsic
relationships between the features are crucial in medical data sets
and thus become a natural candidate as additional knowledge/advice
in guided model learning for faithful data generation.
      </p>
      <p>
        Our approach builds upon a GAN architecture [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] where a
random noise vector is provided to the generator which tries to generate
examples as close to the real distribution as possible. The
discriminator tries to distinguish between real examples and ones generated
by the generator. The generator tries to maximize the probability
that the discriminator makes a mistake and the discriminator tries to
minimize its mistakes thereby resulting in a min-max optimization
problem which can be solved by a mini-max algorithm. We adopt
the Wasserstein GAN (WGAN) architecture1 [
        <xref ref-type="bibr" rid="ref20 ref3">3, 20</xref>
        ] that focuses
      </p>
      <sec id="sec-5-1">
        <title>1We use ‘GAN’ to indicate ‘W-GAN’</title>
        <p>on defining a distance/divergence (Wasserstein or earth movers
distance) to measure the closeness between the real distribution and the
model distribution.
3.1</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Human input as inductive bias</title>
      <p>Historically, two approaches have been studied for using guidance
as bias. The first is to provide advice on the labels as constraints
or preferences that controls the search space. Some example advice
rules on the labels include: (3 ≤ feature1 ≤ 5) ⇒ label = 1 and (0.6
≤ feature2 ≤ 0.8) ∧ (4 ≤ feature3 ≤ 5) ⇒ label = 0. Such advice
is more relevant in an discriminative setting but are not ideal for
GANs. Since GANs are shown to be sensitive to the training data
and here the labels are getting generated, they should not be altered
during training. The second is via correlations between features as
preferences (our approach) which allows for faithful representation
of diverse modality.</p>
      <p>Advice injection: After every fixed number of iterations, N, we
calculate the correlation matrix of the generated data G1 and provide
a set of advice  on the correlations between different features.
Consider the following motivating example for the use of correlations as
a form of advice.</p>
      <p>Example: Consider predicting heart attack with 3 features -
cholesterol, blood pressure (BP) and income. The values of the given
features can vary (sometimes widely) between different patients due
to several latent factors (ex, smoking habits). It is difficult to assume
any specific distribution. In other words, it is difcfiult to deduce
whether the values for the features come from the same distribution
(even though the feature values in the data set are similar).
We modify the correlation coefcfiients (for both positive and
negative correlations) between the features by increasing them if the
human advice suggests that two features are highly correlated and
decrease the same if the advice suggests otherwise.</p>
      <p>Example: Continuing the above example, since rise in the
cholesterol level can lead to rise in BP and vice versa, expert advice here
can suggest that cholesterol and BP should be highly correlated.
Also, as income may not contribute directly to BP and cholesterol
levels, another advice here can be to de-correlate cholesterol/BP
and income level.</p>
      <p>
        The example advice rules ∈  are: 1. Correlation(“cholesterol
level",“BP")↑, 2. Correlation(“cholesterol level",“income level")↓
and 3. Correlation(“BP",“income level")↓, where ↑ and ↓ indicate
increase and decrease respectively. Based on the 1st advice we need
to increase the correlation coefficient between cholesterol level and
BP. Then
Here C is the correlation matrix, A is the advice matrix and  is the
factor by which the correlation value is to be augmented. In case
where we need to increase the value of the correlation coefficient, 
should be &gt; 1. We keep  = 1( | C |) . Since -1.0 ≤ ∀ ∈ C ≤ 1.0,
in this case, the value of  ≥ 1.0, leading to enhanced correlation via
Hadamard product. Thus the new correlation matrix Cˆ is,
 1 0.2 0.3   1 1
Cˆ = C ⊙ A = 00..23 0.107 0.107 ⊙  011.3 11
    (2)
 1 0.667 0.3 
= 0.667 1 0.07
 0.3 0.07 1 
 
If the advice says that features have low correlations (2nd rule in
example), we decrease the correlation coefficient. Now,  must be
&lt; 1 and we set  =  (|C |). Since -1 ≤ ∀ ∈ C ≤ 1.0, the value of
 ≤ 1.0. Thus multiplying by  will decrease the correlation value,
and the new correlation matrix is,
After Cˆ1 is constructed, we next generate data satisfying the
constraints. To this effect, we employ the Iman-Conover method [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ],
a distribution free method to define dependencies between
distributional variables based on rank correlations such as Spearman or
Kendell Tau correlations. Since we deal with linear relationships
between the features and assume a normal distribution and that
Pearson coefcfiient has shown to perform equally well with the
Iman-Conover method [
        <xref ref-type="bibr" rid="ref40">40</xref>
        ] due to the close relationship between
Pearson and Spearman correlations, we use the Pearson correlations.
Further, we assume that the features are Gaussian, justified by the
fact that most lab test data is continuous. The Iman-Conover method
consists of the following steps:
      </p>
      <p>
        [Step 1]: Create a random standardized matrix M with values
 ∈ M ∼ Gaussian distribution. This is obtained by the process of
inverse transform sampling described next. Let V1 be a uniformly
distributed random variable and CD F be the cumulative distribution
function. For a sampled point  , CD F ( ) = P ( ≤  ). Thus, to
generate samples, the values  ∼ V are passed through CD F −1 to
obtain the desired values  [CD F −1 ( ) = { |CD F ( ) ≤ ,  ∈
[
        <xref ref-type="bibr" rid="ref1">0, 1</xref>
        ] }]. Thus for Gaussian,
      </p>
      <p>1 ∫  1 ∫ 
CD F ( ) = √2 −∞ exp −22  = √2 0
exp −22 
= [− exp( −2 2 )]0

(4)
The inverse CDF can be thus written as CD F −1 ( ) = 1−exp( −22 ) ≤
 and the desired values  ∈ M can be obtained as  = p2 (1 −  ).
[Step 2]: Calculate the correlation matrix E of M.</p>
      <p>
        [Step 3]: Calculate the Cholesky decomposition F of the
correlation matrix E. Cholesky decomposition [
        <xref ref-type="bibr" rid="ref46">46</xref>
        ] of a positive-definite
matrix is given as the product of a lower triangular matrix and its
conjugate transpose. Note that for Cholesky decomposition to be unique,
the target matrix should be positive definite, (such as the co-variance
matrix) whereas the correlation matrix, used in our algorithm, is only
positive semi-definite. We enforce positive-definiteness by repeated
addition of very small values to the diagonal of the correlation
matrix until positive-definiteness is ensured. Given a symmetric and
positive definite matrix E, its Cholesky decomposition F is such
that E = F · F⊤.
      </p>
      <p>[Step 4]: Calculate the Cholesky decomposition Q of the
correlation matrix obtained after modifications based on human advice, Cˆ.
As above the Cholesky decomposition is such that Cˆ = Q · Q⊤.</p>
      <p>[Step 5]: Calculate the reference matrix T by transforming the
sampled matrix M from step 1 to have the desired correlations of Cˆ,
by using their Cholesky decompositions.</p>
      <p>[Step 6]: Rearrange values in columns of the generated data G1
to have the same ordering as corrresponding column in the reference
matrix T to obtain the final generated data G˜1.</p>
      <p>Cholesky decomposition to model correlations: Given an
randomly generated data set with no correlations P, a correlation matrix
C and its Cholesky decomposition Q, data that faithfully follows
the given correlations ∈ C can be generated by the product of the
obtained lower triangular matrix with the original uncorrelated data
i.e. Pˆ =QP. The correlation of the newly obtained data, Pˆ is,
 (Pˆ ) =  (Pˆ) = E[Pˆ Pˆ ⊤] − E[Pˆ ]E[Pˆ ]⊤</p>
      <p>Pˆ  Pˆ
Since we consider data Pˆ from a Gaussian distribution with zero
mean and unit variance,
 (Pˆ ) =</p>
      <p>E[Pˆ Pˆ ⊤] − E[Pˆ ]E[Pˆ ]⊤
 ˆ</p>
      <p>P
= E[QPQ⊤P⊤] = QE[P P⊤]Q⊤ = QQ⊤ = C
= E[Pˆ Pˆ ⊤] = E[(QP) (QP)⊤]
(5)
(6)
Thus Cholesky decomposition can capture the desired correlations
faithfully and can be used for generating correlated data. Since we
already have a normal sampled matrix M and a calculated correlation
E of M, we need to calculate a reference matrix (step 5).
3.3</p>
    </sec>
    <sec id="sec-7">
      <title>Human-Allied GAN training</title>
      <p>
        Since the human expert advice is provided independent of the GAN
architecture, our method is agnostic of the underlying GAN
architecture. We make use of Wasserstein GAN (WGAN) architecture since
its shown to be more stable while training and can handle mode
collapse [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Only the error backpropagation values differ when we
are using the data generated by the underlying GAN or the data
generated by the Iman-Conover method. Our algorithm starts with
the general process of training a GAN where the generator takes
random noise as an input and generates data which is then passed,
along with the real data, to the discriminator. The discriminator
tries to identify the real and generated data and the error is back
propagated to the generator. After every specified number of
iterations, the correlations between features C in the generated data is
obtained and a new correlation matrix Cˆ, is obtained with respect
to the expert advice (section 3.1). A new data set is generated wrt
Cˆ using the Iman-Conover method (Section 3.2) and then passed to
the discriminator along with the real data set.
4
      </p>
    </sec>
    <sec id="sec-8">
      <title>EXPERIMENTAL EVALUATION</title>
      <sec id="sec-8-1">
        <title>We aim to answer the following questions:</title>
        <p>Q1: Does providing advice to GANs help in generating better
quality data?
Q2: Are GANs with advice effective for data sets that have few
examples?
Q3: How does bad advice affect the quality of generated data?
Q4: How well does human advice handle class imbalance?
Q5: How does our method compare to state-of-the-art GAN
architectures.</p>
        <p>We consider 2 real clinical data sets.
(1) Nephrotic Syndrome is a novel data set of symptoms that
indicate kidney damage. This consists of 50 kidney biopsy
images along with the clinical reports sourced from Dr Lal
PathLabs, India 2. We use the clinical reports that consist of
the values for kidney tissue diagnosis which can confirm the
clinical diagnosis and help to identify high-risk patients and
influence treatment decisions and help medical practitioners</p>
      </sec>
      <sec id="sec-8-2">
        <title>2https://www.lalpathlabs.com/</title>
        <p>
          to plan and prognosticate treatments. The data consists of 19
features with 44 positive and 6 negative examples.
(2) MIMIC database [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ] consists of deidentified information
of patients admitted to critical care units at a large tertiary
care hospital. The features included are predominately time
window aggregations of physiological measurements from
the medical records. We selected relevant lab results, vital
sign observations and feature aggregations. The data consists
of 18 with 5813 positive and 40707 negative examples.
Advice Acquisition: Here we compile the sources from which we
obtain the advice.
        </p>
        <p>(1) Nephrotic Syndrome: This is a novel real data set and the
advice is obtained from a nephrologist in India. According
to the problem statement from the expert, nephrotic syndrome
involves the loss of a lot of protein and nephritic syndrome
involves the loss of a lot of blood through urine. A kidney
biopsy is often required to diagnose the underlying
pathology in patients with suspected glomerular disease. The goal
of the project is to build a clinical support system that
predicts the disease using clinical features, thus reducing the
need of kidney biopsy. Since the data collection is scarce,
a synthetic data set can help in better understanding of the
disease from the clinical features.
(2) MIMIC: The feature set and the expected correlations are
obtained in consultation with trauma experts at a Dallas
hospital.</p>
        <p>
          All experiments were run on a 64-bit Intel(R) Xeon(R) CPU E5-2630
v3 server for 10K epochs. Both the generator and discriminator are
neural networks with 4 hidden layers. To measure the quality of the
generated data we make use of the train on synthetic, test on real
(TSTR) method as proposed in [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ]. We use gradient boosting with
100 estimators and a learning rate of 0.01 as the underlying model.
We train the GAN for 10K epochs and provide correlation advice
every 1K iterations.
        </p>
        <p>Table 1 shows the results of the TSTR method with data generated
with (HA-GAN) and without advice (GAN). It shows that the
data generated with advice has higher TSTR performance than the
data generated without advice across all data sets and all metrics.
Thus, to answer Q1, providing advice to generative adversarial
networks captures the relationship between features better and thus are
able to generate better quality synthetic data.</p>
        <p>Learning with less data: GANs with advice are especially
impressive in nephrotic syndrome data which consists of only 50 examples
across all metrics and is thus very small in size when compared to the
number of samples typically required to train a GAN model. Thus,
we realize an important property of incorporating human guidance in
the GAN model and can answer Q2 affirmatively. The use of advice
opens up the potential of using GANs in presence of sparse data
samples.</p>
        <p>Effect of bad advice: Table 1 also shows the results for data
generated with bad advice (HA-GAN). To simulate bad advice, we
follow a simple process: if the advice says that the correlation
between features should be high, we set the correlations in Cˆ to 0
and if the advice says that the correlation should be low, we set the
correlations in Cˆ to be either 1 or -1 based on whether the original
As results show in table 1, giving bad advice adversely affects the
performance thereby answering Q3.</p>
        <p>The nephrotic syndrome and MIMIC data sets are relatively
unbalanced with a pos to neg ratio of ≈ 8:1 and 1:7 respectively. Most
of the medical data sets, except highly curated data sets, are
unbalanced. A data generator model should be able to handle this
imbalance. Since our method explicitly focuses on the correlations
between features and generates better quality data based on such
relationships between features, our method is quite robust to the
imbalance in the underlying data. This can be seen in the results
in table 1 where advice based data generation outperforms the
nonadvice and bad advice based data generation. Thus, we can answer
Q4 affirmatively.</p>
        <p>
          To answer Q5 we compare our method to 3 GAN architectures,
medGAN [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] which uses an encoder decoder framework for EHR
data generation and its 2 variants medBGAN and medWGAN [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]
and the results are shown in table 1. Our method, with good advice,
outperforms the baseline both domains showing the effectiveness of
our method.
5
        </p>
      </sec>
    </sec>
    <sec id="sec-9">
      <title>CONCLUSION</title>
      <p>
        We presented a new GAN formulation that employs correlation
information between features as advice to generate new correlated
data and train the underlying GAN model. We tested our model
on real clinical data sets and show that incorporating advice helps
generate good quality synthetic medical data. We employ TSTR
method to test the quality of generated data and demonstrated that
the generated data with advice is more aligned with the real data.
There are several future interesting directions. First, providing advice
only when required in an active fashion can allow for significant
reduction in the amount of effort on the human side. Second, there
can be multiple advice options, such as posterior regularization [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ],
that can be used to capture feature relationships explicitly. Third,
although we do not have identifiers in the data, thereby eliminating
the need of differential privacy [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], a general framework that can
uphold the privacy of patient data along the lines of using Cholesky
decomposition [
        <xref ref-type="bibr" rid="ref31 ref7">7, 31</xref>
        ] is a natural next step.
      </p>
    </sec>
    <sec id="sec-10">
      <title>ACKNOWLEDGMENTS</title>
      <p>DSD and SN gratefully acknowledge DARPA Minerva award
FA955019-1-0391. Any opinions, findings, and conclusion or
recommendations expressed in this material are those of the authors and do not
necessarily reflect the view of the DARPA or the US government.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Grigory</given-names>
            <surname>Antipov</surname>
          </string-name>
          , Moez Baccouche, and
          <string-name>
            <surname>Jean-Luc Dugelay</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Face aging with conditional generative adversarial networks</article-title>
          .
          <source>In ICIP.</source>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Martin</given-names>
            <surname>Arjovsky</surname>
          </string-name>
          and
          <string-name>
            <given-names>Leon</given-names>
            <surname>Bottou</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Towards principled methods for training generative adversarial networks</article-title>
          .
          <source>In ICLR.</source>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Martin</given-names>
            <surname>Arjovsky</surname>
          </string-name>
          , Soumith Chintala, and
          <string-name>
            <given-names>Léon</given-names>
            <surname>Bottou</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Wasserstein gan</article-title>
          .
          <source>ICML</source>
          (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Mrinal</given-names>
            <surname>Kanti</surname>
          </string-name>
          <string-name>
            <given-names>Baowaly</given-names>
            ,
            <surname>Chia-Ching</surname>
          </string-name>
          <string-name>
            <given-names>Lin</given-names>
            ,
            <surname>Chao-Lin</surname>
          </string-name>
          <string-name>
            <given-names>Liu</given-names>
            , and
            <surname>Kuan-Ta Chen</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Synthesizing electronic health records using improved generative adversarial networks</article-title>
          .
          <source>JAMA</source>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Darius</given-names>
            <surname>Braziunas</surname>
          </string-name>
          and
          <string-name>
            <given-names>Craig</given-names>
            <surname>Boutilier</surname>
          </string-name>
          .
          <year>2006</year>
          .
          <article-title>Preference elicitation and generalized additive utility</article-title>
          .
          <source>In AAAI.</source>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Anna</surname>
            <given-names>L Buczak</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Steven Babin</surname>
            , and
            <given-names>Linda</given-names>
          </string-name>
          <string-name>
            <surname>Moniz</surname>
          </string-name>
          .
          <year>2010</year>
          .
          <article-title>Data-driven approach for creating synthetic electronic medical records. BMC medical informatics and decision making (</article-title>
          <year>2010</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Jim</given-names>
            <surname>Burridge</surname>
          </string-name>
          .
          <year>2003</year>
          .
          <article-title>Information preserving statistical obfuscation</article-title>
          .
          <source>Statistics and Computing</source>
          (
          <year>2003</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Edward</given-names>
            <surname>Choi</surname>
          </string-name>
          , Siddharth Biswal, Bradley Malin, Jon Duke, Walter F Stewart,
          <string-name>
            <given-names>and Jimeng</given-names>
            <surname>Sun</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Generating Multi-label Discrete Patient Records using Generative Adversarial Networks</article-title>
          .
          <source>In MLHC.</source>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Corinna</given-names>
            <surname>Cortes</surname>
          </string-name>
          and
          <string-name>
            <given-names>Vladimir</given-names>
            <surname>Vapnik</surname>
          </string-name>
          .
          <year>1995</year>
          .
          <article-title>Support-vector networks</article-title>
          .
          <source>Machine Learning</source>
          (
          <year>1995</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Ivo</surname>
            <given-names>D</given-names>
          </string-name>
          <string-name>
            <surname>Dinov</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Volume and value of big healthcare data</article-title>
          .
          <source>Journal of medical statistics and informatics (</source>
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>Cynthia</given-names>
            <surname>Dwork</surname>
          </string-name>
          .
          <year>2008</year>
          .
          <article-title>Differential privacy: A survey of results</article-title>
          .
          <source>In TAMS.</source>
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Cristóbal</surname>
            <given-names>Esteban</given-names>
          </string-name>
          , Stephanie L Hyland, and
          <string-name>
            <given-names>Gunnar</given-names>
            <surname>Rätsch</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Real-valued (medical) time series generation with recurrent conditional gans</article-title>
          .
          <source>arXiv preprint arXiv:1706.02633</source>
          (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>Maayan</given-names>
            <surname>Frid-Adar</surname>
          </string-name>
          , Eyal Klang, Michal Amitai, Jacob Goldberger, and
          <string-name>
            <given-names>Hayit</given-names>
            <surname>Greenspan</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Synthetic data augmentation using GAN for improved liver lesion classification</article-title>
          .
          <source>In ISBI.</source>
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Glenn</surname>
            <given-names>M Fung</given-names>
          </string-name>
          , Olvi L Mangasarian, and Jude W Shavlik.
          <year>2003</year>
          .
          <article-title>Knowledge-based support vector machine classifiers</article-title>
          .
          <source>In NIPS.</source>
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Kuzman</surname>
            <given-names>Ganchev</given-names>
          </string-name>
          , Jennifer Gillenwater,
          <string-name>
            <given-names>Ben</given-names>
            <surname>Taskar</surname>
          </string-name>
          , et al.
          <year>2010</year>
          .
          <article-title>Posterior regularization for structured latent variable models</article-title>
          .
          <source>JMLR</source>
          (
          <year>2010</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Ian</surname>
            <given-names>Goodfellow</given-names>
          </string-name>
          , Yoshua Bengio, and
          <string-name>
            <given-names>Aaron</given-names>
            <surname>Courville</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Deep learning</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Ian</surname>
            <given-names>Goodfellow</given-names>
          </string-name>
          , Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and
          <string-name>
            <given-names>Yoshua</given-names>
            <surname>Bengio</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>Generative adversarial nets</article-title>
          .
          <source>In NIPS.</source>
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>Peter</given-names>
            <surname>Groves</surname>
          </string-name>
          , Basel Kayyali, David Knott, and Steve Van Kuiken.
          <year>2016</year>
          .
          <article-title>The'big data'revolution in healthcare: Accelerating value and innovation</article-title>
          . (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>John</surname>
            <given-names>T Guibas</given-names>
          </string-name>
          ,
          <article-title>Tejpal S Virdi,</article-title>
          and
          <string-name>
            <surname>Peter S Li</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Synthetic medical images from dual generative adversarial networks</article-title>
          .
          <source>arXiv preprint arXiv:1709</source>
          .
          <year>01872</year>
          (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>Ishaan</surname>
            <given-names>Gulrajani</given-names>
          </string-name>
          , Faruk Ahmed, Martin Arjovsky, Vincent Dumoulin, and Aaron C Courville.
          <year>2017</year>
          .
          <article-title>Improved training of wasserstein gans</article-title>
          .
          <source>In NIPS.</source>
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <surname>Haroun</surname>
            <given-names>Habeeb</given-names>
          </string-name>
          , Ankit Anand, Mausam Mausam, and
          <string-name>
            <given-names>Parag</given-names>
            <surname>Singla</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Coarseto-fine lifted MAP inference in computer vision</article-title>
          . In IJCAI.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <surname>Zhiting</surname>
            <given-names>Hu</given-names>
          </string-name>
          , Zichao Yang, Russ R Salakhutdinov, LIANHUI Qin, Xiaodan Liang, Haoye Dong, and Eric P Xing.
          <year>2018</year>
          .
          <article-title>Deep Generative Models with Learnable Knowledge Constraints</article-title>
          . In NeurIPS.
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <surname>Ronald L Iman</surname>
          </string-name>
          and
          <string-name>
            <surname>William-Jay Conover</surname>
          </string-name>
          .
          <year>1982</year>
          .
          <article-title>A distribution-free approach to inducing rank correlation among input variables</article-title>
          .
          <source>Communications in StatisticsSimulation and Computation</source>
          (
          <year>1982</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <surname>Alistair</surname>
            <given-names>EW Johnson</given-names>
          </string-name>
          , Tom J Pollard, Lu Shen,
          <string-name>
            <given-names>H Lehman</given-names>
            <surname>Li-wei</surname>
          </string-name>
          , Mengling Feng, Mohammad Ghassemi, Benjamin Moody, Peter Szolovits, Leo Anthony Celi, and Roger G Mark.
          <year>2016</year>
          .
          <article-title>MIMIC-III, a freely accessible critical care database. Scientific data (</article-title>
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <surname>Tero</surname>
            <given-names>Karras</given-names>
          </string-name>
          , Samuli Laine, and
          <string-name>
            <given-names>Timo</given-names>
            <surname>Aila</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>A style-based generator architecture for generative adversarial networks</article-title>
          .
          <source>In CVPR.</source>
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>Harold</given-names>
            <surname>William</surname>
          </string-name>
          Kuhn and
          <string-name>
            <given-names>Albert William</given-names>
            <surname>Tucker</surname>
          </string-name>
          .
          <year>1953</year>
          .
          <article-title>Contributions to the Theory of Games.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <surname>Gautam</surname>
            <given-names>Kunapuli</given-names>
          </string-name>
          , Phillip Odom, Jude W Shavlik, and
          <string-name>
            <given-names>Sriraam</given-names>
            <surname>Natarajan</surname>
          </string-name>
          .
          <year>2013</year>
          .
          <article-title>Guiding autonomous agents to better behaviors through human advice</article-title>
          .
          <source>In ICDM.</source>
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <surname>Quoc</surname>
            <given-names>V Le</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Alex J Smola</surname>
            , and
            <given-names>Thomas</given-names>
          </string-name>
          <string-name>
            <surname>Gärtner</surname>
          </string-name>
          .
          <year>2006</year>
          .
          <article-title>Simpler knowledge-based support vector machines</article-title>
          .
          <source>In ICML.</source>
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <surname>Yann</surname>
            <given-names>LeCun</given-names>
          </string-name>
          , Yoshua Bengio, and
          <string-name>
            <given-names>Geoffrey</given-names>
            <surname>Hinton</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <article-title>Deep learning</article-title>
          .
          <source>Nature</source>
          (
          <year>2015</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>Minjun</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Haozhi</given-names>
            <surname>Huang</surname>
          </string-name>
          , Lin Ma, Wei Liu, Tong Zhang, and
          <string-name>
            <given-names>Yugang</given-names>
            <surname>Jiang</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Unsupervised image-to-image translation with stacked cycle-consistent adversarial networks</article-title>
          .
          <source>In ECCV.</source>
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>Yaping</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Minghua</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Qiwei</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>and Wei</given-names>
            <surname>Zhang</surname>
          </string-name>
          .
          <year>2011</year>
          .
          <article-title>Enabling multilevel trust in privacy preserving data mining</article-title>
          .
          <source>TKDE</source>
          (
          <year>2011</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [32]
          <string-name>
            <given-names>Yujia</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Kevin</given-names>
            <surname>Swersky</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Rich</given-names>
            <surname>Zemel</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <article-title>Generative moment matching networks</article-title>
          .
          <source>In ICML.</source>
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          [33]
          <string-name>
            <surname>Ming-Yu</surname>
            <given-names>Liu</given-names>
          </string-name>
          , Thomas Breuel, and
          <string-name>
            <given-names>Jan</given-names>
            <surname>Kautz</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Unsupervised image-to-image translation networks</article-title>
          .
          <source>In NIPS.</source>
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          [34]
          <string-name>
            <surname>Ming-Yu Liu</surname>
            and
            <given-names>Oncel</given-names>
          </string-name>
          <string-name>
            <surname>Tuzel</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Coupled generative adversarial networks</article-title>
          .
          <source>In NIPS.</source>
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          [35]
          <string-name>
            <surname>Faisal</surname>
            <given-names>Mahmood</given-names>
          </string-name>
          , Richard Chen, and
          <string-name>
            <surname>Nicholas</surname>
          </string-name>
          J Durr.
          <year>2018</year>
          .
          <article-title>Unsupervised reverse domain adaptation for synthetic medical images via adversarial training</article-title>
          .
          <source>IEEE transactions on medical imaging</source>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          [36]
          <string-name>
            <given-names>V. K.</given-names>
            <surname>Mansinghka</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Kemp</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. B.</given-names>
            <surname>Tenenbaum</surname>
          </string-name>
          , and
          <string-name>
            <given-names>T. L.</given-names>
            <surname>Griffiths</surname>
          </string-name>
          .
          <year>2006</year>
          .
          <article-title>Structured Priors for Structure Learning</article-title>
          .
          <source>In UAI.</source>
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          [37]
          <string-name>
            <surname>Xudong</surname>
            <given-names>Mao</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Qing</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Haoran</given-names>
            <surname>Xie</surname>
          </string-name>
          , Raymond YK Lau, Zhen
          <string-name>
            <surname>Wang</surname>
          </string-name>
          , and Stephen Paul Smolley.
          <year>2017</year>
          .
          <article-title>Least squares generative adversarial networks</article-title>
          .
          <source>In ICCV.</source>
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          [38] Tom M Mitchell.
          <year>1980</year>
          .
          <article-title>The need for biases in learning generalizations</article-title>
          . Department of Computer Science, Laboratory for Computer Science Research, Rutgers Univ. New Jersey.
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          [39]
          <string-name>
            <surname>Takeru</surname>
            <given-names>Miyato</given-names>
          </string-name>
          , Toshiki Kataoka, Masanori Koyama, and
          <string-name>
            <given-names>Yuichi</given-names>
            <surname>Yoshida</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Spectral normalization for generative adversarial networks</article-title>
          .
          <source>ICLR</source>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref40">
        <mixed-citation>
          [40]
          <string-name>
            <given-names>Klemen</given-names>
            <surname>Naveršnik</surname>
          </string-name>
          and
          <string-name>
            <given-names>Klemen</given-names>
            <surname>Rojnik</surname>
          </string-name>
          .
          <year>2012</year>
          .
          <article-title>Handling input correlations in pharmacoeconomic models</article-title>
          . Value in Health (
          <year>2012</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref41">
        <mixed-citation>
          [41]
          <string-name>
            <given-names>P.</given-names>
            <surname>Odom</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Khot</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Porter</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S.</given-names>
            <surname>Natarajan</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <article-title>Knowledge-Based Probabilistic Logic Learning</article-title>
          .
          <source>In AAAI.</source>
        </mixed-citation>
      </ref>
      <ref id="ref42">
        <mixed-citation>
          [42]
          <string-name>
            <given-names>Phillip</given-names>
            <surname>Odom</surname>
          </string-name>
          and
          <string-name>
            <given-names>Sriraam</given-names>
            <surname>Natarajan</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <article-title>Active advice seeking for inverse reinforcement learning</article-title>
          .
          <source>In AAAI.</source>
        </mixed-citation>
      </ref>
      <ref id="ref43">
        <mixed-citation>
          [43]
          <string-name>
            <given-names>Phillip</given-names>
            <surname>Odom</surname>
          </string-name>
          and
          <string-name>
            <given-names>Sriraam</given-names>
            <surname>Natarajan</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Human-guided learning for probabilistic logic models</article-title>
          .
          <source>Frontiers in Robotics and AI</source>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref44">
        <mixed-citation>
          [44]
          <string-name>
            <surname>Michela</surname>
            <given-names>Paganini</given-names>
          </string-name>
          , Luke de Oliveira, and
          <string-name>
            <given-names>Benjamin</given-names>
            <surname>Nachman</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>CaloGAN: Simulating 3D high energy particle showers in multilayer electromagnetic calorimeters with generative adversarial networks. Physical Review D (</article-title>
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref45">
        <mixed-citation>
          [45]
          <string-name>
            <surname>Alec</surname>
            <given-names>Radford</given-names>
          </string-name>
          , Luke Metz, and
          <string-name>
            <given-names>Soumith</given-names>
            <surname>Chintala</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Unsupervised representation learning with deep convolutional generative adversarial networks</article-title>
          .
          <source>ICLR</source>
          (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref46">
        <mixed-citation>
          [46]
          <string-name>
            <surname>Ernest</surname>
            <given-names>M</given-names>
          </string-name>
          <string-name>
            <surname>Scheuer and David S Stoller</surname>
          </string-name>
          .
          <year>1962</year>
          .
          <article-title>On the generation of normal random vectors</article-title>
          .
          <source>Technometrics</source>
          (
          <year>1962</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref47">
        <mixed-citation>
          [47]
          <string-name>
            <surname>Bernhard</surname>
            <given-names>Schölkopf</given-names>
          </string-name>
          , Patrice Simard,
          <string-name>
            <surname>Alex J Smola</surname>
            , and
            <given-names>Vladimir</given-names>
          </string-name>
          <string-name>
            <surname>Vapnik</surname>
          </string-name>
          .
          <year>1998</year>
          .
          <article-title>Prior knowledge in support vector kernels</article-title>
          .
          <source>In Advances in neural information processing systems</source>
          .
          <volume>640</volume>
          -
          <fpage>646</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref48">
        <mixed-citation>
          [48]
          <string-name>
            <surname>Rittika</surname>
            <given-names>Shamsuddin</given-names>
          </string-name>
          , Barbara M Maweu,
          <string-name>
            <given-names>Ming</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>and Balakrishnan</given-names>
            <surname>Prabhakaran</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Virtual patient model: an approach for generating synthetic healthcare time series data</article-title>
          .
          <source>In ICHI.</source>
        </mixed-citation>
      </ref>
      <ref id="ref49">
        <mixed-citation>
          [49]
          <string-name>
            <surname>Jude</surname>
            <given-names>W</given-names>
          </string-name>
          <string-name>
            <surname>Shavlik and Geoffrey G Towell</surname>
          </string-name>
          .
          <year>1989</year>
          .
          <article-title>Combining explanation-based learning and artificial neural networks</article-title>
          .
          <source>In Proceedings of the sixth international workshop on Machine learning. Elsevier.</source>
        </mixed-citation>
      </ref>
      <ref id="ref50">
        <mixed-citation>
          [50]
          <string-name>
            <surname>Geoffrey</surname>
            <given-names>G</given-names>
          </string-name>
          <string-name>
            <surname>Towell and Jude W Shavlik</surname>
          </string-name>
          .
          <year>1994</year>
          .
          <article-title>Knowledge-based artificial neural networks</article-title>
          .
          <source>Artificial intelligence</source>
          (
          <year>1994</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref51">
        <mixed-citation>
          [51]
          <string-name>
            <surname>Yan</surname>
            <given-names>Wang</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Biting Yu</surname>
          </string-name>
          , Lei Wang,
          <string-name>
            <surname>Chen Zu</surname>
          </string-name>
          , David S Lalush, Weili Lin,
          <string-name>
            <surname>Xi Wu</surname>
          </string-name>
          ,
          <string-name>
            <surname>Jiliu Zhou</surname>
            ,
            <given-names>Dinggang</given-names>
          </string-name>
          <string-name>
            <surname>Shen</surname>
            , and
            <given-names>Luping</given-names>
          </string-name>
          <string-name>
            <surname>Zhou</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>3D conditional generative adversarial networks for high-quality PET image estimation at low dose</article-title>
          .
          <source>NeuroImage</source>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref52">
        <mixed-citation>
          [52]
          <string-name>
            <surname>Zongwei</surname>
            <given-names>Wang</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Xu Tang</surname>
            , Weixin Luo, and
            <given-names>Shenghua</given-names>
          </string-name>
          <string-name>
            <surname>Gao</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Face aging with identity-preserved conditional generative adversarial networks</article-title>
          .
          <source>In CVPR.</source>
        </mixed-citation>
      </ref>
      <ref id="ref53">
        <mixed-citation>
          [53]
          <string-name>
            <surname>Jiajun</surname>
            <given-names>Wu</given-names>
          </string-name>
          , Chengkai Zhang, Tianfan Xue, Bill Freeman, and
          <string-name>
            <given-names>Josh</given-names>
            <surname>Tenenbaum</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Learning a probabilistic latent space of object shapes via 3d generativeadversarial modeling</article-title>
          .
          <source>In NIPS.</source>
        </mixed-citation>
      </ref>
      <ref id="ref54">
        <mixed-citation>
          [54]
          <string-name>
            <given-names>S.</given-names>
            <surname>Yang</surname>
          </string-name>
          and
          <string-name>
            <given-names>S.</given-names>
            <surname>Natarajan</surname>
          </string-name>
          .
          <year>2013</year>
          .
          <article-title>Knowledge Intensive Learning: Combining Qualitative Constraints with Causal Independence for Parameter Learning in Probabilistic Models</article-title>
          .
          <source>In ECMLPKDD.</source>
        </mixed-citation>
      </ref>
      <ref id="ref55">
        <mixed-citation>
          [55]
          <string-name>
            <surname>Jun-Yan</surname>
            <given-names>Zhu</given-names>
          </string-name>
          , Taesung Park,
          <source>Phillip Isola, and Alexei A Efros</source>
          .
          <year>2017</year>
          .
          <article-title>Unpaired image-to-image translation using cycle-consistent adversarial networks</article-title>
          .
          <source>In ICCV.</source>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>