<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Blood Glucose Prediction for Type 1 Diabetes Using Generative Adversarial Networks</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Taiyu Zhu</string-name>
          <email>taiyu.zhu17@imperial.ac.uk</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Xi Yao</string-name>
          <email>x.yao19@imperial.ac.uk</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Kezhi Li</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Pau Herrero</string-name>
          <email>pherrero@imperial.ac.uk</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Pantelis Georgiou</string-name>
          <email>pantelis@imperial.ac.uk</email>
        </contrib>
      </contrib-group>
      <abstract>
        <p>Maintaining blood glucose in a target range is essential for people living with Type 1 diabetes in order to avoid excessive periods in hypoglycemia and hyperglycemia which can result in severe complications. Accurate blood glucose prediction can reduce this risk and enhance early interventions to improve diabetes management. However, due to the complex nature of glucose metabolism and the various lifestyle related factors which can disrupt this, diabetes management still remains challenging. In this work we propose a novel deep learning model to predict future BG levels based on the historical continuous glucose monitoring measurements, meal ingestion, and insulin delivery. We adopt a modified architecture of the generative adversarial network that comprises of a generator and a discriminator. The generator computes the BG predictions by a recurrent neural network with gated recurrent units, and the auxiliary discriminator employs a one-dimensional convolutional neural network to distinguish between the predictive and real BG values. Two modules are trained in an adversarial process with a combination of loss. The experiments were conducted using the OhioT1DM dataset that contains the data of six T1D contributors over 40 days. The proposed algorithm achieves an average root mean square error (RMSE) of 18:34 0:17 mg/dL with a mean absolute error (MAE) of 13:37 0:18 mg/dL for the 30-minute prediction horizon (PH) and an average RMSE of 32:31 0:46 mg/dL with a MAE of 24:20 0:42 for the 60-minute PH. The results are compared for clinical relevance using the Clarke error grid which confirms the promising performance of the proposed model.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>INTRODUCTION</title>
      <p>
        Diabetes is a chronic metabolic disorder that affects more than 400
million people worldwide with an increasing global prevalence [
        <xref ref-type="bibr" rid="ref27">27</xref>
        ].
Due to an absence of insulin production from the pancreatic cells,
people living with Type 1 diabetes (T1D) require long-term
selfmanagement through exogenous insulin delivery to maintain blood
glucose (BG) levels in a normal range. In this regard, accurate
glucose prediction has great potential to improve diabetes
management, enabling proactive actions to reduce the occurrence of adverse
glycemic events, including hypoglycemia and hyperglycemia.
      </p>
      <p>
        In recent years, empowered by the advances in wearable devices
and data-driven techniques, different BG prediction algorithms have
been proposed and validated in clinical practice [
        <xref ref-type="bibr" rid="ref29">29</xref>
        ]. Among these,
continuous glucose monitoring (CGM) is an essential technology
that measures BG levels and provides readings in real-time. CGM
has produced a vast amount of BG data with its increasing use in the
diabetes population. Taking advantage of this, the emergence of deep
learning algorithms for BG prediction has achieved recent success
and outperformed several conventional machine learning approaches
in terms of accuracy [
        <xref ref-type="bibr" rid="ref1 ref16 ref17 ref23 ref28">1, 16, 17, 23, 28</xref>
        ]. Generally, the major
challenge of BG prediction lies in accounting for the intra- and
interperson variability that leads to various glucose responses under
different conditions [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ]. Furthermore, many external events and factors
can influence glucose dynamics, such as meal ingestion, physical
exercise, psychological stress, and illness. Deep learning is powerful
at extracting hidden representations from large-scale raw data [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ],
making it suitable for accounting for the complexity of glucose
dynamics in diabetes.
      </p>
      <p>
        In this work, we propose a novel deep learning model for BG
prediction using a modified generative adversarial network (GAN). As a
recent breakthrough in the field of deep learning, GANs have shown
promising performance on various tasks, such as generating realistic
images [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], synthesizing electronic health records [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] and predicting
financial time series [
        <xref ref-type="bibr" rid="ref31">31</xref>
        ]. Normally, a GAN framework is composed
of two deep neural networks (DNNs) models as the generator and the
discriminator, respectively. They are trained simultaneously through
an adversarial process [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. The proposed generator captures feature
maps of the multi-variant physiological waveform data and
generates predictive BG samples, while the discriminator is designed to
distinguish the real data from generated ones. To model the temporal
dynamics of BG data, we adopt a recurrent neural network (RNN)
in the generator and a one-dimensional convolutional neural network
(CNN) in the discriminator with dilation factors in each DNN layer
to expand receptive fields, which have been verified as adequate
network structures for BG prediction in our previous works [
        <xref ref-type="bibr" rid="ref17 ref33 ref5">5, 17, 33</xref>
        ].
2
2.1
      </p>
    </sec>
    <sec id="sec-2">
      <title>METHODS</title>
    </sec>
    <sec id="sec-3">
      <title>Dataset and Pre-processing</title>
      <p>
        The data that we used to develop the model is the OhioT1DM
dataset, provided by the Blood Glucose Level Prediction (BGLP)
Challenge [
        <xref ref-type="bibr" rid="ref20 ref21">20, 21</xref>
        ]. It was produced by collecting BG-relevant data
on 12 people with T1D over an eight-week period. The first half of
the cohort released for the 2018 BGLP challenge was used for model
pre-training, and we focus on the performance of the rest six
individuals that numbered 540, 544, 552, 567, 584, and 596. The dataset
contains BG levels collected by CGM readings every five minutes,
insulin delivery from insulin pumps, self-reported events (such as
meal, work, sleep, psychological stress, and physical exercise) via
a smartphone app and physical activity by a sensor band. However,
Pre-processing
      </p>
      <p>Dataset</p>
      <p>XT</p>
      <sec id="sec-3-1">
        <title>Raw Data</title>
        <p>Input
✕
X(t+1-L):t
GRU Cell
σ
✕
1σ</p>
      </sec>
      <sec id="sec-3-2">
        <title>Generator</title>
        <p>+
✕
tanh</p>
        <p>Synthetic Sequence
Gt+1,Gt+2, … , GGˆt+tw+w
Prediction</p>
        <p>Gˆt+w
Gt+w
Supervised Loss</p>
      </sec>
      <sec id="sec-3-3">
        <title>Discriminator</title>
        <p>Real/Synthetic
Adversarial Loss
there are unavoidable differences between the collected data and
actual physiological states. For example, the CGM sensor measures
interstitial fluid glucose level and then estimate BG levels by applying
signal processing techniques, such as filtering and calibration
algorithms. The meal and insulin are discrete values manually input by
users, instead of series of carbohydrates and insulin on board.</p>
        <p>It should be noted that the dataset contains many missing gaps
and outliers affecting BG levels, both in the training and testing
sets, mainly due to CGM signal loss, sensor noise (e.g.,
compression artifacts), or some usage reasons, such as sensor replacement
and calibration. To compensate for some of the missing data, we
apply linear interpolation to fill the missing sequences in the training
sets, while we only extrapolate missing values in the testing set to
ensure that the future information is not involved as partial inputs
in the prediction. We then align processed BG samples and other
features, e.g. exogenous events, with the same resolution of CGM
measurements, and normalize them to form a N -step time series:
XN = [x1; : : : ; xN] 2 RN d, where x is a d-dimensional vector
mapping the multivariate data at each timestep.
2.2</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Problem Formulation</title>
      <p>
        Considering a target prediction horizon (PH) (e.g. 30 or 60 minutes),
the goal of the predictor is to estimate the future BG levels Gt+w
of individuals given past and current physiological states, where w is
the number of timesteps determined by PH and CGM resolution (e.g.
5 minutes). Hence, the objective of predictor is consistent with that
of GANs, aiming to learn the DNN approximator p^ from the pattern
of glucose dynamics p measured in the human body, which can be
expressed by the form of the Kullback-Leibler divergence [
        <xref ref-type="bibr" rid="ref30">30</xref>
        ]:
min D((p(Gt+wjX1:t)jj(p^(Gt+wjX1:t))
p^
(1)
where D is a measurement of the distance between distributions.
Thus, we need to select highly-related data features to represent
the physiological state. Referring to some previous work and
hyperparameter tuning [
        <xref ref-type="bibr" rid="ref16 ref17 ref22 ref23">16, 17, 22, 23</xref>
        ], we use X , [G; M; I] as the
physiological time series, where G is pre-processed CGM
measurements (mg/dL); M denotes the carbohydrate amount of meal
ingestion (g); and I is the bolus insulin delivery (U). In order to reduce
the bias in the supervised learning, we set the changes of BG levels
in PH as the training targets of the generator: Gt = Gt+w Gt.
Then the predictive BG level G^t+w from the generator is defined as
follows:
      </p>
      <p>
        G^t+w = fG (Xt+1 L:t) + Gt
(2)
where fG represents the parameters of the generator. Instead of
using the whole series, we divide X into small contiguous sequences
with a length of L as a sliding window, then feed them into the deep
generative model in a form of mini-batches, aiming at improving
stability and generalization of the model [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. According to the feature
selection in [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ] and the model validation, we empirically set L = 18
which indicates that the input contains 1.5 hours historical data.
2.3
      </p>
    </sec>
    <sec id="sec-5">
      <title>System Architecture</title>
      <p>
        The RNN-based algorithms performed well in BG level prediction
in previous studies [
        <xref ref-type="bibr" rid="ref1 ref23 ref28">1, 23, 28</xref>
        ]. Thus, we instantiate a three-layer
RNN with 32 hidden units to build the generator, which can be
seen as a typical setup of time series GANs [
        <xref ref-type="bibr" rid="ref24 ref30 ref9">9, 24, 30</xref>
        ]. In general,
vanilla RNN architecture faces the problem of gradient vanishing
and exploding, making it difficult to capture long-term dependencies.
Thus, the gated RNN units are proposed to meet this challenge using
element-wise gating functions [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], including long short-term
memory (LSTM) units [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] and gated recurrent units (GRUs) [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
Compared to the vanilla RNN, the gated units are able to control the flow
of information inside units by a set of gates, which allows an
easier backward propagation process. Compared to the LSTM, the GRU
was proposed more recently and removed the memory unit. This cell
structure uses less parameters and computes the output more
efficiently [
        <xref ref-type="bibr" rid="ref32">32</xref>
        ] . During the hyper-parameter tuning, GRU-based
algorithms also achieved the best predictive outcomes, so we naturally
adopt GRU cells in the RNNs .
      </p>
      <p>As depicted in Figure 1, the multi-dimensional input is fed into a
RNN with GRU cells given a state length of L. Then the data is
processed by a set of hidden neurons to calculate the last cell state Ct.
A fully connected (FC) layer with weights WF C and a bias bF C are
used to model the final scalar output: G^t = WF C Ct + bF C .
Finally, after adding the current BG level to predictive glucose change,
we obtain the output G^t+w.</p>
      <p>
        In general, the prediction performance degrades with the increase
of PH, due to the complicated physiological conditions of people
with T1D and the uncertainties of exogenous events between t and
t+w. For instance, if there was a meal intake with large carbohydrate
20-30 minutes before t + w, the BG level would raise fast and make
the target Gt suddenly increase. These cases occur frequently in
the daytime with a large PH, which could affect a supervised learning
model to achieve global optimum. This motivated us to make use
of the information between t and t + w during the training process
to investigate the contiguous glucose change. Therefore, we append
the predictive BG level to the end of series Gt+1:t+w 1 to form a
synthetic sequence y^ and use Gt+1:t+w as the corresponding real
sequence y. Then we introduce a CNN-based discriminator to extract
features and distinguish the real from synthetic sequences, benefiting
from the good classification ability of CNNs [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. There are three
one-dimensional (1-D) causal CNN layers employed with rectified
linear unit (ReLU) activation and 32 hidden units to compute the final
binary output. The discriminator is expected to classify the real and
synthetic sequences by 1 and 0, while the generator is pitted against
the discriminator and aims to estimate a BG value that is close to the
real BG distribution over the PH. Thus the loss of discriminator is
computed by cross-entropy. Consequently, this adversarial training
contains two loss functions LG and LD for the generator and the
discriminator respectively, which are given by
      </p>
      <p>LD =</p>
      <p>LG =
1LSL +</p>
      <p>
        2m
1 Xm [ log fD(y(i))
m i=1
m
X log(1
i=1
(log(1
fD(y^(i)));
fD(y^(i))))];
(3)
(4)
where fD represents the calculation in the discriminator; LSL
is the means square error loss of supervised learning: LSL =
Pim=1(Gt(+i)w G^t(+i)w)2; 1 and 2 are used to adjust the ratio
between supervised loss and adversarial loss [
        <xref ref-type="bibr" rid="ref31">31</xref>
        ]; and m stands for
the mini-batch size. In practice, we employ two separate Adam
optimizer [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] to minimize LG and LD with batch size of 512 and
learning rate of 0.0001.
      </p>
      <p>
        Moreover, we introduce dilation to both the RNN and the CNN
layers [
        <xref ref-type="bibr" rid="ref26 ref3">3, 26</xref>
        ], which has shown the promising performance of BG
level prediction in previous work [
        <xref ref-type="bibr" rid="ref17 ref32 ref33 ref5">5, 17, 32, 33</xref>
        ]. By skipping certain
number connections between neurons, the receptive field of the DNN
layers can be exponentially increased, which is helpful to capture
long-term temporal dependencies in the BG series. In particular, the
dilation of layer l is set to rl = 2l 1, increasing from the bottom
layer to the top layer. The computation of DNN layers are defined as
follows:
h(l) = fN (ht( )rl ; int(l 1))
      </p>
      <p>t
where ht(l) and int(l 1) are the output and input of layer l at timestep
t; fN denotes the computation in hidden neurons, referring to
convolution and cell operation in CNN and RNN layers, respectively. As
a feed-forward neural network, the CNN hidden units fetch all the
inputs from the layer at a lower level ( = l 1), whereas RNNs
skip cell state by rl 1 timesteps to perform the recursive operation
( = l).
2.4</p>
    </sec>
    <sec id="sec-6">
      <title>Training and Validation</title>
      <p>
        The training and testing sets are separately provided by the BGLP
challenge, which contains the data for around 40 and 10 days,
respectively. To tune the hyper-parameters by grid search, we
validated the models by the same range of hyper-parameters values as
in our previous work [
        <xref ref-type="bibr" rid="ref32">32</xref>
        ]. We considered many validation methods,
such as simple splitting, k-fold cross-validation, and blocked
crossvalidation [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Due to the temporal dependencies and limited size of
the training set, we use the last 20% data of the training set to validate
the models and guarantee that future information is not involved in
current prediction. The early-stop technique is applied to avoid
overfitting; we stop the training process when the validation loss keeps
increasing. In particular, we set the maximum number of epochs to
3000 with stopping patience of 50. The data sufficiency and
overfitting occurrences are further investigated by means of the learning
curves.
2.5
      </p>
    </sec>
    <sec id="sec-7">
      <title>Metrics</title>
      <p>A set of metrics is applied to evaluate the performance of the GAN
model, including root mean square error (RMSE) (mg/dL), mean
absolute error (MAE) (mg/dL), which are denoted as:</p>
      <p>RMSE = tuu N1
v</p>
      <p>N
X(Gk
k=1</p>
      <p>G^k)2; MAE =
After tuning the hyper-parameters, we tested the model on the
testing sets. Table 1 shows the RMSE and MAE results for the PH of
30 minutes and 60 minutes. Considering the randomness of the
initial weights in DNNs, we conducted 10 simulations and reported
results by Mean SD, where SD is the standard deviation. The
average (AVG) RMSE and MAE over all 6 contributors respectively
achieve 18:34 0:17 and 13:37 0:18 mg/dL for 30-minute PH,
and 32:21 0:46 and 24:20 0:42 for 60-minute PH. The best
RMSE and MAE results in experiments are also presented in the last
row, which are slightly smaller than the average results. It is noted the
standard deviation of multiple simulations is small, which indicates
the stability of the model.
(5)</p>
      <p>A C E
0 0 50 100 150 200 250</p>
      <p>Reference Concentration (mg/3d0L0) 350 400</p>
      <p>A C E
0 0 50 100 150 200 250</p>
      <p>Reference Concentration (mg/3d0L0) 350 400
(a) 30-minute PH
(b) 60-minute PH
As shown in Table 2, the majority of the CEG points are located
in zones A and B. These zones signify that the data is within 20%
value of the reference, where the treatment suggestions are
appropriate regardless of the prediction error. It indicates the high clinical
accuracy of the proposed model. The percentage of zone D is small
for the 30-minute PH and increases for the 60-minute PH. The points
in zone D mean the predictive model missed the hypoglycemia or
hyperglycemia events and could lead to poor treatment. In Figure 2b,
the most error points are concentrated on the bottom-right corner of
the left panel of zone D. It reveals that the model outputs higher
predictions when BG levels enter the hypoglycemia region, which is
undesirable in the clinical setting. Figure 3 shows the
corresponding BG curves for the contributor 544, where the findings from CEG
analysis can be validated, and time lags between the predictions and
measurements can be observed. The overestimation is observed in
several BG regions with low BG levels or a sharp decrease. Aligning
the error region with the timesteps, we find that some of the
misestimation occurs in nocturnal hypoglycemia. Similar findings are
identified by the CEG analysis and BG curves of the other
contributors. Therefore, future work will include training and switching
between different models for different glucose regions, evaluated by
more advanced error grid analysis.</p>
      <p>
        During the experiments, we explored Tikhonov regularization to
filter out the outliers in training sets, as described in [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. However, it
was prone to degrade the validation performance but largely reduce
the training loss. Then we used the 2018 OhioT1DM dataset [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ] and
the in silico datasets from UVA/Padova T1D simulator [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] for model
pre-training. The simulator produced data of an average virtual adult
subject with the scenarios defined in [
        <xref ref-type="bibr" rid="ref32">32</xref>
        ] over 360 simulated days.
The population model was trained by 5 epochs and then fine-tuned
300
)
L
/gd250
m
(
lve200
e
L
se150
o
c
luG100
50
1:00
350
300
)
L
/gd250
m
(
lve200
e
L
se150
o
c
luG100
Measurement
      </p>
      <p>Prediction
(a) 30-minute PH
13:00
Time
13:00</p>
      <p>
        Time
5:00
9:00
17:00
21:00
1:00
(b) 60-minute PH
by subject-specific data, but the average validation RMSE slightly
increased by around 0.5 mg/dL, compared with the models without
pre-training. As shown in Table 1, there are two groups: one
including contributors 544, 552, and 596 with better RMSE and MAE
performance, and the other including contributors 540, 567 and 584.
We introduced the data from the former group to pre-train a
population model for the latter group, but the RMSE almost remained
unchanged. Thus, one explanation of the pre-training performance is
that large inter-person variability exists. For example, in the testing
set, contributor 552 has a gap of 1415 missing data points ( 5 days),
and contributor 567 did not record the meal ingestion, for which we
reduced the dimension of the input data. To this end, multiple
prepossessing methods are needed to mitigate these missing or
incorrect inputs, such as the detection of unannounced meals. In addition,
as future work, we consider incorporating personalized
physiological and behavioral models [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ], such as insulin and carbohydrate on
board, to better explain the observed variability.
      </p>
      <p>
        Compared with the RNN prediction model in our previous
work [
        <xref ref-type="bibr" rid="ref32">32</xref>
        ], the GAN model achieved better validation performance
and smaller RMSE for most of the data contributors in the training
process, especially for the 60-minute PH. During the testing phase,
the GAN model can output the predictions without using the
discriminator. Hence, the complexity of the proposed model is similar to that
of the conventional RNN models, which can be easily implemented
on smartphone applications [
        <xref ref-type="bibr" rid="ref16 ref17">16, 17</xref>
        ] to provide real-time predictions
and control insulin pump via Bluetooth connectivity. The code
corresponding to this work is available at:
https://bitbucket.org/deeplearning-healthcare/glugan.
In this work, a novel deep learning model using a modified GAN
architecture is designed to predict BG levels for people with T1D.
We developed the personalized models and conducted multiple
evaluations for each data contributor in the OhioT1DM dataset. The
proposed model achieves promising prediction performance for
30minute and 60-minute PH in terms of average RMSE and MAE. The
CEG analysis further indicates good clinical accuracy, but there are
opportunities for enhancement. In particular the model falls short
sometimes in capturing a small number of hypoglycemia events.
Nevertheless, the model is able capture most of the individual
glucose dynamics and has clear potential to be adopted in actual clinical
applications.
      </p>
    </sec>
    <sec id="sec-8">
      <title>ACKNOWLEDGEMENTS</title>
      <p>The work is supported by EPSRC EP/P00993X/1 and the President’s
PhD Scholarship at Imperial College London.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Alessandro</given-names>
            <surname>Aliberti</surname>
          </string-name>
          , Irene Pupillo, Stefano Terna, Enrico Macii, Santa Di Cataldo, Edoardo Patti, and Andrea Acquaviva, '
          <article-title>A multi-patient data-driven approach to blood glucose prediction'</article-title>
          ,
          <source>IEEE Access</source>
          ,
          <volume>7</volume>
          ,
          <fpage>69311</fpage>
          -
          <lpage>69325</lpage>
          , (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Christoph</given-names>
            <surname>Bergmeir</surname>
          </string-name>
          and Jose´ M Ben´ıtez, '
          <article-title>On the use of cross-validation for time series predictor evaluation'</article-title>
          ,
          <source>Information Sciences</source>
          ,
          <volume>191</volume>
          ,
          <fpage>192</fpage>
          -
          <lpage>213</lpage>
          , (
          <year>2012</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Shiyu</surname>
            <given-names>Chang</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yang</surname>
            <given-names>Zhang</given-names>
          </string-name>
          , Wei Han,
          <string-name>
            <given-names>Mo</given-names>
            <surname>Yu</surname>
          </string-name>
          , Xiaoxiao Guo, Wei Tan,
          <string-name>
            <surname>Xiaodong Cui</surname>
          </string-name>
          , Michael Witbrock,
          <article-title>Mark A Hasegawa-Johnson, and Thomas S Huang, 'Dilated recurrent neural networks'</article-title>
          ,
          <source>in Advances in Neural Information Processing Systems</source>
          , pp.
          <fpage>77</fpage>
          -
          <lpage>87</lpage>
          , (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Zhengping</given-names>
            <surname>Che</surname>
          </string-name>
          , Yu Cheng, Shuangfei Zhai, Zhaonan Sun, and Yan Liu, '
          <article-title>Boosting deep learning risk prediction with generative adversarial networks for electronic health records'</article-title>
          ,
          <source>in 2017 IEEE International Conference on Data Mining (ICDM)</source>
          , pp.
          <fpage>787</fpage>
          -
          <lpage>792</lpage>
          . IEEE, (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Jianwei</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Kezhi</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Pau</given-names>
            <surname>Herrero</surname>
          </string-name>
          , Taiyu Zhu, and Pantelis Georgiou, '
          <article-title>Dilated recurrent neural network for short-time prediction of glucose concentration</article-title>
          .',
          <source>in The 3rd KDH workshop, IJCAI-ECAI</source>
          <year>2018</year>
          , pp.
          <fpage>69</fpage>
          -
          <lpage>73</lpage>
          , (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>Kyunghyun</given-names>
            <surname>Cho</surname>
          </string-name>
          , Bart Van Merrie¨nboer, Dzmitry Bahdanau, and Yoshua Bengio, '
          <article-title>On the properties of neural machine translation: Encoder-decoder approaches'</article-title>
          ,
          <source>arXiv preprint arXiv:1409.1259</source>
          , (
          <year>2014</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Junyoung</given-names>
            <surname>Chung</surname>
          </string-name>
          , Caglar Gulcehre, Kyunghyun Cho, and Yoshua Bengio, '
          <article-title>Empirical evaluation of gated recurrent neural networks on sequence modeling'</article-title>
          ,
          <source>in NIPS 2014 Workshop on Deep Learning</source>
          ,
          <year>December 2014</year>
          , (
          <year>2014</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>William</surname>
            <given-names>L Clarke</given-names>
          </string-name>
          , Daniel Cox,
          <string-name>
            <surname>Linda A Gonder-Frederick</surname>
            ,
            <given-names>William</given-names>
          </string-name>
          <string-name>
            <surname>Carter</surname>
          </string-name>
          , and Stephen L Pohl, '
          <article-title>Evaluating clinical accuracy of systems for self-monitoring of blood glucose', Diabetes care</article-title>
          ,
          <volume>10</volume>
          (
          <issue>5</issue>
          ),
          <fpage>622</fpage>
          -
          <lpage>628</lpage>
          , (
          <year>1987</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9] Cristo´bal Esteban, Stephanie L Hyland,
          <article-title>and Gunnar Ra¨tsch, 'Realvalued (medical) time series generation with recurrent conditional gans'</article-title>
          ,
          <source>arXiv preprint arXiv:1706.02633</source>
          , (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Ian</surname>
            <given-names>Goodfellow</given-names>
          </string-name>
          , Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio, '
          <article-title>Generative adversarial nets'</article-title>
          ,
          <source>in Advances in neural information processing systems</source>
          , pp.
          <fpage>2672</fpage>
          -
          <lpage>2680</lpage>
          , (
          <year>2014</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>Sepp</given-names>
            <surname>Hochreiter</surname>
          </string-name>
          and
          <article-title>Ju¨rgen Schmidhuber, 'Long short-term memory'</article-title>
          ,
          <source>Neural computation</source>
          ,
          <volume>9</volume>
          (
          <issue>8</issue>
          ),
          <fpage>1735</fpage>
          -
          <lpage>1780</lpage>
          , (
          <year>1997</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Elad</surname>
            <given-names>Hoffer</given-names>
          </string-name>
          , Itay Hubara, and Daniel Soudry, '
          <article-title>Train longer, generalize better: closing the generalization gap in large batch training of neural networks'</article-title>
          ,
          <source>in Advances in Neural Information Processing Systems</source>
          , pp.
          <fpage>1731</fpage>
          -
          <lpage>1741</lpage>
          , (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Phillip</surname>
            <given-names>Isola</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jun-Yan</surname>
            <given-names>Zhu</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tinghui Zhou</surname>
          </string-name>
          , and
          <article-title>Alexei A Efros, 'Imageto-image translation with conditional adversarial networks'</article-title>
          ,
          <source>in Proceedings of the IEEE conference on computer vision and pattern recognition</source>
          , pp.
          <fpage>1125</fpage>
          -
          <lpage>1134</lpage>
          , (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Diederik</surname>
            <given-names>P</given-names>
          </string-name>
          <string-name>
            <surname>Kingma and Jimmy Ba</surname>
          </string-name>
          , '
          <article-title>Adam: A method for stochastic optimization</article-title>
          .',
          <source>International Conference on Learning Representations</source>
          <year>2015</year>
          ,
          <fpage>1</fpage>
          -
          <lpage>15</lpage>
          , (
          <year>2015</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Yann</surname>
            <given-names>LeCun</given-names>
          </string-name>
          , Yoshua Bengio, and Geoffrey Hinton, '
          <article-title>Deep learning'</article-title>
          ,
          <source>nature</source>
          ,
          <volume>521</volume>
          (
          <issue>7553</issue>
          ),
          <fpage>436</fpage>
          -
          <lpage>444</lpage>
          , (
          <year>2015</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>Kezhi</given-names>
            <surname>Li</surname>
          </string-name>
          , John Daniels, Chengyuan Liu,
          <article-title>Pau Herrero-Vinas, and Pantelis Georgiou, 'Convolutional recurrent neural networks for glucose prediction'</article-title>
          ,
          <source>IEEE journal of biomedical and health informatics</source>
          , (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>Kezhi</given-names>
            <surname>Li</surname>
          </string-name>
          , Chengyuan Liu, Taiyu Zhu, Pau Herrero, and Pantelis Georgiou, '
          <article-title>Glunet: A deep learning framework for accurate glucose forecasting'</article-title>
          ,
          <source>IEEE journal of biomedical and health informatics</source>
          , (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Chengyuan</surname>
            <given-names>Liu</given-names>
          </string-name>
          , Josep Veh´ı, Parizad Avari, Monika Reddy, Nick Oliver, Pantelis Georgiou, and
          <article-title>Pau Herrero, 'Long-term glucose forecasting using a physiological model and deconvolution of the continuous glucose monitoring signal'</article-title>
          ,
          <source>Sensors</source>
          ,
          <volume>19</volume>
          (
          <issue>19</issue>
          ),
          <volume>4338</volume>
          , (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>Chiara</given-names>
            <surname>Dalla</surname>
          </string-name>
          <string-name>
            <surname>Man</surname>
          </string-name>
          , Francesco Micheletto, Dayu Lv, Marc Breton, Boris Kovatchev, and Claudio Cobelli, '
          <article-title>The uva/padova type 1 diabetes simulator: new features'</article-title>
          ,
          <source>Journal of diabetes science and technology</source>
          ,
          <volume>8</volume>
          (
          <issue>1</issue>
          ),
          <fpage>26</fpage>
          -
          <lpage>34</lpage>
          , (
          <year>2014</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>C.</given-names>
            <surname>Marling</surname>
          </string-name>
          and
          <string-name>
            <given-names>R.</given-names>
            <surname>Bunescu</surname>
          </string-name>
          , '
          <article-title>The OhioT1DM dataset for blood glucose level prediction: Update 2020', in The 5th KDH workshop</article-title>
          ,
          <year>ECAI 2020</year>
          , (
          <year>2020</year>
          ). CEUR proceedings in press, available at http://smarthealth.cs.ohio.edu/bglp/OhioT1DM-dataset-paper.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>Cindy</given-names>
            <surname>Marling and Razvan C. Bunescu</surname>
          </string-name>
          , '
          <article-title>The OhioT1DM dataset for blood glucose level prediction'</article-title>
          ,
          <source>in The 3rd KDH workshop, IJCAIECAI</source>
          <year>2018</year>
          , pp.
          <fpage>60</fpage>
          -
          <lpage>63</lpage>
          , (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <surname>Cooper</surname>
            <given-names>Midroni</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Peter J Leimbigler</surname>
          </string-name>
          , Gaurav Baruah, Maheedhar Kolla,
          <string-name>
            <surname>Alfred J Whitehead</surname>
          </string-name>
          , and Yan Fossat, '
          <article-title>Predicting glycemia in type 1 diabetes patients: experiments with XGBoost'</article-title>
          ,
          <source>in The 3rd KDH workshop, IJCAI-ECAI</source>
          <year>2018</year>
          , pp.
          <fpage>79</fpage>
          -
          <lpage>84</lpage>
          , (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <surname>Sadegh</surname>
            <given-names>Mirshekarian</given-names>
          </string-name>
          , Hui Shen,
          <string-name>
            <given-names>Razvan</given-names>
            <surname>Bunescu</surname>
          </string-name>
          , and Cindy Marling, '
          <article-title>Lstms and neural attention models for blood glucose prediction: Comparative experiments on real and synthetic data'</article-title>
          ,
          <source>in 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC)</source>
          , pp.
          <fpage>706</fpage>
          -
          <lpage>712</lpage>
          . IEEE, (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <surname>Olof</surname>
            <given-names>Mogren</given-names>
          </string-name>
          , '
          <article-title>C-rnn-gan: Continuous recurrent neural networks with adversarial training'</article-title>
          ,
          <source>arXiv preprint arXiv:1611.09904</source>
          , (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <surname>Silvia</surname>
            <given-names>Oviedo</given-names>
          </string-name>
          , Josep Veh´ı, Remei Calm, and Joaquim Armengol, '
          <article-title>A review of personalized blood glucose prediction strategies for t1dm patients'</article-title>
          ,
          <source>International journal for numerical methods in biomedical engineering</source>
          ,
          <volume>33</volume>
          (
          <issue>6</issue>
          ),
          <year>e2833</year>
          , (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>Tom</given-names>
            <surname>Le</surname>
          </string-name>
          <string-name>
            <given-names>Paine</given-names>
            , Pooya Khorrami,
            <surname>Shiyu</surname>
          </string-name>
          <string-name>
            <given-names>Chang</given-names>
            ,
            <surname>Yang</surname>
          </string-name>
          <string-name>
            <given-names>Zhang</given-names>
            , Prajit Ramachandran,
            <surname>Mark A Hasegawa-Johnson</surname>
          </string-name>
          , and Thomas S Huang, '
          <article-title>Fast wavenet generation algorithm'</article-title>
          ,
          <source>arXiv preprint arXiv:1611.09482</source>
          , (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <surname>Pouya</surname>
            <given-names>Saeedi</given-names>
          </string-name>
          , Inga Petersohn, Paraskevi Salpea, Belma Malanda, Suvi Karuranga, Nigel Unwin, Stephen Colagiuri, Leonor Guariguata, Ayesha A Motala,
          <string-name>
            <given-names>Katherine</given-names>
            <surname>Ogurtsova</surname>
          </string-name>
          , et al.,
          <article-title>'Global and regional diabetes prevalence estimates for 2019 and projections for 2030 and 2045: Results from the international diabetes federation diabetes atlas'</article-title>
          ,
          <source>Diabetes research and clinical practice</source>
          ,
          <volume>157</volume>
          ,
          <fpage>107843</fpage>
          , (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <surname>Qingnan</surname>
            <given-names>Sun</given-names>
          </string-name>
          , Marko V Jankovic,
          <string-name>
            <given-names>Lia</given-names>
            <surname>Bally</surname>
          </string-name>
          , and Stavroula G Mougiakakou,
          <article-title>'Predicting blood glucose with an LSTM and Bi-LSTM based deep neural network'</article-title>
          ,
          <source>in 2018 14th Symposium on Neural Networks and Applications (NEUREL)</source>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>5</lpage>
          . IEEE, (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>Ashenafi</given-names>
            <surname>Zebene</surname>
          </string-name>
          <string-name>
            <given-names>Woldaregay</given-names>
            ,
            <surname>Eirik</surname>
          </string-name>
          <string-name>
            <surname>A</surname>
          </string-name>
          ˚rsand, Sta˚le Walderhaug, David Albers,
          <string-name>
            <given-names>Lena</given-names>
            <surname>Mamykina</surname>
          </string-name>
          , Taxiarchis Botsis, and Gunnar Hartvigsen, '
          <article-title>Data-driven modeling and prediction of blood glucose dynamics: Machine learning applications in type 1 diabetes'</article-title>
          ,
          <source>Artificial intelligence in medicine</source>
          , (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <surname>Jinsung</surname>
            <given-names>Yoon</given-names>
          </string-name>
          , Daniel Jarrett, and Mihaela van der Schaar, '
          <article-title>Time-series generative adversarial networks'</article-title>
          ,
          <source>in Advances in Neural Information Processing Systems</source>
          , pp.
          <fpage>5509</fpage>
          -
          <lpage>5519</lpage>
          , (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <surname>Xingyu</surname>
            <given-names>Zhou</given-names>
          </string-name>
          , Zhisong Pan, Guyu Hu,
          <string-name>
            <given-names>Siqi</given-names>
            <surname>Tang</surname>
          </string-name>
          , and Cheng Zhao, '
          <article-title>Stock market prediction on high-frequency data using generative adversarial nets'</article-title>
          , Mathematical Problems in Engineering,
          <year>2018</year>
          , (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [32]
          <string-name>
            <surname>Taiyu</surname>
            <given-names>Zhu</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Kezhi</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Jianwei</given-names>
            <surname>Chen</surname>
          </string-name>
          , Pau Herrero, and Pantelis Georgiou, '
          <article-title>Dilated recurrent neural networks for glucose forecasting in type 1 diabetes'</article-title>
          ,
          <source>Journal of Healthcare Informatics Research</source>
          ,
          <fpage>1</fpage>
          -
          <lpage>17</lpage>
          , (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          [33]
          <string-name>
            <surname>Taiyu</surname>
            <given-names>Zhu</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Kezhi</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Pau</given-names>
            <surname>Herrero</surname>
          </string-name>
          , Jianwei Chen, and Pantelis Georgiou, '
          <article-title>A deep learning algorithm for personalized blood glucose prediction</article-title>
          .',
          <source>in The 3rd KDH workshop, IJCAI-ECAI</source>
          <year>2018</year>
          , pp.
          <fpage>64</fpage>
          -
          <lpage>78</lpage>
          , (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>