<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Blood Glucose Level Prediction as Time-Series Modeling using Sequence-to-Sequence Neural Networks</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Ananth</forename><surname>Bhimireddy</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Indiana University Purdue University Indianapolis</orgName>
								<address>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Priyanshu</forename><surname>Sinha</surname></persName>
							<email>inha@outlook.com</email>
							<affiliation key="aff1">
								<orgName type="institution">Mentor Graphics India Pvt. Ltd</orgName>
								<address>
									<country key="IN">India</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Bolu</forename><surname>Oluwalade</surname></persName>
							<email>boluwala@iupui.edu</email>
							<affiliation key="aff2">
								<orgName type="institution">Indiana University Purdue University Indianapolis</orgName>
								<address>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Judy</forename><forename type="middle">Wawira</forename><surname>Gichoya</surname></persName>
							<affiliation key="aff3">
								<orgName type="institution">Emory University School of Medicine</orgName>
								<address>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Saptarshi</forename><surname>Purkayastha</surname></persName>
							<affiliation key="aff4">
								<orgName type="institution">Indiana University Purdue University Indianapolis</orgName>
								<address>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Blood Glucose Level Prediction as Time-Series Modeling using Sequence-to-Sequence Neural Networks</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">7916D63B7EAAE15CDE89B84BFAC9746C</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-25T00:23+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Blood glucose prediction</term>
					<term>Time-series model</term>
					<term>Wearable devices</term>
					<term>Transfer learning</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The management of blood glucose levels is critical in the care of Type 1 diabetes subjects. In extremes, high or low levels of blood glucose are fatal. To avoid such adverse events, there is the development and adoption of wearable technologies that continuously monitor blood glucose and administer insulin. This technology allows subjects to easily track their blood glucose levels with early intervention, preventing the need for hospital visits. The data collected from these sensors is an excellent candidate for the application of machine learning algorithms to learn patterns and predict future values of blood glucose levels. In this study, we developed artificial neural network algorithms based on the OhioT1DM training dataset that contains data on 12 subjects. The dataset contains features such as subject identifiers, continuous glucose monitoring data obtained in 5 minutes intervals, insulin infusion rate, etc. We developed individual models, including LSTM, BiLSTM, Convolutional LSTMs, TCN, and sequence-to-sequence models. We also developed transfer learning models based on the most important features of the data, as identified by a gradient boosting algorithm. These models were evaluated on the OhioT1DM test dataset that contains 6 unique subject's data. The model with the lowest RMSE values for the 30-and 60minutes was selected as the best performing model. Our result shows that sequence-to-sequence BiLSTM performed better than the other models. This work demonstrates the potential of artificial neural networks algorithms in the management of Type 1 diabetes.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">INTRODUCTION</head><p>Diabetes Mellitus is a chronic disease characterized by high blood glucose levels. According to the 2020 CDC National Diabetes Statistics Report, about 34.2 million people or 10.2 percent of Americans have diabetes <ref type="bibr" target="#b2">[3]</ref>. Diabetes is classified as Type 1, Type 2 and Gestational Diabetes. Insulin is an enzyme produced in the pancreas that helps in blood glucose absorption into cells. In Type 1 diabetes (T1DM), the pancreas produces little or no insulin. In contrast, in Type 2 diabetes (T2DM), the pancreas produces a small amount of insulin, or the body is resistant to the effect of insulin. This absence or resistance to insulin leads to an increase in blood glucose levels, known as hyperglycemia. Symptoms of hyperglycemia include excessive thirst, excessive urination, sweating, etc. Diabetic ketoacidosis is a serious complication of uncontrolled hyperglycemia that can lead to death. On the other hand, elevated insulin levels in the body can cause low levels of blood glucose, a condition known as hypoglycemia. Dizziness, weakness, coma, or eventually, death can occur in uncontrolled hypoglycemia. Insulin and glucose control are critical to the management of diabetes, and hence titration of the administered insulin doses is critical in management of diabetes patients.</p><p>Glucose levels vary according to the patient's diet, and activities throughout the day. Sensors have been developed to estimate blood glucose levels at various time intervals. These sensors are useful in diabetes management because they provide longitudinal data about subjects' blood glucose and show the distinctive patterns throughout the day. These sensors are frequently coupled with the use of insulin pumps to deliver short-acting insulin continuously (basal rate) and specific insulin quantity after a meal for appropriate glycemic control. Although the sensors and insulin pumps have helped to improve patient care, patients are typically unaware of an impending adverse event of severe hyperglycemia or hypoglycemia. These adverse effects commonly occur when patients are asleep. There is an opportunity for the development of accurate prediction models using previously collected sensor data to estimate future values of blood glucose levels to prevent the occurrence of adverse events.</p><p>In this study, we utilize the OhioT1DM dataset, which contains blood glucose values of twelve T1DM subjects collected at intervals over a total time span of eight weeks <ref type="bibr" target="#b12">[13]</ref>. These individuals had an insulin pump with continuous glucose monitoring (CGM), wearing a physical activity band and self-reporting life events using a smartphone application. CGM blood glucose data were obtained at 5-minute intervals <ref type="bibr" target="#b11">[12]</ref>. We developed multiple models for predicting glucose values at 30 and 60 minutes in the future, using the CGM values, and mean Root Mean Square Error (RMSE) as an evaluation metric. The code for this study can be found at https://github.com/iupui-soic/bglp2</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">RELATED WORK</head><p>Models like LSTM and RNN have been used for forecasting in <ref type="bibr" target="#b9">[10]</ref> which has been improved since then. The paper explores short-term load forecasting for individual electric customers and proposes an LSTM based framework to tackle the issue. Machine Learning models such as XGBoost have been used to predict glycemia in type-1 diabetic patients in <ref type="bibr" target="#b13">[14]</ref>. This paper experiments primarily with XGBoost algorithm to predict the blood glucose levels at a 30minute horizon in the OhioT1DM dataset. Features from pre-trained TimeNets have been used for clinical predictions in <ref type="bibr" target="#b6">[7]</ref>. This paper uses pre-training a network for some supervised or unsupervised tasks on datasets, and then fine-tuning via transfer learning for a related end-task to leverage the resources of labeled data in making predictions. This paper points out that training deep learning models such as RNNs and LSTMs requires large labeled data and is computationally expensive.</p><p>Deep learning models like Recurrent Neural Network (RNN) have been used on the OhioTIDM dataset to predict future blood glucose values <ref type="bibr" target="#b15">[16]</ref>, including the BGLP Challenge at KDH@IJCAI-ECAI 2018 (http://ceur-ws.org/Vol-2148/). In some cases, these data-driven models use only the CGM values or use physiological data such as the insulin concentration, amount of carbohydrate in meals and physical activities. Chen et al. created a data-driven 3-layer dilated recurrent neural network model with a mean RMSE of 18.9, with a range of 15.2995 and 22.7104 <ref type="bibr" target="#b3">[4]</ref>. They concluded that the missing data and the continuous fluctuations in the data influenced the model's performance. Their model bettered the Convolutional Neural Network (CNN) that gave an average RMSE of 21.726 for six subjects <ref type="bibr" target="#b20">[21]</ref>. Bertachi et al. predicted blood glucose levels using Artificial Neural Networks with the inclusion of physiological data <ref type="bibr" target="#b0">[1]</ref>. Their results were not significantly dissimilar to those obtained in the data-driven models. All the previous studies demonstrate that a lower RMSE is obtained at the 30-minute prediction when compared to the 60-minutes values. We postulate that a hybrid approach of combining both the data-driven and physiological models could improve on the performance of the individual models and incorporate this in our approach.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">METHODS</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Dataset Description</head><p>A detailed description of the dataset has been previously published in the OhioT1DM dataset paper <ref type="bibr" target="#b11">[12]</ref>. We used the data provided on 12 subjects for training, and 6 test subjects. Furthermore, the parameters basal and temp basal were merged into a single parameter.</p><p>We converted both the training and testing datasets from XML to CSV, preserving the time intervals. We did not use interpolation on the datasets as rules of the competition have prohibited interpolation. We tried using the forward and backward filling to fill the null values in the datasets, but they are creating additional time intervals which becomes a problem in testing datasets. So, no re-sampling technique was used in this paper to preserve the time intervals of the samples. The data pre-processing is performed on only 6 patients (test and train datasets) whose results are to be predicted. Subject-548 has the highest number of training records (12150) and Subject-552 has the lowest no. of records(9080). The no.of features varied from subject to subject which causes unevenness while training time series models. So, we have added features to subjects as required to ease the process of training and predicting. Missing data is handled in all columns by inputting the null values with zeros(0). We did not disturb the glucose value column as required by the competition.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Description of ML models</head><p>We used the following deep learning models to predict the blood glucose levels of each subject. The data pre-processing and model development are summarized in figure <ref type="figure" target="#fig_0">1</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.1">Long Short-Term Memory (LSTM) Networks</head><p>LSTMs were originally introduced by Hochreiter and Schmidhuber <ref type="bibr" target="#b7">[8]</ref> and later refined and popularized <ref type="bibr" target="#b16">[17]</ref>  <ref type="bibr" target="#b19">[20]</ref>. LSTMs are a special kind of RNNs, capable of learning long-term dependencies. This quality of LSTMs helps memorize useful parts of the sequence and the model learns parameters more efficiently, making it useful for time series models.</p><p>We trained two models using LSTMs, one with 5-min interval data and another with 30-min interval data. In each model, we used all the available features at time t to predict the glucose value at time t+1. Before fitting the dataset, we scaled the dataset using MinMaxScaler from scikit-learn <ref type="bibr" target="#b14">[15]</ref>.</p><p>The LSTM model was built using the Keras <ref type="bibr" target="#b5">[6]</ref> platform. We used 128 LSTM units, followed by a dense layer (150 units), dropout layer (0.20), dense layer (100 units), dropout layer (0.15), dense layer (50 units), dense layer (20 units) and a final layer with one unit (for prediction). We used ReLU as the activation function with Adam optimizer. The loss was calculated in mean squared error (MSE) and later converted into Root Mean Squared Error (RMSE). The model was trained for 200 epochs with a batch size of 32. The results of the model for each subject are provided in the results table.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.2">Bi-directional Long Short-Term Memory (BiLSTM) Networks</head><p>As our input text data is static, and the entire sequence is available at the same time, we implemented a BiLSTM model to observe how processing the sequence from either direction affected the accuracy.</p><p>The architecture of the BiLSTM model is similar to the LSTM model and is thus useful for time series prediction.</p><p>The data processing and model parameters for BiLSTM and LSTM model were similar with an exception in the model's first layer, where the scaled data was inputted into the Bidirectional LSTM with 128 units. The model was trained for 200 epochs with a batch size of 32. The results of the model for each subject are provided in the results table.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.3">Temporal Convolutional Networks (TCN)</head><p>TCNs were originally introduced by Lea and Videl <ref type="bibr" target="#b10">[11]</ref> in 2016. TCNs are extremely useful in capturing the high-level temporal relationships in sequential networks. The TCN architecture allows capturing long-range spatio-temporal relationships. TCN's help capture the blood glucose level of subjects who usually have routine lifestyles, as TCNs can capture hierarchical relationships at low, intermediate, and high time scales.</p><p>The data processing steps are similar to the LSTM model. The TCN model was built using the Keras platform, but the depth of the model is relatively simpler compared to the LSTM and BiLSTM model. The scaled data was inputted into a TCN layer and then connected to a dense layer with one unit for the output. The model used Adam optimizer and MSE for calculating loss which was later converted to RMSE. The model was trained for 10 epochs and the obtained results are provided in the results table.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.4">Convolutional LSTM</head><p>Convolutional LSTMs (ConvLSTM) were introduced by Xingjian Shi et al. <ref type="bibr" target="#b17">[18]</ref> in 2015. ConvLSTMs are created by extending the fully connected LSTM to have convolutional structure in both the inputto-state and state-to-state transitions. ConvLSTM network captures </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.5">Description of Sequence-to-Sequence Models</head><p>Sequence to sequence models were first introduced by Google in 2014 <ref type="bibr" target="#b18">[19]</ref>. These models map fixed-length input with fixed-length output where length of input and output differ. Sequence-to-sequence models consist of three parts:</p><p>• Encoder: Encoder consists of stacks of several recurrent (LSTM) units where each unit takes a single element of the input sequence, extracts information from it and propagates it to the next unit. • Encoded Vector: This is the intermediate step and the final hidden layer of the encoder. It also acts as the first hidden layer for the decoder to make predictions. This vector encapsulates all information from all input samples and provides this information to the decoder. • Decoder: This consists of stacks of recurrent unit where each unit predicts output at time step t. Each unit accepts a hidden layer as input and produces output as well as its own hidden state.</p><p>The main advantage of this architecture is that it can map sequences of different lengths to each other. We applied 3 variants of sequence to sequence models viz., sequence-to-sequence LSTM, sequence-to-sequence Bi-LSTM, and sequence-to-sequence CNN-LSTM.</p><p>For training of sequence-to-sequence models, we split the data into windows of 60 minutes. This approach is intuitive and helpful as BGLP values can be predicted 1-hour ahead. It is also helpful while modelling as the model can be used to predict blood glucose values at specific time periods, let's say (after 10 minutes) or a whole sequence of blood glucose values.</p><p>To evaluate these sequence-to-sequence models we used walk forward validation. Here the model predicts the next one hour and then the actual data for one hour is given to make prediction for next one hour. See below table 1 for more illustration: For our training, we kept our input size (number of prior observations required to make next predictions) as 30 minutes data to predict next 60 minutes data. Each sequence-to-sequence model used in our work is described below:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Sequence-to-Sequence LSTM</head><p>In this model, we used 200 LSTM cells for the encoder and decoder. This layer was followed with 2 dense layers containing 150 and 1 units wrapped in a TimeDistributed layer. The model was trained for 80 epochs with batch size of 40. We used Adam optimizer <ref type="bibr" target="#b8">[9]</ref> with learning rate of 0.01 and loss function as MSE.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Sequence-to-Sequence Bi-LSTM</head><p>In this model, we used 100 Bi-LSTM cells (LSTM cells wrapped in Bidirectional wrapper) for the encoder and decoder. This layer was followed with 2 dense layers containing 150 and 1 units wrapped in a TimeDistributed layer. The model was trained for 80 epochs with batch size of 40. Similar to sequence-to-sequence LSTM, Adam optimizer was used with 0.01 learning rate and "mean squared error" as the loss function.</p><p>3. Sequence-to-Sequence CNN-LSTM In this model, we used 2 1D Convolutional layer with filter size of 128 and 64 respectively. Convolutional layers were followed with MaxPooling 1D layer and flatten layer for the encoder. 200 LSTM cells are used for the decoder. This layer was followed with 2 dense layers containing 100 and 1 units wrapped in a TimeDistributed layer. The model was trained for 80 epochs with batch size of 40. Similar to above models, Adam optimizer was used with 0.01 learning rate and "mean squared error" as the loss function.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.6">Transfer Learning</head><p>For transfer learning, we first found the most relevant and common features among all 12 subjects. The importance of features was found using a Gradient Boosting algorithm. We set the cumulative frequency to 0.99 for feature selection and the 5 most important and common features are as follows: 'finger stick value', 'basal rate value', 'galvanic skin response value', 'skin temperature value', 'bolus dose value' Using the above-identified important features only, we trained our model on a randomly selected subject (567) and subsequently, finetuned each sequence-to-sequence model on each subject. The final model was used for prediction on the test data. The configurations of each model was similar to the sequence-to-sequence model as described above.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">RESULTS</head><p>Figures <ref type="figure">2 and 3</ref> shows the comparison between the actual and the predicted values obtained from our best performing model i.e. sequenceto-sequence BiLSTM model. From the figures, it is clear that our model closely predicts the values of the test data by following similar peaks and troughs. The RMSE of the 30 minutes horizon predictions of the models are presented in tables 3, 4 and 5. From the tables, the sequenceto-sequence (Seq2Seq) models performed better than all the other models with an average RMSE of 23.5 for Seq-2-Seq LSTM, 21.8 for Seq-2-Seq BiLSTM, and 23.0 for Seq-2-Seq CNN-LSTM. Table 2 describes the RMSE and MAE values for the sequence-tosequence BiLSTM model which performed better than the individual and transfer learning models.</p><p>From Table <ref type="table" target="#tab_1">2</ref>, the value for the RMSE varies from 18.2 in subject 552 to 29.7 in subject 584 at 30 minutes, and between 28.3 in subject 596 to 42.6 in subject 584 at 60 minutes. The MAE values lies between 12.7 to 18.1 at 30 minutes and between 19.3 and 31.0 at 60 minutes.</p><p>Subject 584 in figure <ref type="figure">3</ref> shows more fluctuations, has a higher peak and lower troughs than subject 552 in figure <ref type="figure">2</ref>. The effect of these variations are reflected in our results as patient 584 has the highest RMSE value while subject 552 has the lowest RMSE values. It is evident that the levels of variations in the individual subjects contributes significantly to the differences in RMSE values of the individual subjects in our model.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">CONCLUSION AND FUTURE SCOPE</head><p>In this paper, we present results of application of deep learning models to make predictions of blood glucose values. Potential benefits such as the prevention of adverse events associated extreme glucose values serve as a source of motivation for these efforts. Overall, sequence-to-sequence models especially Bi-LSTM have the best performance as these models are best at mapping sequences irrespective of their lengths. Our performance is affected by fluctuations in glucose values and also with missing data as described in previous experiments. Given the overall success of transfer learning, we also evaluated the potential of single model prediction via transfer learning approach. The transfer learning approach was inferior to the sequence to sequence models. Compared to the previous paper for BGLP Challenege, we observed that two papers <ref type="bibr" target="#b1">[2]</ref> and <ref type="bibr" target="#b4">[5]</ref> have better results than our pro-posed results. But <ref type="bibr" target="#b4">[5]</ref> have used interpolation as a part of data processing which is against the rules of the competition and <ref type="bibr" target="#b1">[2]</ref> did not mention details about data processing. Our future work will be to improve the transfer learning model as we are provided with more common features among all subjects, so that we can create a generic model for predicting blood glucose levels. However, the development of a generic model can be challenging because of confounding factors such as variations in sensor types, lifestyles, physiology and genetics. It is therefore pertinent that these factors are considered in future endeavors.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 .</head><label>1</label><figDesc>Figure 1. Process Architecture spatio-temporal correlations better and usually outperform Fully Connected LSTM networks. The scaled data was reshaped and inputted to a convolution layer with 32 filters of kernel size 1, followed by LSTM layer with 128 units, Dense layer with 150 units, Dropout layer with 0.2 dropout rate, Dense layer with 100 units, Dropout layer with 0.15 rate, Dense layer with 50 units, Dense layer with 16 units and finally a Dense layer with 1 unit for prediction. ReLU activation function was used in all layers. The model used Adam optimizer and the MSE loss was further converted to RMSE for model comparison. The model was trained for 200 epochs with a batch size of 32. The obtained results are provided in the results table.</figDesc><graphic coords="3,92.92,71.12,406.47,192.96" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 .Figure 3 .</head><label>23</label><figDesc>Figure 2. Actual and Predicted values of subject 552 at 30 minutes horizons</figDesc><graphic coords="5,67.52,71.12,457.22,213.37" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 .</head><label>1</label><figDesc>Description of Sequence-to-Sequence input and prediction values</figDesc><table><row><cell>Input</cell><cell>Prediction</cell></row><row><cell>1st 60 minutes data</cell><cell>2nd 60 minutes data</cell></row><row><cell>[1st + 2nd] 60 minutes data</cell><cell>3rd 60 minutes data</cell></row><row><cell>[1st + 2nd + 3rd] 60 minutes data</cell><cell>4th 60 minutes data</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 .</head><label>2</label><figDesc>RMSE and MAE results of Sequence-to-Sequence BiLSTM model</figDesc><table><row><cell></cell><cell>RMSE</cell><cell></cell><cell>MAE</cell><cell></cell></row><row><cell>Subjects</cell><cell cols="4">30 Mins 60 Mins 30 Mins 60 Mins</cell></row><row><cell>584</cell><cell>29.7</cell><cell>42.6</cell><cell>18.1</cell><cell>30.0</cell></row><row><cell>567</cell><cell>20.7</cell><cell>35.1</cell><cell>14.4</cell><cell>24.9</cell></row><row><cell>596</cell><cell>18.6</cell><cell>28.3</cell><cell>12.7</cell><cell>19.3</cell></row><row><cell>552</cell><cell>18.2</cell><cell>30.0</cell><cell>13.3</cell><cell>22.2</cell></row><row><cell>544</cell><cell>19.8</cell><cell>32.9</cell><cell>13.7</cell><cell>23.1</cell></row><row><cell>540</cell><cell>24.3</cell><cell>41.4</cell><cell>17.8</cell><cell>31.0</cell></row><row><cell>Mean</cell><cell>21.8</cell><cell>35.0</cell><cell>15.0</cell><cell>25.0</cell></row><row><cell>SD</cell><cell>4.0</cell><cell>5.4</cell><cell>2.1</cell><cell>4.2</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3 .</head><label>3</label><figDesc>RMSE values for individual models at 30 minutes Horizon</figDesc><table><row><cell cols="3">Subjects LSTM BiLSTM</cell><cell>TCN</cell><cell>ConvLSTM</cell></row><row><cell>584</cell><cell>27.97</cell><cell>29.06</cell><cell>26.55</cell><cell>26.57</cell></row><row><cell>567</cell><cell>25.65</cell><cell>27.04</cell><cell>25.71</cell><cell>26.04</cell></row><row><cell>596</cell><cell>19.47</cell><cell>20.30</cell><cell>18.95</cell><cell>21.08</cell></row><row><cell>552</cell><cell>20.62</cell><cell>20.30</cell><cell>17.14</cell><cell>17.73</cell></row><row><cell>544</cell><cell>21.34</cell><cell>22.06</cell><cell>80.67</cell><cell>20.94</cell></row><row><cell>540</cell><cell>30.21</cell><cell>31.72</cell><cell>25.94</cell><cell>27.35</cell></row><row><cell>Mean</cell><cell>25.0</cell><cell>24.4</cell><cell>34.7</cell><cell>23.2</cell></row><row><cell>SD</cell><cell>3.99</cell><cell>4.98</cell><cell>21.13</cell><cell>3.43</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4 .</head><label>4</label><figDesc>RMSE values for sequence-to-sequence models at 30 Minutes horizon</figDesc><table><row><cell></cell><cell cols="2">Seq-2-Seq Seq-2-Seq</cell><cell>Seq-2-Seq</cell></row><row><cell>Subjects</cell><cell>LSTM</cell><cell>BiLSTM</cell><cell>CNN-LSTM</cell></row><row><cell>567</cell><cell>29.2</cell><cell>20.7</cell><cell>29.0</cell></row><row><cell>540</cell><cell>25.0</cell><cell>24.3</cell><cell>23.1</cell></row><row><cell>544</cell><cell>18.5</cell><cell>19.8</cell><cell>19.2</cell></row><row><cell>596</cell><cell>18.6</cell><cell>18.6</cell><cell>19.0</cell></row><row><cell>584</cell><cell>30.7</cell><cell>29.7</cell><cell>31.0</cell></row><row><cell>552</cell><cell>19.2</cell><cell>18.1</cell><cell>17.2</cell></row><row><cell>Mean</cell><cell>23.5</cell><cell>21.8</cell><cell>23.0</cell></row><row><cell>SD</cell><cell>5.0</cell><cell>4.0</cell><cell>5.2</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 5 .</head><label>5</label><figDesc>RMSE values for sequence to Sequence transfer learning model at 30 Minutes horizon</figDesc><table><row><cell></cell><cell cols="2">Seq-2-Seq Seq-2-Seq</cell><cell>Seq-2-Seq</cell></row><row><cell>Subjects</cell><cell>LSTM</cell><cell>BiLSTM</cell><cell>CNN-LSTM</cell></row><row><cell>567</cell><cell>33.5</cell><cell>30.8</cell><cell>32.6</cell></row><row><cell>540</cell><cell>39.7</cell><cell>32.7</cell><cell>44.2</cell></row><row><cell>544</cell><cell>23.1</cell><cell>31.5</cell><cell>21.8</cell></row><row><cell>596</cell><cell>19.2</cell><cell>18.0</cell><cell>17.9</cell></row><row><cell>584</cell><cell>36.8</cell><cell>37.3</cell><cell>38.1</cell></row><row><cell>552</cell><cell>21.8</cell><cell>14.3</cell><cell>17.7</cell></row><row><cell>Mean</cell><cell>29.0</cell><cell>27.4</cell><cell>28.7</cell></row><row><cell>SD</cell><cell>7.9</cell><cell>8.3</cell><cell>10.2</cell></row></table></figure>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Prediction of blood glucose levels and nocturnal hypoglycemia using physiological models and artificial neural networks</title>
		<author>
			<persName><forename type="first">Arthur</forename><surname>Bertachi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lyvia</forename><surname>Biagi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Iván</forename><surname>Contreras</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ningsu</forename><surname>Luo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Josep</forename><surname>Vehí</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">KHD@ IJCAI</title>
				<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="85" to="90" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Prediction of blood glucose levels and nocturnal hypoglycemia using physiological models and artificial neural networks</title>
		<author>
			<persName><forename type="first">Arthur</forename><surname>Bertachi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lyvia</forename><surname>Biagi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Iván</forename><surname>Contreras</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ningsu</forename><surname>Luo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Josep</forename><surname>Vehí</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">KHD@IJCAI</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<title level="m">Estimates of diabetes and its burden in the United States</title>
				<imprint>
			<date type="published" when="2020">2020. 2020</date>
		</imprint>
		<respStmt>
			<orgName>CDC ; National Diabetes Statistics Report</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Dilated recurrent neural network for short-time prediction of glucose concentration</title>
		<author>
			<persName><forename type="first">Jianwei</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kezhi</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Pau</forename><surname>Herrero</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Taiyu</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Pantelis</forename><surname>Georgiou</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">KHD@ IJCAI</title>
				<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="69" to="73" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Dilated recurrent neural network for short-time prediction of glucose concentration</title>
		<author>
			<persName><forename type="first">Jianwei</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kezhi</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Pau</forename><surname>Herrero</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Taiyu</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Pantelis</forename><surname>Georgiou</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">KHD@IJCAI</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<title/>
		<author>
			<persName><forename type="first">Chollet</forename><surname>Franc</surname></persName>
		</author>
		<ptr target="https://keras.io" />
		<imprint>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title level="m" type="main">Using features from pre-trained timenet for clinical predictions</title>
		<author>
			<persName><forename type="first">Priyanka</forename><surname>Gupta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Pankaj</forename><surname>Malhotra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lovekesh</forename><surname>Vig</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Gautam</forename><surname>Shroff</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2018-07">07. 2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Long short-term memory</title>
		<author>
			<persName><forename type="first">Sepp</forename><surname>Hochreiter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jürgen</forename><surname>Schmidhuber</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Neural Comput</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="issue">8</biblScope>
			<biblScope unit="page" from="1735" to="1780" />
			<date type="published" when="1997-11">November 1997</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<title level="m" type="main">Adam: A method for stochastic optimization</title>
		<author>
			<persName><forename type="first">P</forename><surname>Diederik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jimmy</forename><surname>Kingma</surname></persName>
		</author>
		<author>
			<persName><surname>Ba</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Short-term residential load forecasting based on lstm recurrent neural network</title>
		<author>
			<persName><forename type="first">W</forename><surname>Kong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><forename type="middle">Y</forename><surname>Dong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Jia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">J</forename><surname>Hill</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Smart Grid</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="841" to="851" />
			<date type="published" when="2019-01">Jan 2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<title level="m" type="main">Temporal convolutional networks: A unified approach to action segmentation</title>
		<author>
			<persName><forename type="first">Colin</forename><surname>Lea</surname></persName>
		</author>
		<author>
			<persName><forename type="first">René</forename><surname>Vidal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Austin</forename><surname>Reiter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Gregory</forename><forename type="middle">D</forename><surname>Hager</surname></persName>
		</author>
		<idno>CoRR, abs/1608.08242</idno>
		<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title level="m" type="main">The ohiot1dm dataset for blood glucose level prediction: Update</title>
		<author>
			<persName><forename type="first">Cindy</forename><surname>Marling</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Razvan</forename><surname>Bunescu</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">The ohiot1dm dataset for blood glucose level prediction</title>
		<author>
			<persName><forename type="first">Cindy</forename><surname>Marling</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Razvan</forename><forename type="middle">C</forename><surname>Bunescu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">KHD@ IJCAI</title>
				<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="60" to="63" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<title level="m" type="main">Predicting glycemia in type 1 diabetes patients: Experiments with xgboost</title>
		<author>
			<persName><forename type="first">Cooper</forename><surname>Midroni</surname></persName>
		</author>
		<author>
			<persName><surname>Peter Leimbigler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yan</forename><surname>Gaurav</surname></persName>
		</author>
		<author>
			<persName><surname>Fossat</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2018-07">07. 2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Scikit-learn: Machine learning in Python</title>
		<author>
			<persName><forename type="first">F</forename><surname>Pedregosa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Varoquaux</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gramfort</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Michel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Thirion</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Grisel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Blondel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Prettenhofer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Weiss</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Dubourg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Vanderplas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Passos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Cournapeau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Brucher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Perrot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Duchesnay</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Machine Learning Research</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="page" from="2825" to="2830" />
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Deep learning-based data imputation on time-variant data using recurrent neural network</title>
		<author>
			<persName><forename type="first">M</forename><surname>Sangeetha</surname></persName>
		</author>
		<author>
			<persName><surname>Senthil Kumaran</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Soft Computing</title>
		<imprint>
			<biblScope unit="page" from="1" to="12" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<title level="m" type="main">Fundamentals of recurrent neural network (RNN) and long short-term memory (LSTM) network</title>
		<author>
			<persName><forename type="first">Alex</forename><surname>Sherstinsky</surname></persName>
		</author>
		<idno>CoRR, abs/1808.03314</idno>
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<title level="m" type="main">Convolutional LSTM network: A machine learning approach for precipitation nowcasting</title>
		<author>
			<persName><forename type="first">Xingjian</forename><surname>Shi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Zhourong</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Hao</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Dit-Yan</forename><surname>Yeung</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Wai-Kin</forename><surname>Wong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Wang-Chun</forename><surname>Woo</surname></persName>
		</author>
		<idno>CoRR, abs/1506.04214</idno>
		<imprint>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Sequence to sequence learning with neural networks</title>
		<author>
			<persName><forename type="first">Ilya</forename><surname>Sutskever</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Oriol</forename><surname>Vinyals</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Quoc V</forename><surname>Le</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in neural information processing systems</title>
				<imprint>
			<date type="published" when="2014">2014</date>
			<biblScope unit="page" from="3104" to="3112" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Text classification improved by integrating bidirectional LSTM with two-dimensional max pooling</title>
		<author>
			<persName><forename type="first">Peng</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Zhenyu</forename><surname>Qi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Suncong</forename><surname>Zheng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jiaming</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Hongyun</forename><surname>Bao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Bo</forename><surname>Xu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers</title>
				<meeting>COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers<address><addrLine>Osaka, Japan</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2016-12">December 2016</date>
			<biblScope unit="page" from="3485" to="3495" />
		</imprint>
	</monogr>
	<note>The COLING 2016 Organizing Committee</note>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">A deep learning algorithm for personalized blood glucose prediction</title>
		<author>
			<persName><forename type="first">Taiyu</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kezhi</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Pau</forename><surname>Herrero</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jianwei</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Pantelis</forename><surname>Georgiou</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">KHD@ IJCAI</title>
				<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="64" to="78" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
