<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Multi-lag Stacking for Blood Glucose Level Prediction</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Heydar Khadem</string-name>
          <email>h.khadem@shef</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mohammed Benaissa</string-name>
          <email>m.benaissa@shef</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Electronic and Electrical Engineering</institution>
          ,
          <addr-line>Univer-</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Department of Oncology and Metabolism, University of Sheffield</institution>
          ,
          <country country="UK">UK</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>This work investigates blood glucose level prediction for type 1 diabetes in two horizons of 30 and 60 minutes. Initially, three conventional regression tools-partial least square regression (PLSR), multilayer perceptron, and long short-term memory-are deployed to create predictive models. They are trained once on 30 minutes and once on 60 minutes of historical data resulting in six basic models for each prediction horizon. A collection of these models are then set as base-learners to develop three stacking systems; two uni-lag and one multi-lag. One of the uni-lag systems uses the three basic models trained on 30 minutes of lag data; the other uses those trained on 60 minutes. The multi-lag system, on the other hand, leverages the basic models trained on both lags. All three stacking systems deploy a PLSR as meta-learner. The results obtained show: i) the stacking systems outperform the basic models, ii) among the stacking systems, the multi-lag shows the best predictive performance with a root mean square error of 19.01 mg/dl and 33.37 mg/dl for the prediction horizon of 30 and 60 minutes, respectively.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>INTRODUCTION</title>
      <p>
        Diabetes mellitus is a metabolic disorder and a significant cause
of morbidity and mortality worldwide [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. As yet, there is no cure
developed for diabetes; and management of the corresponding
lifeimpeding conditions is recommended as the most successful way to
control the disease [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. In fact, the occurrence of the associated
complications can be suspended or even prevented by effective
management of the disease [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ].
      </p>
      <p>
        Among different types of diabetes, the importance of the
selfmanagement for type 1 diabetes mellitus (T1DM) is accentuated
[
        <xref ref-type="bibr" rid="ref19 ref8">8, 19</xref>
        ]. The key factor in T1DM management is to control the blood
glucose level (BGL) within the normal range [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. BGL predictive
models could contribute to achieving this goal. They can help avert
adverse glycaemic events by forecasting them and giving patients the
chance to take corrective actions ahead of time [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
      </p>
      <p>
        The importance of the development of BGL predictive models in
T1DM management has spurred research into this field [
        <xref ref-type="bibr" rid="ref16 ref22">16, 22</xref>
        ].
According to the knowledge requirement, predictive models can be
classified as; physiological, data-driven, and hybrid models [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ].
Datadriven models interpret trends in sequences of data to make
estimations of future BGLs. Machine learning approaches are broadly
adopted in this area [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ].
      </p>
      <p>
        Mirshekarian et al. [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] developed a model to predict blood
glucose in 30-minute and 60-minute horizons using a recursive neural
network (RNN) with long short- term memory (LSTM) units. The
model explored BGL, insulin, food, and activity information as
inputs. For the same prediction horizons, Bertachi et al. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] and Georga
et al. [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], in separate studies, proposed predictive models. Bertachi
et al. applied an artificial neural network contemplating glucose,
insulin, carbohydrate and physical activity as inputs for their system.
BGL profile, insulin, carbohydrate intake and physical activity were
inputs for a support vector regression (SVR) in the model developed
by Georga et al. Investigating continuous glucose monitoring (CGM)
data by recursive and direct deep learning approaches, Xie et al. [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ]
recommended a model for BGL prediction. Martinsson et al. [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]
proposed an automatic forecast model for a prediction horizon of up
to 60 minutes using RNN. The model used only the information from
past BGLs as input. Bunescu et al. [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] created descriptive features to
train a SVR using a physiological model of blood glucose dynamics.
Carbohydrate intake, insulin administration, and the current and past
BGLs were inputs of their model. Despite extensive research devoted
to the development of predictive models, the performance of the
proposed models remains a challenge [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
      </p>
      <p>In this work, we contributed to the improvement of BGL
prediction for T1DM by applying a multi-lag stacking methodology.
Initially, three conventional regression tools—partial least squares,
multilayer perceptron, and long-short term memory—were applied
to forecast BGLs in horizons of 30 and 60 minutes. Each tool was
trained twice; once on a lag of 30 minutes and once on a lag of 60
minutes of CGM data. Therefore, six basic models were created for
each prediction horizon. For each horizon, three stacking systems
were then developed where predictions from a selection of the basic
models were used as features to train a new regression. The first two
stacking systems followed a uni-lag approach. They used predictions
from the three base models trained on a history of 30 minutes and 60
minutes, respectively. The third system was multi-lag and used
predictions from all six base models. The stacking systems resulted in
appreciable improvements in predictive accuracy as compared to the
basic predictive models. The third stacking system showed a
predictive performance better than the other systems.</p>
      <p>This is the first paper, to our knowledge, that has combined
models with different time-lags to generate a multi-lag BGL prediction
system.
2</p>
    </sec>
    <sec id="sec-2">
      <title>DATASET</title>
      <p>
        The Ohio T1DM dataset comprises several features collected from
12 individuals with type 1 diabetes in 8 weeks [
        <xref ref-type="bibr" rid="ref13 ref14">14, 13</xref>
        ]. The last
ten days’ worth of data for each contributor was considered as the
test set. Data for a cohort of six subjects was released in 2018 for the
first BGL prediction challenge [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]; data for another six subjects was
released in 2020 for the second challenge [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ].
      </p>
      <p>In this work, the 2020’s data was investigated for developing and
evaluating predictive models. Among the collected features were
CGM data every 5 minutes, which was the only feature explored in
this work. A brief description of the CGM data in the Ohio T1DM
dataset released for 2020 BGL prediction challenge is displayed in
Table 1.
The first pre-processing task was taking care of missing data.
Missing data in the training set was imputed applying a simple linear
interpolation. Alternatively, for the test set, a linear extrapolation was
employed. This was to ensure the model is not contaminated by
observing future data in its pre-processing stage.</p>
      <p>The next pre-processing step was transferring the time series
forecasting problem to a supervised learning task. To this end, a rolling
window consisting of a lag and future data was used as explanatory
and dependent variables respectively. To give an illustration, for
forecasting BGL of 30 minutes later using a history of 60 minutes, for
example, we used a window with the length of 18. As a consequence
of the 5-minute interval between data points, it therefore follows that
the first 12 data points in the window were explanatory variables, and
the rest were dependent variables.
3.2</p>
    </sec>
    <sec id="sec-3">
      <title>Prediction methods</title>
      <p>First, six basic predictive models were created by means of three
conventional regression tools. Subsequently, employing stacking
learning, three more advanced predictive systems were developed where a
collection of the basic models were considered as base-learners and
a partial least squares regression as meta-learner. All proposed
models/systems were personalised to individuals.
3.2.1</p>
      <sec id="sec-3-1">
        <title>Basic models</title>
        <p>Initially, for each prediction horizon of 30 and 60 minutes, the
following three conventional regressions tools were employed to
generate six basic predictive models—two models by each tool. For this
purpose, these tools were trained once on a history of 30 and once on
a history of 60 minutes.</p>
      </sec>
      <sec id="sec-3-2">
        <title>Partial least squares regression (PLSR)</title>
        <p>
          PLSR, as a basic linear regression, holds substantial popularity
in different applications due to its easy-to-apply nature and
minimal computation time requirement. In a previous work, we applied
PLSR for glucose quantification which provided promising results
[
          <xref ref-type="bibr" rid="ref12">12</xref>
          ].
        </p>
        <p>
          In this work, PLSR was used as one of the regression tools. For
the number of components, different values ranging from 1 to the
length of the input variable were tried. Each time, the predicted
residual sum of squares (P RESS) was calculated as follows. The
number of components (A) resulting in the minimum value for
P RESS=(N A 1) was then selected [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ].
        </p>
        <p>P RESS =</p>
        <p>N
X(yi
i=1
y^i)2
where, N is the size of the evaluation set , yi is reference value,
and y^i is predicted value.</p>
      </sec>
      <sec id="sec-3-3">
        <title>Multilayer perceptron (MLP)</title>
        <p>
          An MLP [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ] with an architecture of one hidden layer including
100 nodes and an output layer was implemented. ReLU was used
as the activation function for he hidden layer, Adam as the
optimiser, and mean absolute error as the loss function. Learning rate
was 0.01, and the training process was based on 100 epochs.
        </p>
      </sec>
      <sec id="sec-3-4">
        <title>Long short-term memory (LSTM)</title>
        <p>
          We used a Vanila LSTM [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ] composed of a single hidden LSTM
layer with 200 nodes, a fully connected layer with 100 nodes, and
an output layer. ReLU was the activation function for both hidden
layers, mean squared error was the loss function, and Adam was
the optimizer. The model trained on 100 epochs with a learning
rate of 0.01.
(1)
3.2.2
        </p>
      </sec>
      <sec id="sec-3-5">
        <title>Stacking systems</title>
        <p>
          Ensemble learning is a machine learning technique that combines
decisions from several models to create a new model. Stacking
(Figure 1) is an ensemble approach that uses predictions from multiple
base-learners (first level models) as features to train a meta-learner
(second level model). This meta-learner then makes the final
predictions on the test set [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ].
        </p>
        <p>In this paper, for each prediction horizon of 30 and 60 minutes,
three stacking systems comprised of two uni-lag and one multi-lag
were developed.</p>
      </sec>
      <sec id="sec-3-6">
        <title>System 1</title>
        <p>The three basic models trained on a history of 30 minutes were
the base-learners of this uni-lag system and a PLSR was its
metalearner.</p>
      </sec>
      <sec id="sec-3-7">
        <title>System 2</title>
        <p>This system was also uni-lag. It was similar to system 1, except it
used the three basic models trained on a history of 60 minutes in
place of 30 minutes as base-learners.</p>
      </sec>
      <sec id="sec-3-8">
        <title>System 3</title>
        <p>In this multi-lag system, all the six basic models were considered
as the base-learners and again a PLSR was the meta-learner. By
performing a multi-lag approach the idea was to help capture a
broader frequency range of BGL dynamics.
3.3</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Evaluation</title>
      <p>The test set was held out, and the train set was used to create the
predictive models/systems. The developed models/systems were then
utilised to predict the test data. The set of evaluation points starts 60
minutes after the beginning of the test set. First evaluation points
would be otherwise similar to the training data, and it can affect the
reliability of the results. Hence, the number of evaluated points for
each patient is 12 less than the number of test examples mentioned
in Table 1. Root mean square error (RMSE) and mean absolute
error (MAE) were calculated as follows and then used as evaluation
metrics.</p>
      <p>RM SE =</p>
      <p>M AE =
r PN
i=1(yi</p>
      <p>N
PN
i=1 jyi</p>
      <p>N</p>
      <p>y^i)2
y^ij
(2)
(3)
where, N , yi , and y^i carry the same definition as in (1).
4</p>
    </sec>
    <sec id="sec-5">
      <title>RESULTS AND DISCUSSION</title>
      <p>This section presents the evaluation results for both the basic
models and stacking systems. Models/systems with a performance
depended on random initialization ran five times, and corresponding
results have been reported in the form of mean and standard
deviation. Extrapolated points were excluded when calculating the
evaluation metrics. All models were built to predict future BGLs up to
the end of the intended prediction horizon, but only the evaluation
results for the horizon of interest are reported.
4.1
4.1.1</p>
    </sec>
    <sec id="sec-6">
      <title>Prediction horizon of 30 minutes</title>
      <sec id="sec-6-1">
        <title>Basic models</title>
        <p>The results of the RMSE and MAE of the basic predictive models for
the prediction horizon of 30 minutes are displayed in Table 2.</p>
        <p>Based on the average of RMSE and MAE for all patients, LSTM
trained on a history of 30 minutes showed the best performance
among the basic models. PLSR with 60-minute lag was the
secondbest model. All models had satisfactory standard deviations.</p>
        <p>LSTM yielded the best overall predictive accuracy among the three
regression tools. However, the results of the other two tools were also
comparable to that of LSTM. It is worth remarking that PLSR, as a
linear regression tool, was able to generate results comparable to that
of LSTM and even better than that of MLP.</p>
        <p>Among all patients, patient 552 had the best overall evaluation
results. The worst results, on the other hand, belonged to patients
584 and 540.</p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>Prediction horizon of 60 minutes</title>
      <sec id="sec-7-1">
        <title>Basic models</title>
        <p>Evaluation results of the stacking systems for a prediction horizon
of 60 minutes are displayed in Table 5. System 3 proposed the best
overall predictions based on average RMSE and MAE values. The
best result among all patients belonged to patient 596. All systems
had low values of standard deviation.
5</p>
      </sec>
    </sec>
    <sec id="sec-8">
      <title>CONCLUSION</title>
      <p>BGL prediction improved using stacking learning concepts. Initially,
a time series problem was translated into a supervised learning task.
Three conventional regression tools were trained with on different
history length of 30 and 60 minutes, resulting in six basic
predictive models. Predictions from the basic models trained with a history
of 30 minutes were fed as features to a regression to build a
combined learner. The learner was then used to make final predictions on
the test set. The same scenario was repeated using the basic models
trained on 60-minute lag observations. In both cases, the combined
learner was able to make more accurate predictions on the test set.
The overall performance further improved when predictions from all
basic models—trained on both histories of 30 and 60 minutes—were
considered as features to train a new learner.
For data analysis we used Python 3.6, TensorFlow 1.15.0 and
Keras 2.2.5. Pandas, NumPy and Sklearn packages of python
were used. The codes are available at: https://gitlab.com/
Heydar-Khadem/multi-lag-stacking.git</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Florencia</given-names>
            <surname>Aguiree</surname>
          </string-name>
          , Alex Brown, Nam Ho Cho, Gisela Dahlquist, Sheree Dodd, Trisha Dunning, Michael Hirst, Christopher Hwang, Dianna Magliano,
          <string-name>
            <given-names>Chris</given-names>
            <surname>Patterson</surname>
          </string-name>
          , et al.,
          <source>'Idf diabetes atlas'</source>
          , (
          <year>2013</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Ramzi</given-names>
            <surname>Ajjan</surname>
          </string-name>
          , David Slattery, and Eugene Wright, '
          <article-title>Continuous glucose monitoring: A brief review for primary care practitioners'</article-title>
          , Advances in therapy,
          <volume>36</volume>
          (
          <issue>3</issue>
          ),
          <fpage>579</fpage>
          -
          <lpage>596</lpage>
          , (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Muhammad</given-names>
            <surname>Asad</surname>
          </string-name>
          and Usman Qamar, '
          <article-title>A review of continuous blood glucose monitoring and prediction of blood glucose level for diabetes type 1 patient in different prediction horizons (ph) using artificial neural network (ann)'</article-title>
          ,
          <source>in Proceedings of SAI Intelligent Systems Conference</source>
          , pp.
          <fpage>684</fpage>
          -
          <lpage>695</lpage>
          . Springer, (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Arthur</given-names>
            <surname>Bertachi</surname>
          </string-name>
          , Lyvia Biagi, Iván Contreras, Ningsu Luo, and Josep Vehí, '
          <article-title>Prediction of blood glucose levels and nocturnal hypoglycemia using physiological models and artificial neural networks</article-title>
          .',
          <string-name>
            <surname>in</surname>
            <given-names>KHD</given-names>
          </string-name>
          @ IJCAI, pp.
          <fpage>85</fpage>
          -
          <lpage>90</lpage>
          , (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Julio</given-names>
            <surname>Borges</surname>
          </string-name>
          ,
          <source>The Power of Ensembles in Deep Learning</source>
          ,
          <year>2019</year>
          . https://towardsdatascience.com/
          <article-title>the-power-of-ensembles-in-deeplearning-a8900ff42be9.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>Danielle</given-names>
            <surname>Bruen</surname>
          </string-name>
          , Colm Delaney, Larisa Florea, and Dermot Diamond, '
          <article-title>Glucose sensing for diabetes monitoring: recent developments'</article-title>
          ,
          <source>Sensors</source>
          ,
          <volume>17</volume>
          (
          <issue>8</issue>
          ),
          <year>1866</year>
          , (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Razvan</given-names>
            <surname>Bunescu</surname>
          </string-name>
          , Nigel Struble, Cindy Marling, Jay Shubrook, and Frank Schwartz, '
          <article-title>Blood glucose level prediction using physiological models and support vector regression'</article-title>
          ,
          <source>in 2013 12th International Conference on Machine Learning and Applications</source>
          , volume
          <volume>1</volume>
          , pp.
          <fpage>135</fpage>
          -
          <lpage>140</lpage>
          . IEEE, (
          <year>2013</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Mol</given-names>
            <surname>Ecol</surname>
          </string-name>
          , 'HHS Public Access',
          <volume>25</volume>
          (
          <issue>5</issue>
          ),
          <fpage>1032</fpage>
          -
          <lpage>1057</lpage>
          , (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Eleni</surname>
            <given-names>I Georga</given-names>
          </string-name>
          , Vasilios C Protopappas, Diego Ardigò, Demosthenes Polyzos, and
          <string-name>
            <surname>Dimitrios</surname>
            <given-names>I Fotiadis,</given-names>
          </string-name>
          '
          <article-title>A glucose model based on support vector regression for the prediction of hypoglycemic events under free-living conditions'</article-title>
          ,
          <source>Diabetes technology &amp; therapeutics, 15(8)</source>
          ,
          <fpage>634</fpage>
          -
          <lpage>643</lpage>
          , (
          <year>2013</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>Sepp</given-names>
            <surname>Hochreiter</surname>
          </string-name>
          and
          <article-title>Jürgen Schmidhuber, 'Long short-term memory'</article-title>
          ,
          <source>Neural computation</source>
          ,
          <volume>9</volume>
          (
          <issue>8</issue>
          ),
          <fpage>1735</fpage>
          -
          <lpage>1780</lpage>
          , (
          <year>1997</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>George</surname>
            <given-names>S Jeha</given-names>
          </string-name>
          , Lefkothea P Karaviti,
          <string-name>
            <surname>Barbara Anderson</surname>
          </string-name>
          ,
          <string-name>
            <surname>EO'Brian Smith</surname>
            ,
            <given-names>Susan</given-names>
          </string-name>
          <string-name>
            <surname>Donaldson</surname>
          </string-name>
          ,
          <string-name>
            <surname>Toniean S McGirk</surname>
          </string-name>
          , and Morey W Haymond, '
          <article-title>Continuous glucose monitoring and the reality of metabolic control in preschool children with type 1 diabetes'</article-title>
          ,
          <source>Diabetes Care</source>
          ,
          <volume>27</volume>
          (
          <issue>12</issue>
          ),
          <fpage>2881</fpage>
          -
          <lpage>2886</lpage>
          , (
          <year>2004</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Heydar</surname>
            <given-names>Khadem</given-names>
          </string-name>
          , Mohammad R Eissa, Hoda Nemat, Osamah Alrezj, and Mohammed Benaissa, '
          <article-title>Classification before regression for improving the accuracy of glucose quantification using absorption spectroscopy'</article-title>
          ,
          <source>Talanta</source>
          ,
          <volume>211</volume>
          ,
          <fpage>120740</fpage>
          , (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>Cindy</given-names>
            <surname>Marling</surname>
          </string-name>
          and Razvan Bunescu, '
          <article-title>The ohiot1dm dataset for blood glucose level prediction:</article-title>
          <source>Update</source>
          <year>2020</year>
          '.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>Cindy</given-names>
            <surname>Marling and Razvan C Bunescu</surname>
          </string-name>
          , '
          <article-title>The OhioT1DM Dataset For Blood Glucose Level Prediction</article-title>
          .',
          <source>in 3rd International Workshop on Knowledge Discovery in Healthcare Data</source>
          , pp.
          <fpage>60</fpage>
          -
          <lpage>63</lpage>
          , (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>John</surname>
            <given-names>Martinsson</given-names>
          </string-name>
          , Alexander Schliep, Björn Eliasson, Christian Meijner, Simon Persson, and Olof Mogren, '
          <article-title>Automatic blood glucose prediction with confidence using recurrent neural networks'</article-title>
          ,
          <source>in 3rd International Workshop on Knowledge Discovery in Healthcare Data, KDH@ IJCAIECAI</source>
          <year>2018</year>
          ,
          <issue>13</issue>
          <year>July 2018</year>
          , pp.
          <fpage>64</fpage>
          -
          <lpage>68</lpage>
          , (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Cooper</surname>
            <given-names>Midroni</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Peter J. Leimbigler</surname>
            , Gaurav Baruah, Maheedhar Kolla,
            <given-names>Alfred J.</given-names>
          </string-name>
          <string-name>
            <surname>Whitehead</surname>
          </string-name>
          , and Yan Fossat, '
          <article-title>Predicting glycemia in type 1 diabetes patients: Experiments with XGBoost'</article-title>
          ,
          <source>CEUR Workshop Proceedings</source>
          ,
          <volume>2148</volume>
          ,
          <fpage>79</fpage>
          -
          <lpage>84</lpage>
          , (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Sadegh</surname>
            <given-names>Mirshekarian</given-names>
          </string-name>
          , Razvan Bunescu, Cindy Marling, and Frank Schwartz, '
          <article-title>Using lstms to learn physiological models of blood glucose behavior'</article-title>
          ,
          <source>in 2017 39th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC)</source>
          , pp.
          <fpage>2887</fpage>
          -
          <lpage>2891</lpage>
          . IEEE, (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Fionn</surname>
            <given-names>Murtagh</given-names>
          </string-name>
          , '
          <article-title>Multilayer perceptrons for classification and regression'</article-title>
          ,
          <source>Neurocomputing</source>
          ,
          <volume>2</volume>
          (
          <issue>5-6</issue>
          ),
          <fpage>183</fpage>
          -
          <lpage>197</lpage>
          , (
          <year>1991</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Shauna</surname>
            <given-names>S Roberts</given-names>
          </string-name>
          , '
          <article-title>Type 1 diabetes'</article-title>
          ,
          <source>Diabetes Forecast</source>
          ,
          <volume>55</volume>
          ,
          <fpage>19</fpage>
          , (
          <year>2002</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>Svante</surname>
            <given-names>Wold</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Michael</given-names>
            <surname>Sjöström</surname>
          </string-name>
          , and Lennart Eriksson, '
          <article-title>Pls-regression: a basic tool of chemometrics', Chemometrics and intelligent laboratory systems</article-title>
          ,
          <volume>58</volume>
          (
          <issue>2</issue>
          ),
          <fpage>109</fpage>
          -
          <lpage>130</lpage>
          , (
          <year>2001</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>Ashenafi</given-names>
            <surname>Zebene</surname>
          </string-name>
          <string-name>
            <surname>Woldaregay</surname>
          </string-name>
          , Eirik Årsand, Taxiarchis Botsis, David Albers,
          <string-name>
            <given-names>Lena</given-names>
            <surname>Mamykina</surname>
          </string-name>
          , and Gunnar Hartvigsen, '
          <article-title>Data-driven blood glucose pattern classification and anomalies detection: machinelearning applications in type 1 diabetes'</article-title>
          ,
          <source>Journal of medical Internet research</source>
          ,
          <volume>21</volume>
          (
          <issue>5</issue>
          ),
          <year>e11030</year>
          , (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>Jinyu</given-names>
            <surname>Xie</surname>
          </string-name>
          and
          <string-name>
            <given-names>Qian</given-names>
            <surname>Wang</surname>
          </string-name>
          , '
          <article-title>Benchmark machine learning approaches with classical time series approaches on the blood glucose level prediction challenge</article-title>
          .',
          <string-name>
            <surname>in</surname>
            <given-names>KHD</given-names>
          </string-name>
          @ IJCAI, pp.
          <fpage>97</fpage>
          -
          <lpage>102</lpage>
          , (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <surname>Zhi-Hua</surname>
            <given-names>Zhou</given-names>
          </string-name>
          ,
          <article-title>Ensemble methods: foundations and algorithms</article-title>
          , CRC press,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>