<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Workshop on Artificial Intelligence and Formal Verification, Logics, Automata and Synthesis (OVERLAY),
September</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Bayesian Neural Predictive Monitoring</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Luca Bortolussi</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Francesca Cairoli</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Nicola Paoletti</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>S. A. Smolka</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>S. D. Stoller</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science, Royal Holloway University</institution>
          ,
          <addr-line>London</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Department of Computer Science, Stony Brook University</institution>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Department of Mathematics and Geoscience, University of Trieste</institution>
          ,
          <addr-line>Trieste</addr-line>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Modelling and Simulation Group, Saarland University</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2020</year>
      </pub-date>
      <volume>25</volume>
      <issue>2020</issue>
      <fpage>95</fpage>
      <lpage>100</lpage>
      <abstract>
        <p>Neural State Classification (NSC) is a recently proposed method for runtime predictive monitoring of Hybrid Automata (HA) using deep neural networks (DNNs). NSC predictors have very high accuracy, yet are prone to prediction errors that can negatively impact reliability. To overcome this limitation, we present Neural Predictive Monitoring (NPM), a technique that complements NSC predictions with estimates of the predictive uncertainty. These measures yield principled criteria for the rejection of predictions likely to be incorrect, without knowing the true reachability values. We also present an active learning method that significantly reduces the NSC predictor's error rate and the percentage of rejected predictions. NPM is based on the use of Bayesian techniques, Bayesian Neural Networks and Gaussian Processes, to learn respectively the predictor and the rejection rule. Our approach is highly efficient, with computation times on the order of milliseconds, and effective, managing in our experimental evaluation to successfully reject almost all incorrect predictions.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>where 1 is the indicator function. A state x ∈ X is called positive w.r.t. a predictor F : X → {0, 1} if
F (x) = 1. Otherwise, x is called negative (w.r.t. F ).</p>
      <p>
        Finding F ∗, i.e., finding a function approximation with minimal error probability, is indeed a classical
machine learning problem, a supervised classification problem in particular, with F ∗ being the classifier,
i.e., the function mapping HA state inputs x into one of two classes: positive or negative.
Training set. In supervised learning, one minimizes a measure of the empirical prediction error w.r.t. a
training set. In our case, the training set Zt is obtained from a finite sample Xt of X by labelling the training
inputs x ∈ Xt using some reachability oracle, that is, a hybrid automata reachability checker like [
        <xref ref-type="bibr" rid="ref1 ref10 ref6 ref8">1, 6, 8, 10</xref>
        ].
Hence, given a sample Xt of X , the training set is defined by Zt = {(x, 1 (Reach(D, x, T )) | x ∈ Xt}.
Prediction errors. No machine learning technique can completely avoid prediction errors. Therefore,
we have to deal with predictive monitors F that are prone to prediction errors, which are of two kinds:
false positives (FPs) and false negatives (FNs). These errors are respectively denoted by predicates fn(x)
and fp(x), whereas pe(x) = fn(x) ∨ fp(x) denotes a generic prediction error.
      </p>
      <p>Let f be the discriminant function associated to the classifier F , i.e., the function that maps the inputs
into class likelihoods. A central objective of this work is to derive, given a predictor f , a rejection criterion
Rf able to identify states x that are wrongly classified by F , i.e., FNs and FPs, without knowing the true
reachability value of x. Further, Rf should be optimal, that is, it should ensure minimal probability of
rejection errors w.r.t. the state distribution X .</p>
      <p>Our solution relies on enriching each prediction with a measure of predictive uncertainty: given f , we
define a function uf : X → U mapping an HA state x ∈ X into some measure uf (x) of the uncertainty of
f about x. The only requirements are that uf must be point-specific and should not use any knowledge
about the true reachability value. The function uf is then used to build an optimal error detection
criterion, as explained below.</p>
      <p>Problem 2 (Uncertainty-based error detection). Given a reachability predictor f , a distribution X over
HA states X, a predictive uncertainty measure uf : X → U over some uncertainty domain U , and a
kind of error e ∈ {pe, fn, fp}, find an optimal error detection rule Gf∗,e : U → {0, 1}, i.e., a function that
minimizes the probability that Gf∗,e makes a mistake in recognizing errors of type e:</p>
      <p>Prx∼X (e(x) 6= Gf∗,e(uf (x))).</p>
      <p>As for Problem 1, we can obtain a sub-optimal solution Gf,e to Problem 2 by expressing the latter as
a supervised learning problem, where the inputs are, once again, sampled according to X and labelled
using a reachability oracle. These observations need to be independent from the above introduced training
set Zt, i.e., the set used to learn the reachability predictor f .</p>
      <p>For e ∈ {pe, f p, f n}, the final rejection rule Rf,e for detecting HA states where the reachability
prediction should not be trusted, and thus rejected, is readily obtained by the composition of the
uncertainty measure and the error detection rule Rf,e = Gf,e ◦ uf : X → {0, 1}, where Rf,e(x) = 1 if the
prediction on state x is rejected; Rf,e(x) = 0, otherwise.</p>
      <p>Validation set. As done for the training set, a set Zv is obtained from a sample Xv of X , independent
from Xt. Given an uncertainty measure uf , we build a validation set Wve = {(uf (x), 1(e(x))) | x ∈ Xv}
which is used to learn Gf,e. The inputs of Wve, referred to as Uv, are the uncertainty measures evaluated
on Xv: Uv = {uf (x) | x ∈ Xv}. Each input uf (x) ∈ Uv is then labelled respectively with 1 or 0 depending
on whether or not the classifier f makes an error of type e.</p>
      <p>A graphical scheme of the full NPM pipeline is presented in 1.</p>
      <p>State distribution</p>
      <p>+
Reachability oracle</p>
      <p>Validation set
#</p>
      <p>Active
Learning
Training set
"</p>
      <p>Neural State
Classifier</p>
      <p>Uncertainty
quantification</p>
      <p>Error detection
criterion</p>
    </sec>
    <sec id="sec-2">
      <title>Methods</title>
      <p>
        In the NSC method of [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] deep neural networks are used to solve Problem 1. In this work, we rely on
Bayesian neural networks (BNN) [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] which combine the best of probabilistic models and neural networks
by learning a posterior distribution over the weights of the neural network. BNNs benefit from the
statistical guarantees of probabilistic models and from the ability of neural networks to be universal
function approximators. In a nutshell, a prior over the parameters is used as input to a neural network, the
output of the neural network is then used to compute the likelihood – with a specified distribution – and
from these it is possible to compute the posterior distribution of the parameters, either by Hamiltionian
Monte Carlo (HMC) sampling [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] or by Variational Inference (VI) [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. The rationale behind this choice is
that BNNs are well suited to solve Problem 2 as they are able to capture predictive uncertainty via the
posterior predictive distribution [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], which accounts for the uncertainty about the parameters. The function
uf maps a point x ∈ X to the mean μx and the standard deviation σx of the empirical approximation of
the predictive distribution.
      </p>
      <p>Bayesian uncertainty measure. Given observations Zt and a BNN fw with w ∼ p(w|Zt), we define
the Bayesian uncertainty measure uf : X → U ⊆ R2 as the function mapping inputs x into the empirical
mean, μx, and variance, σx2, of the predictive distribution p(yˆ | x, Zt), where yˆ is the class predicted by
fw, where w is a sample from the posterior p(w | Zt). Formally, ∀x ∈ X, uf (x) = (μx, σx2).</p>
      <p>
        An unseen input x must have sufficiently low uncertainty values in order for the prediction to be
accepted. We seek to find those uncertainty values that optimally separate the points in Uv – the validation
set – in relation to their classes, that is, separate points yielding errors from those that do not. Optimal
thresholds can be automatically identified by solving an additional supervised learning problem, in this
work we leverage Gaussian Process classification (GPC) [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ].
2.1
      </p>
      <p>Active Learning
As the accuracy of a classifier increases with the quality and the quantity of observed data, adding samples
to Zt will generate a more accurate predictor, and similarly, adding samples to Wv will lead to more
precise error detection. Ideally, one wants to maximize accuracy while using the least possible amount of
additional samples, because obtaining labeled data is expensive (in NPM, labelling each sample entails
solving a reachability checking problem), and the size of the datasets affects the complexity and the
dimension of the problem. Therefore, to improve the accuracy of our learning models efficiently, we need
a strategy to identify the most “informative” additional samples.</p>
      <p>For this purpose, we propose an uncertainty-aware active learning solution, where the re-training
points are derived by first sampling a large pool of unlabeled data, and then considering only those points
where the current predictor f is still uncertain according to the rejection rule Rf .</p>
      <p>Active learning algorithm. We query the HA reachability checker to label the points rejected by Rf
and we divide them into two groups: one group is added to the training set Zt, producing the augmented
dataset Zta; the other is added to the validation set Zv, producing Zva. The first step consists in retraining
the reachability predictor on Zta. Let fa denote the new predictor. The second step requires extracting
the augmented validation dataset Wva from Za. To do so, we must first train a new uncertainty measure
v
ufa so as to reflect the new predictor fa. Then, the new error detection rule Gfa is trained on Wva. In
conclusion, this process leads to an updated rejection rule Rfa = Gfa ◦ ufa , which is expected to have a
reduced rate of incorrect rejections.</p>
      <p>
        A frequentist approach is also possible [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]: the NSC is a deterministic Neural Network, the uncertainty
is measured using the theory of conformal predictions [
        <xref ref-type="bibr" rid="ref13 ref17 ref3">3, 13, 17</xref>
        ] and the rejection rule is described by a
support vector classifier [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
3
      </p>
    </sec>
    <sec id="sec-3">
      <title>Experimental Results</title>
      <p>
        We experimentally evaluate the proposed method on a benchmark of six hybrid system models with
varying degrees of complexity. We consider four deterministic case studies: the model of the spiking neuron
(SN) action potential [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] and the classic inverted pendulum (IP) on a cart, which are two-dimensional
models with non-linear dynamics, the artificial pancreas (AP) [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], which is a six-dimensional non-linear
model, and the helicopter model (HC) [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], which is a linear model with 29 state variables. In addition,
we analyze two non-deterministic models with non-linear dynamics: a cruise controller (CC) [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], whose
input space has four dimensions, and a triple water tank (TWT)1, which is a three-dimensional model.
Performance measures. We want our method to be capable of working at runtime, which means it
must be extremely fast in making predictions and deciding whether to trust them. The time required
to train the reachability predictor and the error detection rule does not affect its runtime efficiency, as
it is performed in advance (offline) only once. Furthermore, to avoid overfitting, we did not tune the
architecture (i.e., number of neurons and hidden layers) to optimize the performance for our data and,
for the sake of simplicity, we choose the same architecture for all the case studies. However, each case
study requires a fine tuning of the the inference hyper-parameters. VI parameters differ from HMC
hyper-parameters. Keeping that in mind, the relevant performance metrics for NPM are the accuracy
of the reachability predictor F , the error detection rate, that measures the proportion of errors made on
the test set by F that are actually recognized by Rf , and the overall rejection rate of the rejection rule
Rf , that measures the overall proportion of test points rejected by Rf . Clearly, we want our method to
be reliable and thus, detect the majority of prediction errors (high detection rate) without being overly
conservative, i.e., keeping a low rejection rate, in order not to reduce its effectiveness.
Computational costs. Training an NPM requires training the state classifier, generating the datasets
Wv, which requires computing the uncertainty values for each point in Zv, and finally training the error
rejection rule. All these steps are performed offline. On the other hand, the time required to evaluate
the NPM, on a given test input x∗, is composed of the time needed to make a prediction and the time
needed to choose whether to accept it or not, which take on the order of milliseconds. NPM’s efficiency is
not directly affected by the complexity of the system under analysis but only by the complexity of the
underlying learning problems.
      </p>
      <p>Evaluation. We consider three configurations: the initial configuration, where the predictor and error
detection rules are derived via supervised learning; the active configuration, where the initial models
are retrained via our uncertainty-aware active learning; and the passive configuration, where the initial
models are retrained using a uniform sampling strategy to augment the dataset and with the same number
1http://dreal.github.io/benchmarks/networks/water/
of observations of the active configuration. In this way, we can evaluate the benefits of employing an
uncertainty-based criterion to retrain our models.</p>
      <p>Our experimental evaluation demonstrates that our reachability predictors attain high accuracies,
consistently above 97.4% and the benefits of active learning are visible from an overall reduction of the
rejection rate and an overall increase in the state-classifier accuracy compared to the passive approach.
Furthermore, VI scales better than HMC with respect to the dimension of the system and we observe that
on average VI outperforms HMC: the initial VI approach yield an NSC accuracy of 98.95%, a rejection
rate of 5.19% and a recognition rate of 86.18%. The passive results introduce only minor improvements,
whereas active learning yields a significant reduction in the rejection rate (from 5.19% to 2.36%), an
increase in the NSC accuracy (from 98.95% to 99.32%) and an increase in the overall recognition rate
(from 86.18% to 91.85%).</p>
      <sec id="sec-3-1">
        <title>Det. rate</title>
      </sec>
      <sec id="sec-3-2">
        <title>Rej. rate</title>
      </sec>
      <sec id="sec-3-3">
        <title>Acc. Det. rate</title>
      </sec>
      <sec id="sec-3-4">
        <title>Artificial Pancreas (AP)</title>
        <p>Initial VI 99.29 95.77
Active VI 99.71 96.55
Passive VI 99.47 100.00</p>
      </sec>
      <sec id="sec-3-5">
        <title>Inverted Pendulum (IP)</title>
        <p>Initial VI 99.58 80.95
Active VI 99.58 85.71
Passive VI 99.71 75.81</p>
      </sec>
      <sec id="sec-3-6">
        <title>Cruise Controller (CC)</title>
        <p>Initial VI 99.01 97.98
Active VI 99.84 100.00
Passive VI 99.74 100.00</p>
      </sec>
      <sec id="sec-3-7">
        <title>Triple Water Tank (TWT)</title>
        <p>Initial VI 99.60 77.50
Active VI 99.67 84.85
Passive VI 99.50 75.40</p>
      </sec>
      <sec id="sec-3-8">
        <title>Helicopter (HC)</title>
        <p>Initial VI 98.14 88.71
Active VI 98.90 92.86
Passive VI 98.66 87.54</p>
      </sec>
      <sec id="sec-3-9">
        <title>Spiking Neuron (SN)</title>
        <p>Initial VI 98.18 91.21
Active VI 98.20 91.11
Passive VI 98.20 98.52
Rej. rate</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Conclusion</title>
      <p>Our approach overcomes the computational footprint of reachability checking (infeasible at runtime), while
improving on traditional runtime verification by being able to detect future violations in a preemptive
way. Among future directions, we plan to extend our approach to support more complex and real-world
systems that include noise and partial observability.</p>
    </sec>
    <sec id="sec-5">
      <title>References</title>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>M.</given-names>
            <surname>Althoff</surname>
          </string-name>
          .
          <article-title>An introduction to CORA 2015</article-title>
          .
          <source>In Proc. of the Workshop on Applied Verification for Continuous and Hybrid Systems</source>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>R.</given-names>
            <surname>Alur</surname>
          </string-name>
          .
          <article-title>Formal verification of hybrid systems</article-title>
          .
          <source>In Proceedings of the Ninth ACM International Conference on Embedded Software (EMSOFT)</source>
          , pages
          <fpage>273</fpage>
          -
          <lpage>278</lpage>
          ,
          <year>Oct 2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>V.</given-names>
            <surname>Balasubramanian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.-S.</given-names>
            <surname>Ho</surname>
          </string-name>
          , and
          <string-name>
            <given-names>V.</given-names>
            <surname>Vovk</surname>
          </string-name>
          .
          <article-title>Conformal prediction for reliable machine learning: theory, adaptations and applications</article-title>
          . Newnes,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>M.</given-names>
            <surname>Betancourt</surname>
          </string-name>
          .
          <article-title>A conceptual introduction to hamiltonian monte carlo</article-title>
          .
          <source>arXiv:1701.02434</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>C. M.</given-names>
            <surname>Bishop</surname>
          </string-name>
          .
          <article-title>Pattern recognition and machine learning</article-title>
          . Springer,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>S.</given-names>
            <surname>Bogomolov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Forets</surname>
          </string-name>
          , G. Frehse,
          <string-name>
            <given-names>K.</given-names>
            <surname>Potomkin</surname>
          </string-name>
          , and
          <string-name>
            <given-names>C.</given-names>
            <surname>Schilling</surname>
          </string-name>
          .
          <article-title>Juliareach: a toolbox for set-based reachability</article-title>
          .
          <source>In Proceedings of the 22nd ACM International Conference on Hybrid Systems: Computation and Control</source>
          , pages
          <fpage>39</fpage>
          -
          <lpage>44</lpage>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>T.</given-names>
            <surname>Brihaye</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Doyen</surname>
          </string-name>
          , G. Geeraerts,
          <string-name>
            <given-names>J.</given-names>
            <surname>Ouaknine</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.-F.</given-names>
            <surname>Raskin</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J.</given-names>
            <surname>Worrell</surname>
          </string-name>
          .
          <article-title>On reachability for hybrid automata over bounded time</article-title>
          . In L. Aceto,
          <string-name>
            <given-names>M.</given-names>
            <surname>Henzinger</surname>
          </string-name>
          , and J. Sgall, editors,
          <source>Automata, Languages and Programming</source>
          , pages
          <fpage>416</fpage>
          -
          <lpage>427</lpage>
          , Berlin, Heidelberg,
          <year>2011</year>
          . Springer Berlin Heidelberg.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>X.</given-names>
            <surname>Chen</surname>
          </string-name>
          , E. Ábrahám, and
          <string-name>
            <given-names>S.</given-names>
            <surname>Sankaranarayanan</surname>
          </string-name>
          .
          <article-title>Flow*: An analyzer for non-linear hybrid systems</article-title>
          . In International Conference on Computer Aided Verification, pages
          <fpage>258</fpage>
          -
          <lpage>263</lpage>
          . Springer,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>X.</given-names>
            <surname>Chen</surname>
          </string-name>
          and
          <string-name>
            <given-names>S.</given-names>
            <surname>Sankaranarayanan</surname>
          </string-name>
          .
          <article-title>Model predictive real-time monitoring of linear systems</article-title>
          .
          <source>In Real-Time Systems Symposium (RTSS)</source>
          ,
          <year>2017</year>
          IEEE, pages
          <fpage>297</fpage>
          -
          <lpage>306</lpage>
          . IEEE,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>S.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kong</surname>
          </string-name>
          , and
          <string-name>
            <given-names>E. M.</given-names>
            <surname>Clarke</surname>
          </string-name>
          .
          <article-title>dreal: An smt solver for nonlinear theories over the reals</article-title>
          .
          <source>In International conference on automated deduction</source>
          , pages
          <fpage>208</fpage>
          -
          <lpage>214</lpage>
          . Springer,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>T. A.</given-names>
            <surname>Henzinger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. W.</given-names>
            <surname>Kopke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Puri</surname>
          </string-name>
          , and
          <string-name>
            <given-names>P.</given-names>
            <surname>Varaiya</surname>
          </string-name>
          .
          <article-title>What's decidable about hybrid automata? Journal of computer and system sciences</article-title>
          ,
          <volume>57</volume>
          (
          <issue>1</issue>
          ):
          <fpage>94</fpage>
          -
          <lpage>124</lpage>
          ,
          <year>1998</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>N.</given-names>
            <surname>Paoletti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. S.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. A.</given-names>
            <surname>Smolka</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S.</given-names>
            <surname>Lin</surname>
          </string-name>
          .
          <article-title>Data-driven robust control for type 1 diabetes under meal and exercise uncertainties</article-title>
          .
          <source>In International Conference on Computational Methods in Systems Biology</source>
          , pages
          <fpage>214</fpage>
          -
          <lpage>232</lpage>
          . Springer,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>H.</given-names>
            <surname>Papadopoulos</surname>
          </string-name>
          .
          <article-title>Inductive conformal prediction: Theory and application to neural networks</article-title>
          .
          <source>In Tools in artificial intelligence. InTech</source>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>D.</given-names>
            <surname>Phan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Paoletti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Grosu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. A.</given-names>
            <surname>Smolka</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S. D.</given-names>
            <surname>Stoller</surname>
          </string-name>
          .
          <article-title>Neural state classification for hybrid systems</article-title>
          .
          <source>In Automated Technology for Verification and Analysis</source>
          , volume
          <volume>11138</volume>
          of Lecture Notes in Computer Science, pages
          <fpage>422</fpage>
          -
          <lpage>440</lpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>C. E.</given-names>
            <surname>Rasmussen</surname>
          </string-name>
          and
          <string-name>
            <given-names>C. K.</given-names>
            <surname>Williams</surname>
          </string-name>
          .
          <article-title>Gaussian processes for machine learning</article-title>
          , volume
          <volume>1</volume>
          . MIT press Cambridge,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>S. D.</given-names>
            <surname>Stoller</surname>
          </string-name>
          .
          <article-title>Neural predictive monitoring</article-title>
          .
          <source>In Runtime Verification: 19th International Conference, RV</source>
          <year>2019</year>
          , Porto, Portugal, October 8-
          <issue>11</issue>
          ,
          <year>2019</year>
          , Proceedings, volume
          <volume>11757</volume>
          , page 129. Springer Nature,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>V.</given-names>
            <surname>Vovk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gammerman</surname>
          </string-name>
          , and
          <string-name>
            <given-names>G.</given-names>
            <surname>Shafer</surname>
          </string-name>
          .
          <source>Algorithmic learning in a random world. Springer Science &amp; Business Media</source>
          ,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>