<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>How to Tame Your Anticipatory Algorithm</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Allegra De Filippo</string-name>
          <email>allegra.defilippo@unibo.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Michele Lombardi</string-name>
          <email>michele.lombardi2@unibo.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Michela Milano</string-name>
          <email>michela.milano@unibo.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>DISI, University of Bologna</institution>
          ,
          <addr-line>Bologna</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <fpage>36</fpage>
      <lpage>44</lpage>
      <abstract>
        <p>Sampling-based anticipatory algorithms can be very effective at solving online optimization problems under uncertainty, but their computational cost may be prohibitive in some cases. Given an arbitrary anticipatory algorithm, we present three methods that allow to retain its solution quality at a fraction of the online computational cost, via a substantial degree of offline preparation. Our approaches are obtained by combining: 1) a simple technique to identify likely future outcomes based on past observations; 2) the (expensive) offline computation of a “contingency table”; and 3) an efficient solution-fixing heuristic. We ground our techniques on two case studies: an energy management system with uncertain renewable generation and load demand, and a traveling salesman problem with uncertain travel times. In both cases, our techniques achieve high solution quality, while substantially reducing the online computation time.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Optimization problems under uncertainty arise in many important application domains,
such as production scheduling or energy system management. These problems often
benefit from making all or part of their decisions online, reacting and adapting to
external events. In this context, stochastic online anticipatory algorithms have proved
particularly effective (see e.g. [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]). However, many of such algorithms have a considerable
computational cost, which may be problematic if (as it is often the case) online
decisions must be taken within a short time frame.
      </p>
      <p>In most practical settings, however, a substantial amount of time and information
is available before the online problem is solved, in an offline phase. We refer to this
sort of data as offline information. Usually, it is employed to characterize the uncertain
elements and for sampling likely outcomes (i.e. scenarios). We will show how to exploit
this information at a much deeper level.</p>
      <p>
        This paper reports and summarizes the methods and the results presented in [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] and
accepted for publication at IJCAI 2019.
      </p>
      <p>We propose three hybrid offline/online methods that build over a given,
samplingbased, anticipatory algorithm, and allow to match its solution quality at a fraction of
the online computational cost: 1) a technique to estimate the probability of future
outcomes, given past observations; 2) a scheme for building a “contingency table”, with
precomputed solutions to guide the online choices; and 3) an efficient fixing heuristic
for adapting the precomputed solutions to run-time conditions.</p>
      <p>We ground our approaches on a (numeric) energy management problem with
uncertain loads and generation from Renewable Energy Sources (RES), and on a (discrete)
Copyright c 2019 for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
Traveling Salesman Problem with uncertain travel times. We show how our methods
reach a solution quality comparable with the anticipatory algorithm, with lower (or
dramatically lower) online computational cost.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Background and Related Work</title>
      <p>
        Historically, the literature on optimization under uncertainty has focused on offline
problems [
        <xref ref-type="bibr" rid="ref17 ref19">17, 19</xref>
        ]. These methods usually rely on sampling (yielding a number of
scenarios) to obtain a statistical model of future uncertainty.
      </p>
      <p>
        More recently, improvements in the solution techniques and computational power
have enabled the application of similar methods to online optimization. This lead to
so-called online anticipatory algorithms, which proved very effective at finding robust,
high quality, solutions as uncertainty slowly reveals itself [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. Notable algorithms and
results are described in [
        <xref ref-type="bibr" rid="ref15 ref2 ref3 ref5 ref6">5, 2, 3, 15, 6</xref>
        ].
      </p>
      <p>
        Online anticipatory algorithm typically rely on scenario sampling to estimate the
possible developments for a fixed number of future steps, known as look-ahead horizon.
Larger sample sizes result in higher accuracy, but also in more and bigger (possibly
NP-hard) problems to be solved. Considerable research effort has therefore focused on
improving the efficiency of these algorithms [
        <xref ref-type="bibr" rid="ref14 ref2 ref4">2, 4, 14</xref>
        ]. The approaches from [
        <xref ref-type="bibr" rid="ref12 ref13 ref16">12, 16,
13</xref>
        ] attempt instead to reduce the number of scenarios by increasing their relevance by
taking into account past observations while sampling.
3
      </p>
      <p>“Taming” an Anticipatory Algorithm
Our goal is reducing the online computational cost of a given sampling-based
anticipatory algorithm, referred to as A, by exploiting the existence of an offline phase. Such A
algorithm is the main input for all our methods.</p>
      <p>
        Similarly to [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], we view online optimization under uncertainty as a stochastic
nstage problem. At each stage some uncertainty is resolved, and some decision must be
made. A stage k is associated to a decision variable xk (e.g. the power flows between
loads and generators) and a state variable sk (summarizing the effect of past decisions).
All variables may be vector-valued.
      </p>
      <p>We assume that uncertainty is exogenous, i.e. not affected by the decisions (e.g. the
RES generation does not depend on how we choose to route it), and modeled via a set of
random variables i. Which variables are observed at each stage depends on the state,
and is controlled by a peek function: O = peek(sk) which returns a set O with the
indices of the observed variables. We will use the notation O to denote the observed
variables, and O for the unobserved ones.</p>
      <p>Base Anticipatory Algorithm Sampling-based algorithms rely on scenarios to estimate
future outcomes. Formally, a scenario ! specifies a value i! for all the random
variables. Given a set of scenarios, the system state sk, and values for O corresponding
to the observed uncertainty, we assume that A can compute the decisions for stage k:
xk = A(sk; O; f !
g!2
)
(1)
Algorithm 1 ANTICIPATE (s1; )</p>
      <p>Sample a set of scenarios
for k = 1 : : : n do
xOk==OA[(spke;ekO(;sfk)!g!2 )
sk+1 = next(sk; xk; O)
return s, x
# observe uncertainty
# obtain decisions
# determine next state
(2)
(3)
Once the decisions are computed, the next state can be determined. This is controlled via
a state transition function next that, based on the current state, decisions, and observed
uncertainty, computes:</p>
      <p>sk+1 = next(sk; xk; O)
Given the initial state s1, a set of scenarios , and a set of values sampled from k
(which represent the online observations), we can simulate the execution of the method
via Algorithm 1. The O set is assumed to be initially empty. This generic anticipatory
algorithm will be referred to as ANTICIPATE in the remainder of the paper.</p>
      <p>OFFLINE INFORMATION Defining a representative set of scenarios is critical for
the approach effectiveness and it is usually done by exploiting the available offline
information. Here, we assume that the such offline information is a collection of observed
uncertain values. This definition captures many practical cases (e.g. forecasts or
predictions, historical data, data from test runs). More importantly, this means that the offline
information is in fact a collection of scenarios. We will denote the offline information as
I, index its element with !, and assume (as it is usual) that I is representative of the true
probability distribution of the random variables. A set of scenarios for ANTICIPATE
can be obtained by sampling a number of elements from I.
3.1</p>
      <sec id="sec-2-1">
        <title>Basic Contributions</title>
        <p>All our methods rely on three techniques, which will be described in this section.</p>
        <p>PROBABILITY ESTIMATION FOR SCENARIO SAMPLING: Using a fixed set of
scenarios (as in ANTICIPATE) is beneficial when the i variables are statistically
independent. When they are not, however, the set of scenarios may loose relevance as
uncertainty is resolved. For example, a scenario based on a cloudy day forecast becomes
less likely if fair weather is observed at the beginning of online execution. Formally, at
stage k we wish to sample scenarios that are likely to occur given the past observations,
i.e. to sample the unobserved variables O according to the conditional distribution
P ( O j O). If we draw the scenarios from the offline information (which guarantees
physically meaningfulness), then sampling requires to estimate the conditional
probabilities of the elements in I. From basic probability theory, this is given by the ratio of
two joint probabilities:
8! 2 I;</p>
        <p>P ( O! j O) =</p>
        <p>P ( O O!)</p>
        <p>P ( O)
where P ( O O!) is the probability that observed values occur together with the
remaining predictions from the scenario, and P ( O) is the probability that the values are
observed. The joint probability at the numerator can be approximated via any density
Algorithm 2 BUILDTABLE (s1; AA)
for ! 2 I do</p>
        <p>
          s!; x! = AA(s1; !)
return f !; s!; x!g!2T
estimation method, such as Kernel Density Estimation [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ], Gaussian Mixture Models
[
          <xref ref-type="bibr" rid="ref10">10</xref>
          ], or recent Deep Learning techniques such as Normalizing Flows [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ] and Real
NVP [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ]. Any such method can be trained on the offline information to obtain an
estimator P~( ) for the joint distribution of the random variables. An estimator for the
distribution P ( O) at the denominator can then be derived from P~( ) via
marginalization, i.e. by averaging the contribution of the unobserved variables. We perform this step
over all possible completions of the observed values in the offline information. Overall,
we have:
8! 2 I;
        </p>
        <p>P~( O! j O) =</p>
        <p>P~( O O!)</p>
        <p>P!02T P~( O O!0 )
This estimator defines a discrete distribution over the offline information I. The chosen
marginalization technique guarantees an estimate that is approximately proportional
(not approximately equal) to the true P ( O). Hence, we have that:
8! 2 I;</p>
        <p>P ( O! j O) / P~( O! j O)
Sampling from I according to Equation (4) yields scenarios with a distribution that
takes into account the observed values.</p>
        <p>BUILDING A CONTINGENCY TABLE: If a significant amount of time is available
in the offline phase, we can exploit the offline information more aggressively, by trying
to prepare for each likely future development. Intuitively, we can treat each scenario
! 2 I as if it were an actual sequence of online observations, and process it via some
anticipatory algorithm. By doing this, we build a pool of solutions that can then be
used to guide an online method. The process is outlined in Algorithm 2, which requires
as input the initial state s1 of the system, and a solution algorithm AA, accepting the
same parameters as ANTICIPATE. The result is an augmented version of the offline
information, where each scenario ! is additionally associated to the sequence of states
s! visited by the algorithm and its sequence of decisions x!. We refer to this data
structure as contingency table, and to its elements as traces. We denote the table as T .</p>
        <p>FIXING HEURISTIC: We use the traces from T to guide an efficient fixing heuristic,
which tries to choose decisions having the largest chance of being optimal. Formally, it
solves:
(4)
(5)
(6)
arg maxfP (xk j sk O) : xk 2 Xkg
where P is the probability that the chosen xk is optimal, given the state sk and the
observed uncertainty. The Xk set represents the feasible decision space, which is defined
via problem-dependent constraints and auxiliary variables. Closed-forms for P can be
obtained separately for discrete and numeric problems, based on the contingency table.</p>
      </sec>
      <sec id="sec-2-2">
        <title>Algorithm 3 FIXING (s1, , T )</title>
        <p>for k = 1 : : : n do</p>
        <p>O = O [ peek(sk)</p>
        <p>= top elements in T by descending Equation (9)
Compute pjv and/or p! based on
Solve Equation (7)/(10) to obtain xk
sk+1 = next(sk; xk; O)
return s; x
Overall, in case of discrete decisions, the problem from Equation (6) translates into:
:</p>
        <p>j=1 v2Dj
arg min &lt;8 Xm X log pjvJxkj = vK : xk 2 Xk
pjv =</p>
        <p>P!2T;xk!j=v P (!)</p>
        <p>P!2T P (!)
where J K denotes the truth value of a predicate, Dj is the domain of xkj , and:
Here, P (!) is a compact notation for the probability that we reach the same state as
trace !, and then everything goes according to plan. It can be approximated using:
8! 2 T;</p>
        <p>P (!) / P~(ss!k+1 j sk)P~( O! j O)
where P~( O j O) is the estimator from Equation (4), and P~(ssk+1 j sk) is a second
estimator obtained via similar means. The cost function in Equation (7) is linear if a
one-hot encoding is adopted for xkj , and the size of T affects only the computation of
the pjv values. Overall, the problem is efficient to solve. In case of numeric decisions,
we have instead:
9
=
;
9
=
;
(7)
(8)
(9)
(10)
(11)
8 m
arg min &lt;X</p>
        <p>X p!
:j=1 !2T
1
2 j
with:
p! = P!02T P (!0)</p>
        <p>P (!)
(xkj</p>
        <p>xk!j )2 : xk 2 Xk
The cost function is quadratic and convex, and the problem size is small due to the same
arguments as Equation (7).</p>
        <p>Intuitively, the discrete version of the heuristic is related to minimizing weighted
discrepancies w.r.t. the traces in T , i.e. to weighted Hamming distances. The numeric
version is instead related to weighted Euclidean distances. The pseudo-code for the
heuristic is provided in Algorithm 3. The only difference with the process described
so far is that the pjv and p! probabilities may be computed based on a subset of
the full contingency table. This may be useful to bias the choice of the online decision
according to the most relevant traces.
Algorithm 4 ANTICIPATE-D (s1; )</p>
        <p>Train the P~( ) estimator on I
for k = 1 : : : n do</p>
        <p>O = O [ peek(sk)
Sxakm=pleA(sfkro;mOT;,facc!ogrd!i2ng t)o Equation (4)
sk+1 = next(sk; xk; O)
return s, x
Algorithm 5 CONTINGENCY (s1; )</p>
        <p>Train the P~( ) estimator on I
T = BUILDTABLE(s1; ANTICIPATE)
Train the P~(sksk+1) estimators on T , for all k
s; x = FIXING(s1; ; T )
return s, x
3.2</p>
      </sec>
      <sec id="sec-2-3">
        <title>Hybrid Offline/Online Methods</title>
        <p>Our three solution methods can now be defined with relative ease, by combining the
techniques just described.</p>
        <p>ANTICIPATE-D Our first hybrid method is obtained from ANTICIPATE by simply
replacing the static set of samples with a dynamically adjusted one. The dynamic set
can be populated according to the estimated probabilities from Equation (4), so as not to
loose relevance: this may enable to reach similar solution qualities with fewer scenarios,
at the cost of training an estimator offline. We refer to this approach as ANTICIPATE-D,
and its pseudo-code is in Algorithm 4</p>
        <p>CONTINGENCY The second method is based on the idea of computing robust
solutions for the scenarios in the offline information, and then use them as guidance for
the FIXING heuristic. Robust solutions are obtained by using ANTICIPATE, so that
hopefully the (fast) fixing heuristic will be able to match their quality: the price to pay is a
hefty offline computational effort. We refer to this approach as CONTINGENCY, and its
pseudo-code is reported in Algorithm 5.</p>
        <p>CONTINGENCY-D Our final method is similar to the previous one, except that the
contingency table is populated with non-robust solutions. This is done by using
ANTICIPATE with a single scenario, given by the values of ! (i.e. the pretend online
observations). This technique (referred to as ANTICIPATE1) provides perfect information about
the future, so that achieving robustness is entirely delegated to the FIXING heuristic. The
approach is likely to loose reliability, but has two important advantages: 1) lower offline
computational costs; and 2) while ANTICIPATE is a stochastic algorithm, ANTICIPATE1
is deterministic. So, this method may provide anticipatory-like results even when no
anticipatory algorithm is available. We refer to this method as CONTINGENCY-D, and
its pseudo-code is reported in Algorithm 6.
4</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Grounding the Approaches</title>
      <p>Grounding our approaches requires to specify: 1) the x, s and variables, 2) the peek
and next functions, 3) the sampling-based algorithm A, and 4) the feasible space Xk
Algorithm 6 CONTINGENCY-D (s1; )</p>
      <p>Train the P~( ) estimator on I
T = BUILDTABLE(s1; ANTICIPATE1)
Train the P~(sksk+1) estimators on T , for all k
s; x = FIXING(s1; ; T )
return s, x
for the FIXING heuristic. Additionally, evaluating the solution quality requires to define
5) a cost metric.</p>
      <p>
        We show how this can be done in two case studies: 1) a Virtual Power Plant
energy management problem with numerical decisions; and 2) a combinatorial Traveling
Salesman Problem with uncertain travel times. In both cases, the input anticipatory
algorithm A is given by a Mathematical Programming model, based on the Sample
Average Approximation. The models are slight improvements over those by [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], whose work
brought to attention the interplay between offline and online phases. Both approaches
are serviceable, but not necessarily representative of the state-of-the-art (especially for
the TSP). We focus on the minimal information needed.
5
      </p>
    </sec>
    <sec id="sec-4">
      <title>Experimentation</title>
      <p>We empirically evaluated the three hybrid offline/online methods on realistic instances
for the two case studies. The baseline in both cases are myopic heuristics, obtained by
running ANTICIPATE with an empty set of scenarios.</p>
      <p>
        Our methods are evaluated over different uncertainty realizations, obtained by
sampling the random variables for the loads and RES generation in the VPP, and for the
travel times in the TSP. In both cases, we use models of uncertainty that ensure
realistic statistical dependence between the variables. These models are used to obtain the
offline information I and the sequences of observations for the actual experiments. For
the VPP, grid electricity prices change every 15 minutes, which is also the duration
of our online stages. New offline information (e.g. market prices) becomes available
every day, hence our horizon corresponds to 24 4 = 96 stages. We use (real)
physical bounds for power generation from [
        <xref ref-type="bibr" rid="ref1 ref9">1, 9</xref>
        ]. The initial battery state, efficiency, and
power flow limit are also based on real data [
        <xref ref-type="bibr" rid="ref1 ref9">1, 9</xref>
        ]. The same goes for the data, which
is also from [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. Different instances have then been obtained by manually scaling load
and RES generation. For the TSP we use classical benchmarks1, by including
problems from 10 to 40 nodes. In the TSP each stage represents a visit, hence our horizon
corresponds to the total number of nodes. We use Kernel Density Estimation (KDE
with Gaussian Kernels) to obtain all approximate distributions. As an underlying solver
we use Gurobi 2, which can handle both MILPs and Quadratic Programs. Each
evaluated algorithm (in each considered configuration) is run 50 times, with the same 50
sequences of realizations. We use a time limit of 300 seconds for both the VPP and the
TSP. For each run we record both the time required by each approach and the
corresponding solution cost, and we report their average values over the 50 realizations. In
1 http://myweb.uiowa.edu/bthoa/TSPTWBenchmarkDataSets.htm
2 Available at http://www.gurobi.com
all cases, jIj = jT j = 100, and for the CONTINGENCY method, the contingency table is
built using ANTICIPATE with 20 scenarios. Due to space constraints, we report results
only for a few representative instances.
      </p>
      <p>The offline training times of the KDE models are roughly the same for all the three
hybrid methods ( 65 sec for the VPP and 32 sec for the TSP). Building the
contingency tables for CONTINGENCY takes 6; 000 sec in the VPP and 15; 000 sec
in the TSP, but only 400 and 2; 000 sec for CONTINGENCY-D. In Section 5 we
show the cost/quality tradeoff of the proposed methods and of ANTICIPATE for the VPP
(base instance) and for the TSP (a representative 20 customers instance). The use of a
dynamic set of scenarios allow ANTICIPATE-D to work better than ANTICIPATE. The
CONTINGENCY method is surprisingly close in terms of quality to the original
anticipatory algorithm, especially considered that its online computational cost is dramatically
smaller (one to two orders of magnitude). CONTINGENCY-D performs (as expected)
slightly worse than CONTINGENCY, but it still much better than the myopic heuristic.
Increasing the number of guiding traces is beneficial for both methods, and in particular
for CONTINGENCY-D.
6</p>
    </sec>
    <sec id="sec-5">
      <title>Conclusion</title>
      <p>We have presented three methods that can be applied to a generic anticipatory
algorithm to reduce its online computational effort by exploiting offline information. In
particular, both CONTINGENCY and CONTINGENCY-D are dramatically faster than
ANTICIPATE during online operation. Between the two of them CONTINGENCY is
significantly more reliable in terms of quality, but may require a substantial amount of offline
computation. The ANTICIPATE-D technique provides a modest advantage in terms of
solution time, but can match and even surpass ANTICIPATE in terms of quality.</p>
      <p>The ability to shift part of the computational cost to an offline stage provides a
significant degree of flexibility to stochastic anticipatory algorithm, and likely to increase
their applicability. We believe there is room for improving the scalability and efficiency
of our methods, and achieving this goal is part of our current research directions.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Bai</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Miao</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ran</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ye</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Optimal dispatch strategy of a virtual power plant containing battery switch stations in a unified electricity market</article-title>
          .
          <source>Energies</source>
          <volume>8</volume>
          (
          <issue>3</issue>
          ),
          <fpage>2268</fpage>
          -
          <lpage>2289</lpage>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Bent</surname>
            , R., Van Hentenryck,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Regrets only! online stochastic optimization under time constraints</article-title>
          .
          <source>In: AAAI</source>
          . vol.
          <volume>4</volume>
          , pp.
          <fpage>501</fpage>
          -
          <lpage>506</lpage>
          (
          <year>2004</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Bent</surname>
            , R., Van Hentenryck,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Online stochastic optimization without distributions</article-title>
          .
          <source>In: ICAPS</source>
          . vol.
          <volume>5</volume>
          , pp.
          <fpage>171</fpage>
          -
          <lpage>180</lpage>
          (
          <year>2005</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Borodin</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>El-Yaniv</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          :
          <article-title>Online computation and competitive analysis</article-title>
          . cambridge university press (
          <year>2005</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <issue>5</issue>
          .
          <string-name>
            <surname>Chang</surname>
            ,
            <given-names>H.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Givan</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chong</surname>
            ,
            <given-names>E.K.</given-names>
          </string-name>
          :
          <article-title>On-line scheduling via sampling</article-title>
          .
          <source>In: AIPS</source>
          . pp.
          <fpage>62</fpage>
          -
          <lpage>71</lpage>
          (
          <year>2000</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>De Filippo</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lombardi</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Milano</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Methods for off-line/on-line optimization under uncertainty</article-title>
          .
          <source>In: IJCAI</source>
          . pp.
          <fpage>1270</fpage>
          -
          <lpage>1276</lpage>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>De Filippo</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lombardi</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Milano</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>How to tame your anticipatory algorithm</article-title>
          . In: to be published at IJCAI (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Dinh</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sohl-Dickstein</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bengio</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Density estimation using real nvp</article-title>
          .
          <source>arXiv preprint arXiv:1605.08803</source>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Espinosa</surname>
            ,
            <given-names>A.N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ochoa</surname>
            ,
            <given-names>L.N.</given-names>
          </string-name>
          :
          <article-title>Dissemination document “low voltage networks models and low carbon technology profiles”</article-title>
          .
          <source>Tech. rep.</source>
          , University of Manchester (
          <year>June 2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Gauvain</surname>
            ,
            <given-names>J.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>C.H.</given-names>
          </string-name>
          :
          <article-title>Maximum a posteriori estimation for multivariate gaussian mixture observations of markov chains</article-title>
          .
          <source>IEEE transactions on speech and audio processing 2</source>
          (
          <issue>2</issue>
          ),
          <fpage>291</fpage>
          -
          <lpage>298</lpage>
          (
          <year>1994</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Hentenryck</surname>
            ,
            <given-names>P.V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bent</surname>
          </string-name>
          , R.:
          <article-title>Online stochastic combinatorial optimization</article-title>
          . The MIT Press (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12. John,
          <string-name>
            <given-names>G.H.</given-names>
            ,
            <surname>Langley</surname>
          </string-name>
          ,
          <string-name>
            <surname>P.</surname>
          </string-name>
          :
          <article-title>Static versus dynamic sampling for data mining</article-title>
          .
          <source>In: KDD</source>
          . vol.
          <volume>96</volume>
          , pp.
          <fpage>367</fpage>
          -
          <lpage>370</lpage>
          (
          <year>1996</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Lin</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tang</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yao</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          :
          <article-title>Dynamic sampling approach to training neural networks for multiclass imbalance classification</article-title>
          .
          <source>IEEE Transactions on Neural Networks and Learning Systems</source>
          <volume>24</volume>
          (
          <issue>4</issue>
          ),
          <fpage>647</fpage>
          -
          <lpage>660</lpage>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Mercier</surname>
          </string-name>
          , L.,
          <string-name>
            <surname>Van Hentenryck</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Performance analysis of online anticipatory algorithms for large multistage stochastic integer programs</article-title>
          .
          <source>In: IJCAI</source>
          . pp.
          <fpage>1979</fpage>
          -
          <lpage>1984</lpage>
          (
          <year>2007</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Mercier</surname>
          </string-name>
          , L.,
          <string-name>
            <surname>Van Hentenryck</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Amsaa: A multistep anticipatory algorithm for online stochastic combinatorial optimization. Integration of AI and OR Techniques in Constraint Programming for Combinatorial Optimization Problems pp</article-title>
          .
          <fpage>173</fpage>
          -
          <lpage>187</lpage>
          (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Philpott</surname>
            ,
            <given-names>A.B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>De Matos</surname>
            ,
            <given-names>V.L.</given-names>
          </string-name>
          :
          <article-title>Dynamic sampling algorithms for multi-stage stochastic programs with risk aversion</article-title>
          .
          <source>European Journal of Operational Research</source>
          <volume>218</volume>
          (
          <issue>2</issue>
          ),
          <fpage>470</fpage>
          -
          <lpage>483</lpage>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Powell</surname>
          </string-name>
          , W.B.:
          <article-title>A Unified Framework for Optimization Under Uncertainty</article-title>
          ,
          <source>chap. 3</source>
          , pp.
          <fpage>45</fpage>
          -
          <lpage>83</lpage>
          . INFORMS (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Rezende</surname>
            ,
            <given-names>D.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mohamed</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Variational inference with normalizing flows</article-title>
          .
          <source>arXiv preprint arXiv:1505.05770</source>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Sahinidis</surname>
            ,
            <given-names>N.V.</given-names>
          </string-name>
          :
          <article-title>Optimization under uncertainty: state-of-the-art and opportunities</article-title>
          .
          <source>Computers &amp; Chemical Engineering</source>
          <volume>28</volume>
          (
          <issue>6</issue>
          ),
          <fpage>971</fpage>
          -
          <lpage>983</lpage>
          (
          <year>2004</year>
          ), http://www.sciencedirect.com/science/ article/pii/S0098135403002369, fOCAPO 2003 Special issue
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Silverman</surname>
            ,
            <given-names>B.W.</given-names>
          </string-name>
          :
          <article-title>Density estimation for statistics and data analysis</article-title>
          .
          <source>Routledge</source>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>