<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Parallel Numerical Methods for Ordinary Di erential Equations: a Survey</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Svyatoslav I. Solodushkin</string-name>
          <email>s@mail.ru</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Irina F. Iumanova</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Krasovskii Institute of Mathematics and Mechanics</institution>
          ,
          <addr-line>Yekaterinburg</addr-line>
          ,
          <country>Russia solodushkin</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Ural Federal University</institution>
          ,
          <addr-line>Yekaterinburg</addr-line>
          ,
          <country country="RU">Russia</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Tremendous e orts have been made in the eld of parallel numerical methods for ODE. Special volumes are dedicated to the theme we are going to discuss [1]. The brief survey of some classical results and recent developments is the aim of this paper. We particulary focus on small scale parallelism across the method, but not large scale parallelism across the system.</p>
      </abstract>
      <kwd-group>
        <kwd>parallel numerical methods</kwd>
        <kwd>parallelism across the method small scale parallelism</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>Modern science poses a lot of computationally intensive problems, among them
weather forecast or heart modeling. It is widely believed that the only
appropriate means to solve these problems is to use parallel computers. This fact
dictates the necessity to elaborate special classes of numerical algorithms which
are inherently parallel and designed for use on parallel computers.</p>
      <p>
        In this paper we consider parallel numerical methods for initial value problem
(IVP) for ordinary di erential equations
y0(x) = f (x; y(x));
y(x0) = y0;
(
        <xref ref-type="bibr" rid="ref1">1</xref>
        )
1. parallelism across the system or, that is the same, parallelism across the
space;
2. parallelism across the method or, that is the same, parallelism across the
time.
      </p>
      <p>Methods fallen into the rst group are usually based on more or less obvious
techniques such as exploiting parallelism of evaluation of function f or solving
linear equations which arise at each step of an otherwise standard IVP solver. For
traditional forward-step methods, e. g. explicit Runge{Kutta methods, this
approach implies simple dividing the original system into a number of subsystems
which are processed concurrently by separated computing nodes with
interprocessor communications at the end of each step of IVP solver. It is clear that this
approach is e ective for high dimensional systems only. In practice such systems
appear in molecular dynamics modelling, where thousands of particles should be
taken into account, or after the discretization of time dependent PDE.</p>
      <p>The second group includes numerical algorithms, which allow several
simultaneous function evaluations within each step of IVP solver. These methods are
suitable for multi-core processors or computers with a few processors with fast
interprocessor communication which use shared memory.</p>
      <p>The second group also includes methods which search solution for many step
simultaneously, they are also known as time parallel methods and can be
classi ed into four di erent groups: methods based on multiple shooting, methods
based on domain decomposition and waveform relaxation, space-time multigrid
methods and direct time parallel methods [3].</p>
      <p>A dedicated survey should be devoted to parallel across the space numerical
methods, and this is the reason why they are beyond the scope of present article
which is focused on parallel across the time numerical methods.
2</p>
      <p>
        Predictor-Corrector, Runge{Kutta and Block Type
Methods
Miranker and Linger [5] were probably the rst who proposed small-scale
parallel method of predictor-corrector type. They mentioned that at rst sight the
sequential nature of the numerical methods like (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) does not permit parallel
computation on all of the processors to be performed, in other words the front
of computation is too narrow to take advantage of more than one processor.
      </p>
      <p>p
yn+1 = ync + h2 3f (xn; ync)</p>
      <p>f (xn 1; ync 1) ;
yn+1 = ync + h2 f (xn; ync) + f (xn+1; ynp+1) ;
c
p
Indeed, ync must be computed before yn+1; which in turn must be computed
c
before yn+1 etc.</p>
      <p>
        However, if the predictor is taken in another form, as represented in (
        <xref ref-type="bibr" rid="ref3">3</xref>
        ), then
p
ync and yn+1 can be evaluated in parallel by two processors.
      </p>
      <p>
        ynp+1 = ync 1 + 2hf (xn; ynp);
ync = ync 1 + h2 f (xn 1; ync 1) + f (xn; ynp) ;
(
        <xref ref-type="bibr" rid="ref2">2</xref>
        )
(
        <xref ref-type="bibr" rid="ref3">3</xref>
        )
      </p>
      <p>
        Method (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) is based on couple of Adams{Bashfort and Adams{Moulton
methods of second order, where the rst one is used to estimate yn+1 which
is substituted to the second one to avoid the implicity and, consequently, the
necessity of solving a nonlinear equation. In (
        <xref ref-type="bibr" rid="ref3">3</xref>
        ) the predictor outruns the corrector
and stands one step ahead; this made the parallel implementation possible.
      </p>
      <p>As a development of this idea Miranker and Liniger showed how to
systematically derive a general class of numerical integration methods which can be
executed on multiple processors in parallel, and present a stability and convergence
analysis for those methods. More precisely, they considered predictor-corrector
methods in a PECE mode (one predicted derivative evaluation and one
corrected derivative evaluation) where calculation advances s steps at a time. So,
if 2s processors are available it is possible to arrange tasks in such a way that
each processor performs either a predictor or a corrector calculation.</p>
      <p>Continuing the research started by Mirankel Katz, Franklin and Sen [6]
studied the stability properties of predictor-corrector parallel methods. Speci cally,
they considered the case where there are two processors, i.e. s = 1: They showed
that for a xed Adams{Moulton corrector the optimally stable parallel predictor
(in the sense that the parallel scheme has a maximum stability interval on the
negative real axis) is the Adams{Bashforth predictor shifted to the right by one
integration step. Methods constructed in [6] are not A-stable.</p>
      <p>
        Another type of numerical methods which possess an internal parallelism are
block methods. Block method which nds solution in r points at one step is
called r-dots block method. By analogy with the classi cation into one-step and
multistep for classical methods, block methods could also be divided into these
two groups. Shampine and Watts studied a family of one-step block methods,
and gave one speci c example (
        <xref ref-type="bibr" rid="ref4">4</xref>
        ):
yn+1
yn+2
= h
Shampine and Watts took advantages of this formula which is more accurate at
the end of the block than in the middle and proved (
        <xref ref-type="bibr" rid="ref4">4</xref>
        ) has global convergence
like O(h4) instead of O(h3) one might expect. They also proved the A-stability
of (
        <xref ref-type="bibr" rid="ref4">4</xref>
        ), so, in a sense they overcame the Dahlquist barrier which states that of
the linear multistep methods the trapezoidal rule is the most accurate A-stable
method.
      </p>
      <p>To avoid the necessity of solving nonlinear equation which arise at each step
Shampine and Watts supplied a predictor-corrector form of one-step r-dots block
methods. Good prediction allow them to compute the corrector two times only
rather than iterate until it is satis ed, furthermore suitable predictor-corrector
scheme asymptotically gives the same result as iterating to completion.
Unfortunately block implicit methods when used as a predictor-corrector combination
are even not A( )-stable.</p>
      <p>To expand the stability region Lee and Song [8] proposed a family of explicit
and implicit multistep block methods as well as block predictor corrector schemes
based on couples of these methods where explicit one is used as predictor.</p>
      <p>Feldman used collocation approach to derive implicit multistep block
methods of general form [9]. He presented formulas for A-stable one-step 4-dots block
implicit method and mentioned that collocation approach allow to construct
Astable one-step block methods of higher order as well. At the same time multistep
methods constructed in such a way are not A-stable.</p>
      <p>Deep analysis of predictor-corrector methods suggests a simple general
paradigm for achieving parallelism across the time in traditional forward-step
methods: group the stages of a the method into blocks for which all evaluations
associated with each block can be performed simultaneously. This idea could be
illustrated by parallel explicit and diagonally implicit Runge{Kutta (ERK and
DIRK, respectively) methods in a very natural way.</p>
      <p>An s-stage RK method has the following form</p>
      <p>s
yn+1 = yn + h X bikn;i;
i=1</p>
      <p>
        s
kn;i = f xn + cih; yn + h X aij kn;j ; i = 1; :::; s
j=1
(
        <xref ref-type="bibr" rid="ref5">5</xref>
        )
where bi; ci; ai;j are coe cients; ai;j = 0; i j for explicit RK formulas and
ai;j = 0; i &lt; j and at least one aii 6= 0 for diagonally implicit RK formulas. Let
us denote by A = fa)ij g the s s RK matrix.
      </p>
      <p>Suppose, for example, that the coe cients of explicit RK method have the
zero-pattern presented in Table 1.</p>
      <p>Each arrow in the corresponding "production graph" depicted in Fig. 1,
pointing from vertex i to vertex j; stands for a non-zero aji: Here the vertices 2
and 3 are independent and can be processed in parallel. The number of vertices
in the longest chain of successive arrows (in this case 3) is called the number of
e ective sages [13] or number of sequential function evaluations [12].</p>
      <p>In general, let us consider RK matrix A which could be partitioned (possibly
after a permutation of the stages) as
2 D1
where, for ERK methods, Dl; l = 1; : : : ; ; are zero matrixes and, for DIRK
methods, Dl; l = 1; : : : ; ; are (possibly di erent) diagonal matrixes. For any
l = 1; : : : ; all kn;i in l-th block of ERK method could be computed
concurrently as soon as all kn;j in all previous blocks are available. Similarly, for any
l = 1; : : : ; each kn;i in l-th block of DIRK method depends on itself and the
previously competed kn;j in blocks 1; : : : ; l 1: Thus, all kn;i in l-th block could
be computed in parallel by solving the system of uncoupled equations.</p>
      <p>Unfortunately, these ERK schemes do not have a high potential of
parallelism. The following theorem is a severe restriction on parallel methods [12]
Theorem 1. The order p of the ERK method with sequential stages satis es
the inequality p for any number of available processors.</p>
      <p>Let us remark that stability function of these ERK methods is a polinimial and
therefore they can not be A-stable.</p>
      <p>
        On the other hand DIRK methods have better stability properties and allow
one to achive higher order of convergency. For example, method (
        <xref ref-type="bibr" rid="ref6">6</xref>
        ) is L-stable
(but not B-stable) and has a 4-th order.
      </p>
      <p>
        1/2
2/3
1/2
1/3
1/2
0
-5/2
-5/3
-1
0
2/3
5/2
4/3
3/2
0
0
1/2
0
-1
0
0
0
2/3
3/2
(
        <xref ref-type="bibr" rid="ref6">6</xref>
        )
This method consists of two uncoupled blocks, i. e. = 2; of two equations in
each, hence kn;1 and kn;2 could be found in parallel, after that kn;3 and kn;4
could be found in parallel too.
      </p>
      <p>The theorem 1 motivated to pose the question whether it is always possible
to nd ERK method of order p using not more than p e ective stages, assuming
that su ucient number of processors are available, these methods are called
P-optimal. For sequential ERK of order p 4 it is possible to choose coe cient
of the method in such a way that the number of stages s is equal to p; so
these sequential ERK are P-optimal. At the same time even for p 5 the
number of stages is greater than p and increase rapidly, see table 2. Also
the number of stages smin; the number of stages S for which these RK
methods have actually been constructed, the number of e ective stages Seff
are reported. Special techniques allow one to construct methods which are
P-optimal, because some stages are performed in parallel (Table 2). So, these
methods allow one to decrease the computation time with the help of parallelism.
where sequential stages are performed and s processors are available is
assumed. The following theorem discover the connection between the order and
the number of iterations.</p>
      <p>
        Theorem 2. The parallel iterated Runge{Kutta method (
        <xref ref-type="bibr" rid="ref5">5</xref>
        ) in form (
        <xref ref-type="bibr" rid="ref7">7</xref>
        ) is of
order p = min(p0; ); where p0 denotes the order of the basic method.
This theorem shows that the choice = p0 yields P-optimal ERK methods.
      </p>
      <p>The next question is the least number of processors needed to implement an
optimal ERK method. Houwen and Sommejeir [13] took as basic method the
s-stage Gaussian{Legendre type RK method, which has the smallest number of
stages with respect to their order. This allowed them to construct the method
of order p = 2s which is P-optimal on s processors.</p>
      <p>In this scheme one may use linear multistep (LM) predictors reducing the
number of e ective stages. First results based on LM predictors are reported by
Lie [14], using a 4th-order, 2-stage Gauss{Legendre corrector and a 3rd-order
Hermite extrapolation predictor. Future investigation of LM correctors in this
scheme was made in [13].</p>
      <p>In conclusion it should be said that predictor-corrector and block as well
as Runge{Kutta methods are ideally suited to be used on the few cores of a
multicore processor, but they do not have the potential for large scale parallelism.</p>
    </sec>
    <sec id="sec-2">
      <title>Extrapolation Methods</title>
      <p>Many authors note that extrapolation methods posses a high degree of
parallelism and o er an extremely simple technique to obtain for generating high-order
methods [15{18].</p>
      <p>
        Let us consider a basic method of order p for integrating (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) from x0 until
x1 = x0 + h: As usual simple method of low order, e. g. explicit Euler method or
Crank{Nicolson method, is taken as a basic. Denote the numerical approximation
to the exact solution value y(x1) by y(x1; h): Let the error of approximation
allows the series expansion in powers of hq; where q = 2 for symmetric basic
method and q = 1 otherwise. Let us consider a sequence of integration steps
hi = h=i; i = 1; : : : ; r; then the corresponding r-point extrapolation method is
de ned as follow
      </p>
      <p>r
X i = 1;
i=1</p>
      <p>r
y1 = X i y(x0 + h; hi);</p>
      <p>i=1
i=1</p>
      <p>
        i
r
X ji = 0; j = p; p + q; : : : ; p + (r
2)q:
(
        <xref ref-type="bibr" rid="ref8">8</xref>
        )
Theorem 3. Let the basic method providing the values y(x1; hi) be of order p;
then the extrapolation method (
        <xref ref-type="bibr" rid="ref8">8</xref>
        ) has order p + q(r 1):
      </p>
      <p>
        As soon as y1 is computed one can perform the second step using the y1 as a
new initial value at t1; etc. Obviously the di erent terms y(x1; hi) in (
        <xref ref-type="bibr" rid="ref8">8</xref>
        ) could
be computed in parallel. It is clear that the time needed to compute y(x1; ; hi)
is proportional to i; hence to balance the load it is recommended to compute
y(x1; h1) and y(x1; hr) on the rst processor, y(x1; h2) and y(x1; hr 1) on the
second processor and so on. In this way b(r + 1)=2c processors are used.
      </p>
      <p>Based on Richardson extrapolation technique Houwen in [15] constructed
method which requires less processors to be optimal that the
Gauss{Legendrebased parallel iterated RK method. For example, an optimal RK method of
order ten requires only three processors when using Richardson extrapolation
and ve processors when using predictor-corrector iteration of the tenth-order
Gauss{Legendre method.</p>
      <p>Korch et al. considered in [17] the parallel explicit extrapolation methods for
high-dimensional ODEs, up to 108: They analyze and compare the scalability of
several implementations for distributed-memory architectures which use di erent
load balancing strategies and di erent loop structures. By exploiting the special
structure of a large class of ODE systems, the communications costs can be
reduced signi cantly. More then, by processing the micro-steps using a
pipelinelike structure, the locality of memory references could be increased and better
utilization of the cache memory can be achieved.
4</p>
      <p>Multiple Shooting and Parareal Algorithm
According to Gander [3] time parallel time integration methods have received
renewed interest over the last decade because of the advent of massively parallel
computers, which is mainly due to the clock speed limit reached on today's
processors. When solving time dependent partial di erential equations, the time
direction is usually not used for parallelization. But when parallelization in space
saturates, the time direction o ers itself as a further direction for parallelization.
The time direction is however special, and for evolution problems there is a
causality principle: the solution later in time is a ected (it is even determined)
by the solution earlier in time, but not the other way round. Algorithms trying
to use the time direction for parallelization must therefore be special, and take
this very di erent property of the time dimension into account.</p>
      <p>Below we overview only two parallel in time methods.</p>
      <p>
        The basic idea of multiple shooting method proposed in [19] is to apply a
shooting method which was originally developed for boundary value problems
to an initial value problem (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ). To do this one splits the time interval into
subintervals [x0; x1]; [x1; x2]; : : : ; [xn 1; xn]; and then solves on each subinterval
the underlying initial value problem
y00(x) = f (x; y0(x));
y0(0) = y0
y10(x) = f (x; y1(x)); : : :
y1(x1) = U1
yn0 1(x) = f (x; yn 1(x));
yn 1(xn 1) = Un 1
(
        <xref ref-type="bibr" rid="ref9">9</xref>
        )
together with the matching conditions U1 = y0(x1); : : : ; Un 1 = yn 2(xn 1);
where Ui; i = 1; : : : ; n 1 are so called shooting parameters which are unknown.
This leads to the system of non-linear equations
8 U1
&gt;
&gt;&lt; U2
&gt; : : :
&gt;: Un 1
y0(x1) = 0
y1(x2) = 0
yn 1(xn 1) = 0
:
(
        <xref ref-type="bibr" rid="ref10">10</xref>
        )
      </p>
      <p>To determine the shooting parameters the Newton's method could be applied
to this system. Fortunately the Jacobian matrix has a very special band form
and could be inverted easily. The corresponding iteration process has a following
form</p>
      <p>U k+1 = y0</p>
      <p>
        0
(
        <xref ref-type="bibr" rid="ref11">11</xref>
        )
and allow parallel implementation. Chartier and Philippe prove that (
        <xref ref-type="bibr" rid="ref11">11</xref>
        )
converges locally quadratically. Note, that (
        <xref ref-type="bibr" rid="ref11">11</xref>
        ) has a form of Bellen and Zennaro
method [20] who used multiple shooting technique to parallelize discrete
recurrent relations yn+1 = Fn+1(yn) and reported that the speedup of 53 for a
problem with 400 time steps.
      </p>
      <p>Following Lions, Maday and Turinici [21] let us explain the parareal
algorithms on the simple scalar model
Uik); i = 0; 1; 2; : : : ; n
2
y0(x) =
ay(x);
The solution is rst approximated using Backward Euler on the time grid xi
with coarse time step ;</p>
      <p>Yi1+1</p>
      <p>Yi1 =
a Yi1+1;</p>
      <p>Y01 = y0:
The approximate solution values Yi1 are then used to compute on each time
interval [xi; xi+1] exactly and in parallel the solution of
Theorem 4. [21] The parareal scheme is of order k; i. e. there exists Ck such
that
kYik
y(xi)k +</p>
      <p>max
x2[xi;xi+1]
kyik(x)
y(x)k</p>
      <p>Ck
k:</p>
      <p>In particular, this result means that with each iteration of the parareal
algorithm, one obtains a numerical time stepping scheme which has a truncation
error that is one order higher than before. So for a xed iteration number k;
one can obtain high order time integration methods that are naturally parallel.
The same proposition also holds for Forward Euler. Unfortunately in both
discretization schemes the stability of the higher order methods obtained with the
parareal correction scheme degrades with iterations.</p>
      <p>Lions et al. gave two numerical examples: one for a heat equation where they
obtain a simulated speedup of a factor 8 with 500 processors, and one for a
semi-linear advection di usion problem where the obtained speedup was 18.
Acknowledgements This research is supported by RFBR 17-01-00033 and
1701-00392, Russian Science Foundation (RSF) 14-35-00005. We acknowledge the
support by the program 02.A03.21.0006 on 27.08.2013.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <source>Advances in Computational Mathematics Volume 7, Issue</source>
          <volume>1</volume>
          {
          <fpage>2</fpage>
          ,
          <year>1997</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Gear</surname>
            ,
            <given-names>C.W.:</given-names>
          </string-name>
          <article-title>The parallel methods for ordinary di erential equations</article-title>
          ,
          <source>Tech. Rep</source>
          . UIUCDCS R{
          <volume>87</volume>
          {
          <fpage>1369</fpage>
          ,
          <string-name>
            <surname>Comp</surname>
          </string-name>
          . Sci. Dept.,
          <source>Univ. of Illinois</source>
          , Urbana, IL,
          <year>1987</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Gander</surname>
            ,
            <given-names>M.J.:</given-names>
          </string-name>
          <article-title>50 Years of Time Parallel Time Integration</article-title>
          .
          <source>In: in Multiple Shooting and Time Domain Decomposition</source>
          , Springer,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Jackson</surname>
            ,
            <given-names>K.R.:</given-names>
          </string-name>
          <article-title>A Survey of Parallel Numerical Methods for Initial Value Problems for Ordinary Di erential Equations</article-title>
          .
          <source>In: IEEE Transactions on Magnetics</source>
          ,
          <year>1991</year>
          . Vol.
          <volume>27</volume>
          . No. 5, P.
          <volume>3792</volume>
          {
          <fpage>3797</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Miranker</surname>
            ,
            <given-names>W.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Linger</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          :
          <article-title>Parallel methods for the numerical integration of ordinary di erential equations</article-title>
          . In: Math. Comp.,
          <year>1967</year>
          . Vol.
          <volume>21</volume>
          . P.
          <volume>303</volume>
          {
          <fpage>320</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Katz</surname>
            ,
            <given-names>I.N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Franklin</surname>
            ,
            <given-names>M.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sen</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Optimally stable parallel predictors for AdamsMoulton correctors</article-title>
          .
          <source>In: Camp. and Moth</source>
          .
          <source>with Appl.</source>
          ,
          <year>1977</year>
          . Vol 3. pp 217{
          <fpage>233</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Shampine</surname>
            ,
            <given-names>L.F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Watts</surname>
            ,
            <given-names>H.A.</given-names>
          </string-name>
          :
          <article-title>Block Implicit One-Step Methods</article-title>
          . In: Math. Comp.,
          <year>1969</year>
          . V. 23. P.
          <volume>731</volume>
          {
          <fpage>740</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>M.G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Song</surname>
          </string-name>
          , R.W.:
          <article-title>Computation of Stability Region of some Block Methods and Their Application to Numerical Solutions of ODEs</article-title>
          . In: Proceedings of XIV Baikal International Conference, Baikal, Siberia, Russia,
          <year>July 2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Feldman</surname>
            ,
            <given-names>L.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nazarova</surname>
            ,
            <given-names>I.A.</given-names>
          </string-name>
          :
          <article-title>Parallel algorithms for the numerical decision of Caushy's problem for ordinary di erential equations systems</article-title>
          .
          <source>In: Mathematical modelling</source>
          ,
          <year>2006</year>
          . Vol.
          <volume>18</volume>
          . No. 9. P.
          <volume>17</volume>
          {
          <fpage>31</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Iserles</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , N rsett, S.P.:
          <article-title>On the theory of parallel Runge{Kutta methods</article-title>
          .
          <source>IMA J. Numer. Anal.</source>
          ,
          <year>1990</year>
          . Vol.
          <volume>10</volume>
          . P.
          <volume>463</volume>
          {
          <fpage>488</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Jackson</surname>
            ,
            <given-names>K.R.</given-names>
          </string-name>
          , N rsett,
          <string-name>
            <surname>S.P.:</surname>
          </string-name>
          <article-title>The potential of parallelism in Runge{Kutta methods. Part 1: RK formulas in standard form</article-title>
          .
          <source>Report 239/90</source>
          ,
          <year>1992</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Hairer</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wanner</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <article-title>N rsett</article-title>
          , S.P.:
          <string-name>
            <surname>Solving Ordinary Di erential Equations</surname>
            <given-names>I</given-names>
          </string-name>
          ,
          <year>1993</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>van Der Houwen</surname>
            ,
            <given-names>P.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sommeijer</surname>
            ,
            <given-names>B.P.</given-names>
          </string-name>
          :
          <article-title>Parallel iteration of high-order RungeKutta methods with stepsize control</article-title>
          .
          <source>In: Journal of Computational and Applied Mathematics</source>
          ,
          <year>1990</year>
          . Volume
          <volume>29</volume>
          . Issue 1. P.
          <volume>111</volume>
          {
          <fpage>127</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Lie</surname>
            ,
            <given-names>I.:</given-names>
          </string-name>
          <article-title>Some aspects of parallel Runge-Kutta methods</article-title>
          ,
          <source>Report No. 3/87</source>
          , University of Trondheim,
          <source>Division Numerical Mathematics</source>
          ,
          <year>1988</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>van Der Houwen</surname>
          </string-name>
          , P.J.:
          <article-title>Parallel step-by-step methods</article-title>
          .
          <source>In: Applied Numerical Mathematics</source>
          ,
          <year>1993</year>
          . Volume
          <volume>29</volume>
          . Issue 1. P.
          <volume>69</volume>
          {
          <fpage>81</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Rauber</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Runger</surname>
          </string-name>
          , G.:
          <article-title>Load balancing schemes for extrapolation methods</article-title>
          .
          <source>In: Concurrency: Practice and Experience</source>
          ,
          <year>1997</year>
          . Volume
          <volume>9</volume>
          . Issue 3. P.
          <volume>181</volume>
          {
          <fpage>202</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Korch</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rauber</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Scholtes</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Scalability and locality of extrapolation methods on large parallel systems</article-title>
          .
          <source>In: Euro-Par</source>
          <year>2010</year>
          ,
          <string-name>
            <surname>Part</surname>
            <given-names>II</given-names>
          </string-name>
          , LNCS
          <volume>6272</volume>
          ,
          <year>2010</year>
          . P.
          <volume>65</volume>
          {
          <fpage>76</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Ketcheson</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          , bin Waheed, U.:
          <article-title>A comparison of high-order explicit Runge{Kutta, extrapolation, and deferred correction methods in serial and parallel</article-title>
          .
          <source>In: Communications in Applied Mathematics and Computational Science</source>
          ,
          <year>2014</year>
          , Volume
          <volume>9</volume>
          . Issue 2. P.
          <volume>175</volume>
          {
          <fpage>200</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Chartier</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Philippe</surname>
            ,
            <given-names>B.A.</given-names>
          </string-name>
          <article-title>: parallel shooting technique for solving dissipative ODEs</article-title>
          . In: Computing,
          <year>1993</year>
          . Volume
          <volume>51</volume>
          . Issue 3. P.
          <volume>209</volume>
          {
          <fpage>236</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Bellen</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zennaro</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Parallel algorithms for initial-value problems for di erence and di erential equations</article-title>
          .
          <source>In: J. Comput. Appl</source>
          . Math.,
          <year>1989</year>
          . Volume
          <volume>25</volume>
          . Issue 3. P.
          <volume>341</volume>
          -
          <fpage>350</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>Lions</surname>
            ,
            <given-names>J.-L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Maday</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Turinici</surname>
          </string-name>
          , G.:
          <article-title>A parareal in time discretization of PDEs</article-title>
          . In: C.R. Acad. Sci. Paris,
          <string-name>
            <surname>Serie</surname>
            <given-names>I</given-names>
          </string-name>
          ,
          <volume>2001332</volume>
          ,
          <year>2001</year>
          . P.
          <volume>661</volume>
          {
          <fpage>668</fpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>