<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Markov Approximation of Zero-sum Differential Games</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Hence</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Yurii Averboukh Krasovskii Institute of Mathematics and Mechanics S. Kovalevskoy str</institution>
          ,
          <addr-line>16, 620990 Yekaterinburg</addr-line>
          ,
          <country country="RU">Russia</country>
        </aff>
      </contrib-group>
      <fpage>73</fpage>
      <lpage>80</lpage>
      <abstract>
        <p>The paper is concerned with approximations of a value function of a differential game. To this end we introduce a approximation of the original differential game by a continuous-time Markov game with infinite state space. The value function of this game in a given region is approximated by a solution of a finite state Markov game. This yields the approximation of the value function of the original differential game by the finite system of ODEs.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Introduction
results are closed to the problems of thr approximation of the control system by Markov chains considered in
[Boue &amp; Dupuis, 1999], [Kushner &amp; Dupuis, 2001].</p>
      <p>Paper [Averboukh, 2016] is concerned with the general continuous-time stochastic game. A near optimal
strategy for the continuous-time stochastic game is constructed using the value function of a game with a different
dynamics. The general theory implies the approximation of the value function of the differential game by the
infinite system of ODEs. This system is obtained using dynamic programming arguments for the approximating
continuous-time Markov game.</p>
      <p>In the paper we introduce the approximation of the value function of the differential game by the finite system
of ODEs. This approach can be used for numerical schemes. The paper is organized as follows. In Section
2 we recall the general theory of differential games. Section 3 deals with the approximation of the original
differential game by a Markov game with infinite state space. This leads to the approximation of the value
function by infinite system of ODEs. The approximation of the differential game by a Markov game with the
finite state space is considered in Section 4. In this section we obtain an approximation of the value function of
the differential game by the finite system of ODEs.
2
x(t) = f (t, x(t), u(t), v(t)), t ∈ [0, T ], x(t) ∈ Rd, u(t) ∈ U, v(t) ∈ V.
(1)
Here u(t) (respectively, v(t)) is a instantaneous control of the first (respectively, second) player. The aim of the
first (respectively, second) player is to minimize (respectively, maximize) the terminal payoff σ(x(T )).</p>
      <p>We use feedback approach proposed by Krasovskii and Subbotin. To introduce this formalization let us denote
by U [t0] (respectively, V[t0]) the set of all measurable functions from [t0, T ] to U (receptively, V ).</p>
      <p>We assume that
1. the sets U and V are metric compacts;
2. the functions f and σ are continuous;
3. the functions f and σ are bounded by the constant M ;
4. there exists a constant K such that, for any t ∈ [0, T ], x′, x′′ ∈ Rd, u ∈ U , v ∈ V ,
d
dt
d
dt</p>
      <p>∥f (t, x′, u, v) − f (t, x′′, u, v)∥ ≤ K∥x′ − x′′∥;
5. there exists a function α : R → R such that α(δ) → 0 as δ → 0 and, for any t′, t′′ ∈ [0, T ], x ∈ Rd, u ∈ U ,
v ∈ V ,</p>
      <p>∥f (t′, x, u, v) − f (t′′, x, u, v)∥ ≤ α(t′ − t′′);</p>
    </sec>
    <sec id="sec-2">
      <title>6. the function σ is Lipschtiz continuous with the constant R;</title>
      <p>7. (Isaacs’ condition) for any t ∈ [0, T ], x, ξ ∈ Rd, u ∈ U , v ∈ V ,
mu∈iUn mv∈aVx⟨ξ, f (t, x, u, v)⟩ = max min⟨ξ, f (t, x, u, v)⟩.</p>
      <p>v∈V u∈U
Below, let p be an arbitrary function from [0, T ] × Rd to U . We will construct a stepwise control that is
N
constant between the times of control correction. Let t0 be an initial time, ∆ = {ti}i=0 be a partition of [t0, T ].
The times ti are times of control correction. If x0 ∈ Rd, v ∈ V[t0], then let x1[·, t0, x0, p, ∆, v] be a solution of
the problem</p>
      <p>x(t) = f (t, x(t), p(ti−1, x(ti−1), v(t)), t ∈ [ti−1, ti], i = 1, . . . , N, x(t0) = x0.</p>
      <p>Analogously, the strategy of the second player is determined by the function q : [0, T ] × Rd → V . If (t0, x0) is
an initial position, ∆ = {ti}iN=0 is a partition of [t0, T ], u ∈ U [t0], then denote by x2[·, t0, x0, q, ∆, u] a solution of
the problem
x(t) = f (t, x(t), u(t), q(ti−1, x(ti−1)), t ∈ [ti−1, ti], i = 1, . . . , N, x(t0) = x0.
Krasovskii and Subbotin proved that there exist functions p∗ : [0, T ] × Rd → U , q∗ : [0, T ] × Rd → V such that
lim sup{σ(x1[T, t0, x0, p∗, ∆, v]) : d(∆) ≤ δ, v ∈ V[t0]}
δ↓0
=
=</p>
      <p>inf lim sup{σ(x1[T, t0, x0, p, ∆, v]) : d(∆) ≤ δ, v ∈ V[t0]}
p∈U[0;T ]×Rd δ↓0</p>
      <p>sup lim inf{σ(x2[T, t0, x0, q, ∆, u]) : d(∆) ≤ δ, u ∈ U [t0]}
q∈V [0;T ]×Rd δ↓0
= lim inf{σ(x2[T, t0, x0, q∗, ∆, u]) : d(∆) ≤ δ, u ∈ U [t0]} = Val(t0, x0).</p>
      <p>δ↓0</p>
      <sec id="sec-2-1">
        <title>Here BA stands for the set of functions from A to B.</title>
        <p>The function Val is value function of the differential game. It is a viscosity/minimax solution of the
Hamilton</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Jacobi PDE</title>
      <p>∂ϕ
∂t
+ H(t, x, ∇ϕ), ϕ(T, x) = σ(x).</p>
    </sec>
    <sec id="sec-4">
      <title>Here</title>
      <p>H(t, x, ξ) , min max⟨ξ, f (t, x, u, v)⟩.</p>
      <p>u∈U v∈V
(2)</p>
      <p>Note that given a supersolution of (2) one can construct a suboptimal strategy of the first player. Analogous
result holds true for the second player. She is to use a subsolution of (2).
3</p>
      <p>General Theory of Markov Games
To evaluate the value function we will use zero-sum Markov game. In the general case a Markov game is defined
in the following way. Let S be a set. Assume that S is at most countable. Below S stands for the state space. For
any t ∈ [0, T ], u ∈ U , v ∈ V , let Q(t, u, v) = (Qx,y(t, u, v))x,y∈S be a Kolmogorov matrix i.e., for any t ∈ [0, T ],
u ∈ U , v ∈ V , x ∈ S,
and for any y ∈ S, y ̸= x
If u ∈ U [t0], v ∈ V[t0] be open-loop controls of the players, the state of the Markov chain at t ∈ [0, T ] is x, then
under some continuity conditions the probability of the state y ̸= x at time t+ &gt; t is
whereas the probability of keeping of x at [t, t+] is
∑ Qx,y(t, u, v) = 0
y∈S</p>
      <p>Qx,y(t, u, v) ≥ 0.</p>
      <p>Qx,y(τ, u(τ ), v(τ ))dτ + o(τ );</p>
      <p>Qx,x(τ, u(τ ), v(τ ))dτ + o(τ ).
d
dt
∫ t+</p>
      <p>t
1 +
∫ t+</p>
      <p>t
Recall that Qx,x(t, u, v) is always nonpositive. If we denote the probability of the state x at time t by Px(t), we
obtain the vector P (t) = (Px(t))x∈S . Below we assume that P (t) is a row-vector. The dynamics of the vector of
probabilities P (t) is given by the Kolmogorov equation</p>
      <p>P (t) = P (t)Q(t, u(t), v(t)), P (0) = P 0.
(3)
Here P 0 = (Px0)x∈S is an initial distribution.</p>
      <p>Assume that the first (receptively, second) player wishes to minimize (respectively, maximize) the terminal
payoff given by Eσ(X(T )), where X(t) is a state of the Markov chain corresponding to the Kolmogorov matrix
Q(t, u(t), v(t)) at time t.</p>
      <p>We do not put any analog of Isaacs condition on the Markov games. Thus, we are to suppose that one player
is informed about the current player of her partner or use mixed strategies. Within the paper we assume that
the first player has an information only on a current position, whereas the second player is informed about a
current position and a current control of the first player. In this case the strategy of the first player is u(t, x), the
strategy of the second player v(t, x, u). These strategies produce a Markov chain with the Kolmogorov matrix</p>
      <p>Qbx,y(t) , Qx,y(t, u(t, x), v(t, x, u(t, x))).</p>
      <p>Denote the corresponding Markov chain starting at (t∗, x∗) by Xt∗,x∗,u,v(·). The corresponding probability is
denoted by P t∗,x∗,u,v. One may introduce the upper value function by the rule:
η+(t∗, x∗) , min max Et∗,x∗,u,vσ(Xt∗,x∗,u,v(T )).</p>
      <p>u v</p>
      <sec id="sec-4-1">
        <title>Here Et∗,x∗,u,v stands for the expectation according to the probability P t∗,x∗,u,v.</title>
        <p>Under some regularity conditions the value function satisfies Zachrisson equation that is an analog of
Isaacs</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Bellman equation:</title>
      <p>d
dt
η+(t, x) + min max ∑ η+(t, y)Qx,y(t, u, v) = 0, η+(T, x) = σ(x), x ∈ S.</p>
      <p>u∈U v∈V y∈S
Note that (4) is a system of ODE. The existence result was established for the case of finite S in [Zachrisson, 1964].</p>
      <p>If a solution of the (4) exist one can define the strategies of the players as follows:
</p>
      <p>
u∗(t, x) , argmin mv∈aVx ∑ η+(t, y)Qx,y(t, u, v) ;
u∈U y∈S</p>
      <p> 
v∗(t, x, u) , argmax ∑ η+(t, y)Qx,y(t, u, v) .</p>
      <p>v∈V y∈S
Using the standard verification arguments, one can prove that the strategies u∗ and v∗ are optimal, i.e.
η+(t∗, x∗) = Et∗,x∗,u∗,v∗ σ(X(T, t∗, x∗, u∗, v∗)) = min Et∗,x∗,u,v∗ σ(X(T, t∗, x∗, u∗, v∗))</p>
      <p>u
= max Et∗,x∗,u∗,vσ(X(T, t∗, x∗, u∗, v)) = min max Et∗,x∗,u,vσ(X(T, t∗, x∗, u, v)). (5)</p>
      <p>v u v
4</p>
      <p>Approximating Infinite State Markov Chain
Given a differential game with dynamics (1) define the approximating Markov game in the following way.</p>
      <p>Let h be a positive number, f (t, x, u, v) = (f1(t, x, u, v), . . . , fd(t, x, u, v)) and let ei denote the i-th coordinate
vector. Put
χi(t, x, u, v) =  −eei,i,
 0,
fi1(t, x, u, v) &gt; 0,
fi1(t, x, u, v) &lt; 0,
fi1(t, x, u, v) = 0.</p>
      <sec id="sec-5-1">
        <title>Let the state space be S , hZd. Put the Kolmogorov matrix be equal to</title>
        <p>Qxhy(t, u, v) =
 h1 |fi(t, x, u, v)|,</p>
        <p>− h1 ∑id=1 |fi(t, x, u, v)|,
 0,
y = x + hχi(t, x, u, v),
x = y,
y ̸= x, y ̸= x + hχi(t, x, u, v),</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>In this case the Zachrisson equation (4) takes the form:</title>
      <p>ddt ηh+(t, x) + min max ∑d |fi(t, x, u, v)| ηh+(t, x + hχi(t, x, u, v)) − ηh+(t, x) = 0, ηh+(T, x) = σ(x).</p>
      <p>u∈U v∈V i=1 h
Here x ∈ hZd is a parameter.</p>
    </sec>
    <sec id="sec-7">
      <title>The following statement is proved in [Averboukh, 2016].</title>
      <p>Proposition 1. There exists an unique solution of (7).
(4)
(6)
(7)</p>
    </sec>
    <sec id="sec-8">
      <title>The following theorem is also proved in [Averboukh, 2016].</title>
      <p>Theorem 1. There exists a constant C1 determined by the function f such that if ηh+ is a solution of (7), then,
for t0 ∈ [0, T ], x0 ∈ hZd,
√
|Val(t0, x0) − ηh+(t0, x0)| ≤ RC1 h.
(8)
Note that in [Averboukh, 2016] it is proved that C1 = √2dM T e(3+2K)T .</p>
      <p>Further, the optimal strategies of the players take the form:
u∗(t, x) , argmin max ∑ |fi(t, x, u, v)| ηh+(t, x + hχi(t, x, u, v)) − ηh+(t, x) ] ;</p>
      <p>[ d
u∈U v∈V i=1 h
v∗(t, x, u) , argmax ∑ |fi(t, x, u, v)| ηh+(t, x + hχi(t, x, u, v)) − ηh+(t, x) ] .</p>
      <p>[ d
v∈V i=1 h
5</p>
      <p>Approximating finite state Markov chain
Assume that we are interesting in the approximation of the value function in the set</p>
    </sec>
    <sec id="sec-9">
      <title>To this end we consider the Markov chain with the state space Let</title>
      <p>Define the Kolmogorov matrix Q♮,h(t, u, v) = (Q♮x,,hy(t, u, v))x,y ∈ Λρ in the following way:
Λr ,
{</p>
      <p>}
x ∈ hZd : max |xi| ≤ r .</p>
      <p>i=1,d
Λρ ,
Γρ ,
{</p>
      <p>}
x ∈ hZd : max |xi| ≤ ρ .</p>
      <p>i=1,d
{</p>
      <p>}
x ∈ hZd : max |xi| = ρ .</p>
      <p>i=1,d
Q♮x,,hy(t, u, v) , { Qxh,y(t, u, v), x ∈ Λρ−h,</p>
      <p>0, x ∈ Γρ.
d
dt ηh+(t, x) = 0, x ∈ Γρ,
ηh+(T, x) = σ(x).</p>
      <p>We assume that the purposes of the player in this game is the same as above i.e. the first (respectively, second)
player wishes to minimize (respectively, maximize) the terminal payoff Eσ(Y (T )), where Y (t) denotes the state
of the corresponding Markov chain.</p>
      <sec id="sec-9-1">
        <title>For the Kolmogrov matrix Q♮,h(t, u, v) equation (4) takes the form:</title>
        <p>ddt ηh+(t, x) + min max ∑d |fi(t, x, u, v)| ηh+(t, x + hχi(t, x, u, v)) − ηh+(t, x)
u∈U v∈V i=1 h
= 0, x ∈ Λρ−h,
(9)
Note that system (9) is a finite system of ODEs. Thus, it can be solved using a numerical method.</p>
        <p>Let us estimate the difference between ηh(t, x) and η♮,h(t, x) for t ∈ [0, T ], x ∈ Λr. This estimate and
Theorem 1 will lead the approximation of the value function by the solution of the finite system of ODEs.</p>
        <p>If u(t, x) is a strategy of the first player, v(t, x, u) is a strategy of the second player, then denote the
corresponding Markov chain determined by the Kolmogorov matrix (Q♮x,,hy(t, u(t, x), v(t, x, u(t, x)))) and starting at (t∗, x∗)
by Y T∗,x∗,u,v. Further, denote by Pt∗,x∗,u,v and E t∗,x∗,u,v the corresponding probability and the expectation
respectively.</p>
        <p>Without loss of generality we can assume that P t∗,x∗,u,v and Pt∗,x∗,u,v are defined on the same measurable
space.</p>
      </sec>
    </sec>
    <sec id="sec-10">
      <title>Analogously, let</title>
      <p>We have that, for any event C such that C ∩ Atρ∗,x∗,u,v = C ∩ Bρt∗,x∗,u,v = ∅,
Lemma 1. For any t∗ ∈ [0, T ], x∗ ∈ Λr, and any strategies of the players u and v,</p>
      <p>Atρ∗,x∗,u,v = {Xt∗,x∗,u,v(t) ∈/ Λρ−h for some t ∈ [t∗, T ]}.</p>
      <p>Bρt∗,x∗,u,v = {Y t∗,x∗,u,v(t) ∈/ Λρ−h for some t ∈ [t∗, T ]}.</p>
      <p>P t∗,x∗,u,v(C) = Pt∗,x∗,u,v(C).</p>
      <p>P t∗,x∗,u,v(Atρ∗,x∗,u,v) ≤ d
P t∗,x∗,u,v(Bρt∗,x∗,u,v) ≤ d
Lt[u, v]φ(x) = ∑ Qx,y(t, u, v)φ(y).</p>
      <p>y∈S
d
dt</p>
      <p>Eφ(X(t)) = ELt[u(t), v(t)]φ(X(t)).</p>
      <p>Proof. We shall prove estimate (11). To this end given a controlled Markov chain with the Kolmogorov matrix
Q(t, u, v) let us introduce the generator acting on the continuous functions by the following rule:
It follows from [Kolokoltsov, 2011] that if X(t) is a Markov chain corresponding to the Kolmogorov matrix Q
and E is the corresponding probability, then</p>
      <p>For the Markov chain with the Kolmogorov matrix Qh(t, u, v) the generator takes the form:
Lth[u, v]φ =
1 d</p>
      <p>|fi(t, x, u, v)| ∑[φ(x + hχi(t, x, u, v)) − φ(x)].</p>
      <p>h i=1
Further, for x = (x1, . . . , xd), put ψi(x) , ∥xi∥2. Equality (14) implies that for any u, v</p>
      <p>Lt[u, v]ψ(x) = h|χi(t, x, u, v)|2 + 2xiχi(t, x, u, v)|fi(t, x, u, v)|.</p>
    </sec>
    <sec id="sec-11">
      <title>Therefore, (13) yields the following equation:</title>
      <p>d Et∗,x∗,u,v(Xit∗,x∗,u,v(t))2 = hEt∗,x∗,u,v|χi(t, Xit∗,x∗,u,v(t), u, v)|2
dt</p>
      <sec id="sec-11-1">
        <title>Here Xt∗,x∗,u,v(t) stands for the i-th coordinate of Xt∗,x∗,u,v(t).</title>
        <p>i
Since, |χi(t, x, u, v)| = 1, we have that</p>
        <p>+ Et∗,x∗,u,v2Xit∗,x∗,u,v(t)χi(t, Xit∗,x∗,u,v(t), u, v)|fi(t, Xit∗,x∗,u,v(t), u, v)|.
d Et∗,x∗,u,v(Xit∗,x∗,u,v(t))2 ≤ h + Et∗,x∗,u,v2|fi(t, Xit∗,x∗,u,v(t), u, v)||Xit∗,x∗,u,v(t)|.
dt</p>
        <sec id="sec-11-1-1">
          <title>Since fi is bounded by M we have that</title>
          <p>Thus,
d Et∗,x∗,u,v(Xit∗,x∗,u,v(t))2 ≤ h + 2M Et∗,x∗,u,v|Xit∗,x∗,u,v(t)|.
dt
d Et∗,x∗,u,v(Xit∗,x∗,u,v(t))2 ≤ h + M 2 + Et∗,x∗,u,v(Xit∗,x∗,u,v(t))2.
dt
(10)
(11)
(12)
(13)
(14)</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-12">
      <title>Therefore,</title>
    </sec>
    <sec id="sec-13">
      <title>Therefore,</title>
    </sec>
    <sec id="sec-14">
      <title>Using Gronwall’s inequality, we get that</title>
    </sec>
    <sec id="sec-15">
      <title>Using Markov inequality, we get</title>
      <p>Et∗,x∗,u,v(Xit∗,x∗,u,v(t))2 ≤ ∥x∗,i∥2 + (h + M 2)(t − t∗) +
Et∗,x∗,u,v(Xit∗,x∗,u,v(τ ))2dτ.
∫ t
t∗</p>
      <p>Et∗,x∗,u,v(Xit∗,x∗,u,v(t))2 ≤ (∥x∗,i∥2 + (h + M 2)(t − t∗))et−t∗ .
Et∗,x∗,u,v</p>
      <p>sup (Xit∗,x∗,u,v(t′))2 ≤
t′∈[t∗,T ]</p>
      <p>sup
t′∈[t∗,T ]</p>
      <p>Et∗,x∗,u,v(Xit∗,x∗,u,v(t′))2 ≤ (r2 + (h + M 2)T 2)eT .
P
(</p>
      <p>sup |Xit∗,x∗,u,v(t)| ≥ ρ
t∈[t∗,T ]
)
≤
.</p>
      <p>P (Atρ∗,x∗,u,v) ≤
∑ P
i
(</p>
      <p>sup |Xit∗,x∗,u,v(t)| ≥ ρ
t∈[t∗,T ]
)
≤ d
.</p>
    </sec>
    <sec id="sec-16">
      <title>Inequality (12) is proved in the same way.</title>
      <p>Theorem 2. There exists a constant C2 such that for any h one can choose ρ satisfying
|Val(t0, x0) − η♮,h(t0, x0)| ≤ C2h.</p>
      <p>Proof. We shall estimate |ηh(t∗, x∗) − η♮,h(t∗, x∗)|.</p>
      <p>Further, let Atρ∗,x∗,u,v (respectively, Btρ∗,x∗,u,v) be a complement of At∗,x∗,u,v (respectively, Bρt∗,x∗,u,v).
Using Lemma 1, we get
ηh(t∗, x∗) = min max Et∗,x∗,u,vσ(Xt∗,x∗,u,v)</p>
      <p>u v
≤ min max Et∗,x∗,u,vσ(Xt∗,x∗,u,v)1At∗;x∗;u;v + max max Et∗,x∗,u,vσ(Xt∗,x∗,u,v)1At∗;x∗;u;v
u v u v
≤ min max E t∗,x∗,u,vσ(Y t∗,x∗,u,v)1Bt∗;x∗;u;v + M d
u v
(r2 + (h + M 2)T 2)eT
ρ
≤ min max E t∗,x∗,u,vσ(Y t∗,x∗,u,v) + 2M d
u v
= η♮,h(t∗, x∗) + 2M d
.</p>
    </sec>
    <sec id="sec-17">
      <title>The opposite inequality is proved in the same way. Thus,</title>
      <p>|ηh(t∗, x∗) − η♮,h(t∗, x∗)| ≤ 2M d
.</p>
      <p>Once can choose ρ such that 2M d(r2 + (h + M 2)T 2)eT /ρ ≤ h. Using Theorem 1, we get the conclusion of the</p>
    </sec>
    <sec id="sec-18">
      <title>Theorem.</title>
      <p>Acknowledgements
This work was supported by the Russian Science Foundation (project no. 17-11-01093).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <source>[Averboukh</source>
          , 2016] Averboukh,
          <string-name>
            <surname>Y.</surname>
          </string-name>
          (
          <year>2016</year>
          ).
          <article-title>Approximate solutions of continuous-time stochastic games</article-title>
          .
          <source>SIAM J. Control Optim</source>
          .,
          <volume>54</volume>
          (
          <issue>5</issue>
          ):
          <fpage>2629</fpage>
          -
          <lpage>2649</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [Botkin et al., ]
          <string-name>
            <surname>Botkin</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hoffmann</surname>
          </string-name>
          , K.-H., and
          <string-name>
            <surname>Turova</surname>
          </string-name>
          , V. L.
          <article-title>Stable numerical schemes for solving hamiltonjacobibellmanisaacs equations</article-title>
          .
          <source>SIAM J. Control Optim</source>
          .,
          <volume>33</volume>
          (
          <issue>2</issue>
          ):
          <fpage>9921007</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <source>[Boue &amp; Dupuis</source>
          , 1999] Boue,
          <string-name>
            <given-names>M.</given-names>
            and
            <surname>Dupuis</surname>
          </string-name>
          ,
          <string-name>
            <surname>P.</surname>
          </string-name>
          (
          <year>1999</year>
          ).
          <article-title>Markov chain approximations for deterministic control problems with affine dynamics and quadratic cost in the control</article-title>
          .
          <source>Siam J. Numer Anal.</source>
          ,
          <volume>36</volume>
          :
          <fpage>667</fpage>
          -
          <lpage>695</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <source>[Buslaeva</source>
          , 1977] Buslaeva,
          <string-name>
            <surname>L. T.</surname>
          </string-name>
          (
          <year>1977</year>
          ).
          <article-title>Stochastic control in a differential game</article-title>
          .
          <source>J. Appl. Math. Mech.</source>
          ,
          <volume>42</volume>
          (
          <issue>4</issue>
          ):
          <fpage>609</fpage>
          -
          <lpage>623</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <source>[Camilli &amp; Marchi</source>
          , 2012] Camilli,
          <string-name>
            <given-names>F.</given-names>
            and
            <surname>Marchi</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          (
          <year>2012</year>
          ).
          <article-title>Continuous dependence estimates and homogenization of quasi-monotone systems of fully nonlinear second order parabolic equations</article-title>
          .
          <source>Nonlinear Anal.</source>
          ,
          <volume>75</volume>
          (
          <issue>13</issue>
          ):
          <fpage>5103</fpage>
          -
          <lpage>5118</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <source>[Elliott &amp; Kalton</source>
          , 1972] Elliott,
          <string-name>
            <given-names>R. J.</given-names>
            and
            <surname>Kalton</surname>
          </string-name>
          ,
          <string-name>
            <surname>N. J.</surname>
          </string-name>
          (
          <year>1972</year>
          ).
          <article-title>Values in differential games</article-title>
          .
          <source>Bull. Amer. Math. Soc.</source>
          ,
          <volume>78</volume>
          (
          <issue>3</issue>
          ):
          <fpage>427</fpage>
          -
          <lpage>431</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <source>[Jorgensen &amp; Zaccour</source>
          , 2006] Jorgensen,
          <string-name>
            <given-names>S.</given-names>
            and
            <surname>Zaccour</surname>
          </string-name>
          ,
          <string-name>
            <surname>G.</surname>
          </string-name>
          (
          <year>2006</year>
          ).
          <article-title>Developments in differential game theory and numerical methods: economic and management applications</article-title>
          .
          <source>Computational Management Science</source>
          ,
          <volume>4</volume>
          :
          <fpage>159181</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <source>[Kolokoltsov</source>
          , 2011] Kolokoltsov,
          <string-name>
            <surname>V. N.</surname>
          </string-name>
          (
          <year>2011</year>
          ).
          <article-title>Markov processes, semigroups and generators</article-title>
          , volume
          <volume>38</volume>
          of De Gruyter Studies in Mathematics. De Gryuter.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <source>[Kolokoltsov</source>
          , 2012] Kolokoltsov,
          <string-name>
            <surname>V. N.</surname>
          </string-name>
          (
          <year>2012</year>
          ).
          <article-title>Nonlinear Markov games on a finite state space (mean-field and binary interactions)</article-title>
          .
          <source>Int. J. Stat. Probab.</source>
          ,
          <volume>1</volume>
          (
          <issue>1</issue>
          ):
          <fpage>77</fpage>
          -
          <lpage>91</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <source>[Krasovskii &amp; Kotelnikova</source>
          , 2009] Krasovskii,
          <string-name>
            <given-names>N. N.</given-names>
            and
            <surname>Kotelnikova</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. N.</surname>
          </string-name>
          (
          <year>2009</year>
          ).
          <article-title>Unification of differential games, generalized solutions of the hamilton-jacobi equations, and a stochastic guide</article-title>
          .
          <source>Differ</source>
          . Equ.,
          <volume>45</volume>
          (
          <issue>11</issue>
          ):
          <fpage>1653</fpage>
          -
          <lpage>1668</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <source>[Krasovskii &amp; Kotelnikova</source>
          , 2010a]
          <string-name>
            <surname>Krasovskii</surname>
            ,
            <given-names>N. N.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Kotelnikova</surname>
            ,
            <given-names>A. N.</given-names>
          </string-name>
          (
          <year>2010a</year>
          ).
          <article-title>An approach-evasion differential game: stochastic guide</article-title>
          .
          <source>Proc. Steklov Inst. Math., 269 Supplement</source>
          (1 Supplement):
          <fpage>191</fpage>
          -
          <lpage>213</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <source>[Krasovskii &amp; Kotelnikova</source>
          , 2010b]
          <string-name>
            <surname>Krasovskii</surname>
            ,
            <given-names>N. N.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Kotelnikova</surname>
            ,
            <given-names>A. N.</given-names>
          </string-name>
          (
          <year>2010b</year>
          ).
          <article-title>On a differential interception game</article-title>
          .
          <source>Proc. Steklov Inst. Math.</source>
          ,
          <volume>268</volume>
          (
          <issue>1</issue>
          ):
          <fpage>161</fpage>
          -
          <lpage>206</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <source>[Krasovskii &amp; Subbotin</source>
          , 1973] Krasovskii,
          <string-name>
            <given-names>N. N.</given-names>
            and
            <surname>Subbotin</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. I.</surname>
          </string-name>
          (
          <year>1973</year>
          ).
          <article-title>Approximation in a differential game</article-title>
          .
          <source>J. Appl. Math. Mech.</source>
          ,
          <volume>37</volume>
          (
          <issue>2</issue>
          ):
          <fpage>185</fpage>
          -
          <lpage>192</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <source>[Krasovskii &amp; Subbotin</source>
          , 1988] Krasovskii,
          <string-name>
            <given-names>N. N.</given-names>
            and
            <surname>Subbotin</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. I.</surname>
          </string-name>
          (
          <year>1988</year>
          ).
          <article-title>Game-theoretical control problems</article-title>
          . Springer, New York.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <source>[Kushner</source>
          , 2007] Kushner,
          <string-name>
            <surname>H.</surname>
          </string-name>
          (
          <year>2007</year>
          ).
          <article-title>Numerical approximations for stochastic differential games: The ergodic cost criterion</article-title>
          . In Jorgensen, S.,
          <string-name>
            <surname>Quincampoix</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Vincent</surname>
          </string-name>
          , T. L., editors,
          <source>Advances in Dynamic Game Theory Numerical Methods</source>
          , Algorithms, and Applications to Ecology and Economics, pages
          <fpage>617</fpage>
          -
          <lpage>638</lpage>
          . Springer.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <source>[Kushner &amp; Dupuis</source>
          , 2001] Kushner,
          <string-name>
            <surname>H.</surname>
          </string-name>
          <article-title>and</article-title>
          <string-name>
            <surname>Dupuis</surname>
            ,
            <given-names>P. G.</given-names>
          </string-name>
          (
          <year>2001</year>
          ).
          <article-title>Numerical Methods for Stochastic Control Problems in Continuous Time</article-title>
          . Springer, New York.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <source>[Li &amp; Song</source>
          , 2007]
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Song</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          (
          <year>2007</year>
          ).
          <article-title>Markov chain approximation methods on generalized hjb equation</article-title>
          .
          <source>In Proceedings of the 46th IEEE Conference on Decision and Control</source>
          , pages
          <fpage>4069</fpage>
          -
          <lpage>4074</lpage>
          , New Orleans, LA, USA.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          <source>[Subbotin</source>
          , 1995] Subbotin,
          <string-name>
            <surname>A. I.</surname>
          </string-name>
          (
          <year>1995</year>
          ).
          <article-title>Generalized solutions of first-order PDEs. The dynamical perspective</article-title>
          .
          <source>Birkhauser</source>
          , Boston.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          <source>[Subbotin &amp; Chentsov</source>
          , 1981] Subbotin,
          <string-name>
            <given-names>A. I.</given-names>
            and
            <surname>Chentsov</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. G.</surname>
          </string-name>
          (
          <year>1981</year>
          ).
          <article-title>Optimization of guarantee in control problems</article-title>
          . Nauka, Moscow. in Russian.
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          <source>[Varaya &amp; Lin</source>
          , 1969] Varaya,
          <string-name>
            <given-names>P.</given-names>
            and
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <surname>J.</surname>
          </string-name>
          (
          <year>1969</year>
          ).
          <article-title>Existence of saddle points in differential games</article-title>
          .
          <source>Siam J. Control</source>
          ,
          <volume>7</volume>
          (
          <issue>1</issue>
          ):
          <fpage>142</fpage>
          -
          <lpage>157</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          <source>[Zachrisson</source>
          , 1964] Zachrisson,
          <string-name>
            <surname>L. E.</surname>
          </string-name>
          (
          <year>1964</year>
          ).
          <article-title>Markov games</article-title>
          . In Dresher,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Shapley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. S.</given-names>
            , and
            <surname>Tucker</surname>
          </string-name>
          , A. W., editors,
          <source>Advances in game theory</source>
          , pages
          <fpage>211</fpage>
          -
          <lpage>253</lpage>
          . Princeton University Press, Princeton.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>