An Introduction To Stochastic Control
An Introduction To Stochastic Control
F. J. Silva
Dip. di Matematica Guido Castelnuovo
March 2012
Some disperse but useful results
This section is based on [24, Chapter 1]. Let Ω be a nonempty set, and let F ⊂ 2Ω.
We say that
F is a π -system if A, B ∈ F ⇒ A ∩ B ∈ F .
1
π -system A, then they coincide on σ(A). In fact, it is enough to define C = {C ∈
F ; P (C) = Q(C)} and two verify that it is a λ-system.
Corollary 1. [Measurable function w.r.t. the σ -field of a π system] Let A be a π system.
Let H be a linear space of functions from Ω to R such that
with an analogous approximation for φ−. Therefore, by (iii), it is enough to show that
φn ∈ H. But φ is the sum of indicator of elements in σ(A). Therefore it is natural to
consider the set
F := {A ∈ Ω ; IA ∈ H}.
2
and to prove that σ(A) ⊆ F . But this follows from lemma 1, since A is a π -system and
A ⊂ F , which is easily shown to be a λ-system.
Theorem 1. [Dynkin theorem] Let (Ω, F ) and (Ω0, F 0) two measurable spaces, and let
(U, d) be a Polish space. Let ξ : Ω → Ω0 and φ : Ω → U be r.v.’s. Then φ ∈ σ(ξ)-
measurable, i.e. φ−1(B(U )) ⊆ ξ −1(F 0), iff there exists a measurable η : Ω0 → U
such that
φ(ω) = η(ξ(ω)) for all ω ∈ Ω.
Proof: Consider the case U = R (the general case can be obtained using an
isomorphism theorem, see [19]) and define the set
0
H := η(ξ) ; for some F 0-measurable map η : Ω → R .
We have to shown that the set of σ(ξ)/B(R) measurable maps belongs to H. This can
be done by checking the assumptions of corollary 1 with A = σ(ξ) and proving that H
satisfies (i), (ii) and (iii).
3
Lemma 2. [Borel-Cantelli] Let (Ω, F , P) be a probability space and consider a sequence
of events Ai ∈ F . We have
X
P(Ai) < ∞ ⇒ P(∩i ∪j≥i Aj ) = 0.
i
Proof: Straightforward. Note that P(∩i ∪j≥i Aj ) ≤ P(∪j≥iAj ) for all i and use the
convergence of the series.
4
Conditional expectation
This section is based on [24, Chapter 1]. Consider a probability space (Ω, F , P). Let
X ∈ L1(Ω) and G be a sub-σ -field of F . Define the signed measure µ : G → R as
Z
µ(A) := X(ω)dP(ω) for all A ∈ G.
A
By the Radon-Nikodym theorem there exists a unique P|G a.s. r.v. f ∈ L1G (Ω) (i..e. in
particular G -measurable) such that
Z Z
f dP = XdP for all A ∈ G. (2)
A A
E(X|G) := f.
5
Fundamental properties of E(·|G)
All the properties below are simple consequences of (2) (Do them as an exercise!)
(iv) [Take out the measurable part] For X ∈ L2F and Y ∈ L2G , we have
E(Y X|G) = Y E(X|G).
(vii) [Jensen inequality] Let φ convex such that φ(X) ∈ L1(Ω), then
φ(E(X|G)) ≤ E(φ(X)|G).
6
[Conditioning one r.v. X w.r.t. another r.v. ξ ]
Let X ∈ L1F and ξ : (Ω, F ) → (U, B(U )). Note that we can always define
−1
E(X|ξ) := E(X|ξ (B(U ))) = η(ξ) for some B(U )/B(R)- measurable function η,
E(X|ξ = x) := η(x).
Another way to define this is by appealing to the Radon-Nikodym. In fact, let us define in
B(U ) the measure Z
ν(B) := X(ω)dP(ω),
ξ −1 (B)
which is absolutely continuous with respect to Pξ := P ◦ ξ −1 (the image measure of P
dν
under ξ ). Therefore, there exists a Radon -Nikodym derivative dP ε
(unique Pε-a.s.) such
that
dν
Z Z
(x)dPε(x) = X(ω)dP(ω) for all B ∈ B(U ).
B dP ε ξ −1 (B)
7
We define
dν
E(X|ξ = x) := (x).
dPε
−1 dν
η(ξ(ω)) = E(X|ξ (B(U )))(ω) = (ξ(ω)) P|ξ−1(B(U )) − a.s. (3)
dPε
Integrating w.r.t. a set of the form ξ −1(A) and using the definition of Pξ , we obtain that
dν
η(x) = (x) Pε − a.s.
dPε
Note that incidentally, this gives another proof of Dynkin theorem. Let us prove (3). We
8
have
dν dν
R R
ξ −1 (B) dPε
(ξ(ω))dP(ω) = B dPε
(x)dPε(x)
= ν(B)
R
= ξ −1 (B)
X(ω)dP(ω)
E(X|ξ −1(B(U )))(ω)dP(ω).
R
= ξ −1 (B)
P(A|G) := E(IA|G).
Nota that for each B ∈ F we have that P(B|G) is only defined P|G a.s. Thus, it is not
sure that we can find some A ∈ G with P(A) = 1 such that if we fix any ω ∈ A, we
can measure all the sets B ∈ F . However, we can give a sense to this using the concept
of regular conditional probability (see [2] for more on this).
9
[Characterization of E(·|ξ) in terms of {g(ξ) ; g bounded continuous}] Let us now
prove that
The “only if” part is direct. To prove the “if part”, first note that
−1
A := {ξ ([a, b]) ; for some a < b}.
10
If A = ξ −1([a, b]) ∈ A we have
E(IA(ξ)X) = E(I[a,b](ξ)X).
Now, take any sequence of continuous functions g n → I[a,b](ξ), where the convergence is
pointwise. Since E(g n(ξ)X) = 0, by passing to the limit we get that E(IA(ξ)X) = 0
and so assumption (ii) of corollary 1 is verified.
Note that the same proof yields that if G = σ(ξ1, ..., ξn) we have
The interesting fact is that the result is also valid when we condition to a countable
set of r.v’s. More precisely, using the above result and the technique of its proof we get
the following result.
11
sequence of variables ξ1, ξ2, . . . and define G := σ({ξi ; i ∈ N}). Then
12
Stochastic process: Basic definitions
13
(i) [Symmetry condition] For any permutation σ of (t1, · · · , tn) we have
Question: Given a family of functions Ft1,...,ti (·, . . . , ·) satisfying the the symmetry and
compatibility condition, do we have an stochastic process with the desired finite-dimensional
distributions?
Theorem 2. [Kolmogorov’s existence theorem] Given a family Ft1,...,ti (·, . . . , ·)
satisfying (4) and (5), there exists a probability space (Ω, F , P) and a stochastic process
X whose finite dimensional distributions are the desired ones.
14
Given two process X and X 0 we say that X 0 is a modification of X if for all t ≥ 0
we have P(X(t) = X 0(t)) = 1. Note that the negligible sets depend on the time t.
We say that X 0 is a version of X if there exists A ∈ F with P(A) = 1 such that
X(ω, ·) = X 0(ω, ·) for all ω ∈ A.
Despite of the above example, it should be noted that modifications are in fact very
useful. For example, a modification of a process preserves its finite-dimensional laws.
15
[Measurability of process] A stochastic process X in [0, T ] is:
(iii) adapted, if for all t ∈ [0, T ] we have that X(t) is Ft/B(R) measurable.
(iii) progressively measurable (to simplify we will say “progressive”) if for all t ∈ [0, T ]
we have that X : [0, t] × Ω → R is B([0, t]) × Ft/B(R)-measurable.
Evidently, every progressive process is adapted. What happens with the converse? In
general, it is not true. However,
Proposition 2. [Adapted continuous process are progressive] If an adapted process X is
left or right continuous, then it is progressively measurable.
16
It is easy to see that X n are progressive and by the right continuity they converge to X
and so X is progressive.
For the general case, we have the following difficult result (for the proof see [17]) :
Proposition 3. [Progressive modifications] Every adapted process has a progressive
modification.
Now, we construct two natural filtrations in the space W[0, T ] := C[0, T ]. Set
17
Remark 1. Note that by definition W[0, T ] ∈/ B(Wt[0, T ]), this is one reason to consider
σ(B(Wt[0, T ])), the σ -field in W[0, T ], which by definition contains W[0, T ].
Note also that Bt(W[0, T ]) 6= Bt+ (W[0, T ]). In fact, the event
We denote by
A(U ) the set of Bt+ (W[0, T ])- progressive measurable process (6)
with values in a Polish space U , i.e. if f ∈ A(U ), the restriction of f to [0, t] × W[0, T ]
is [0, t] × Bt+ (W[0, T ])/B(U ) measurable.
The following theorem is the extension for filtrations of Dynkin theorem 1. For the
proof see [24].
18
Theorem 3. [Dynkin theorem for filtrations generated by a continuous process] Consider
(Ω, F , P) and a Polish space U . Let ξ : [0, T ] × Ω → R be a continuous process and
denote by Ftξ := σ(ξ(s) ; 0 ≤ s ≤ t) for its associated filtration. Then a process
φ : [0, T ] × Ω → U is {Ftξ }t≥0 adapted iff there exists η ∈ A(U ) such that
Nex we provide a sufficient condition for a.s. Holder continuity (up to modification of
the process) for a general stochastic process.
Then X has a modification which is α-Holder continuous, for every α ∈ (0, γ).
19
Stopping times
Let (Ω, F , {Ft}t≥0, P) be a filtered probability space satisfying the usual conditions.
Fτ := {A ∈ F ; A ∩ {τ ≤ t} ∈ Ft, ∀ t ≥ 0}.
Can we change ≤ by < in the definition of τ and Fτ ? The answer is yes because the
filtration is right-continuous (Exercise).
20
We will only list some interesting properties of stopping times. See [14] for the proofs.
σ + τ ; sup σi; inf σi; lim sup σi; lim inf σi.
i i i i
21
Proposition 5. [Tower property along stopping times] Let τ and σ be two stopping
times. Then
Proposition 6. [Stochastic process and stopping times] Suppose that the filtration
satisfies the natural conditions. Let X be progressive and τ an stopping time. Then
X(τ ) is Fτ measurable and the process X(τ ∧ t) is progressive.
22
Martingales
23
Fundamental properties
24
[Optional sampling theorem, or martingale property over stopping times] Let σ ≤
τ be two bounded stopping times. Then, for any martingale (resp. submartingale,
supermartingale) X(t) we have
25
Multivariate Normal distribution
−d
1 > −1 d
(2π detΣ) 2 exp − 2 (x − µ) Σ (x − µ) for all x ∈ R .
The particular form of the density yields that if Σ is diagonal then the coordinates of X
are one dimensional independent variables with normal distribution. (you can factorize the
multiple integrals in order to calculate the distributions).
Proposition 7. [Characterization of a multivariate gaussian] X is a multivariate
gaussian iff for every b ∈ Rd we have that b>X is a univariate gaussian.
26
Brownian motion
An adapted Rm valued process X(t) is called a Brownian motion over [0, ∞) if for
all 0 ≤ s < t we have that:
(iii) X(t) − X(s) is normally distributed with mean 0 and covariance matrix (t − s)I.
Therefore,
27
In particular, it is equivalent to say that we have m independents one dimensional Brownian
motions. In order to simplify the exposition we will work only with a one dimensional
Brownian motion, but we will come back to the multidimensional case a bit before treating
stochastic differential equations.
It is seen that W, endowed with this metric, is a Polish space. We say that C ⊆ W is a
cylinder in W if for j ∈ N and 0 ≤ t1 < . . . tj , there exists E ∈ B(Rj ) such that
28
We call C the set of cylinders.
which belongs to σ(C). Since W is separable, open sets are countably unions of balls and
thus B(W ) ⊆ σ(C).
Let µ be probability measure in (R, B(R)) and consider the density function of a
N (0, t),
1 |x|2
− 2 − 2t
f (x, t) := (2πt) e for t > 0, x ∈ R.
29
We define Pµ as follows: For any Ei ∈ B(R) we let
R R
Pµ({ξ ∈ W ; ξ(t1) ∈ E1, . . . ξ(tj ) ∈ Ej }) := µ(x ) E f (x1 − x0, t)dx1 . . .
R R 0 1
. . . R f (xj − xj−1, tj − tj−1)dxj .
(9)
It is seen that Pµ is additive. The difficult part consists in proving that it is σ -additive.
This was shown by Wiener in 1923. Using lemma 4 we can extend Pµ to (W, B(W)).
This extension of µ (which of course it is unique) is called a Wiener measure with initial
distribution µ. Then, on the filtered probability space (W, B(W), {Bt(W)}t≥0, Pµ), we
obtain a Brownian motion by defining
30
where E ∈ B(Rj ). We set B(R[0,∞)) for the σ -field generated by the cylinders. We
define a family of a posteriori finite dimensional-distributions replacing in (9) each Ei
by (−∞, xj ]. It is easy to show that this family satisfies (4) and (5). Therefore, by
applying theorem 2 we obtain a stochastic process X that has (9) as finite dimensional
distributions.
The only delicate point is that a priori there is no a.s. continuity. In fact, it is easy
to see that W ∈ / B(R[0,∞)) (we only have countably information which is not enough to
establish continuity, since R[0,∞) is too large) and that W contains only the empty set as
a measurable set belonging to B(R[0,∞)).
31
(iii) [Approximation by random walks] Given a family {Y (i) ; i = 1, . . . , n} of
independent r.v’s with P(Y (i) = 1) = P(Y (i) = −1) = 21 we define the symmetric
random walk as
k
X
M (0) := 0 and M (k) := Y (j) for k = 0, . . . , n.
j=1
Let us properly scale the process (recalling the Central Limit Theorem) defining
n 1
W (t) := √ M (nt) for all t ≥ 0.
n
Another proof of the existence of the Brownian motion, consists in proving the convergence
of this process to a process that satisfies the desired properties. This is the so-called
Donsker’s invariance principle (see [5]).
32
Set tk := k/n. Note that the following properties are clear (easy exercise!)
In particular we have
n 2 n n
(W (tk )) − [W , W ](tk ) is a discrete time martingale.
Return to the Brownian motion W (t): We now prove the analogous properties for the
limit case, i.e. for W (t).
E(W (t)|Fs) = E(W (t)−W (s)+W (s)|Fs) = W (s)+E(W (t)−W (s)|Fs) = W (s).
33
(ii) Let π n = (tni)i≥0 (t0 = 0) be a sequence of partitions of R+. Let ∆tni :=
tni+1 − tni and suppose that
n n
|π | := sup |∆ti | → 0 as n ↑ ∞.
i≥1
πn
X n n 2
QV (t) := |W (ti ∧ t) − W (ti−1 ∧ t)| .
i≥1
n
Proposition 8. We have that QV π (t) → t in L2(Ω), for all t ≥ 0.
Proof: We have
2 2
πn P n n 2 n n
E QV (t) − t = E i≥1 |W (ti ∧ t) − W (ti−1 ∧ t)| − (ti ∧ t − ti−1 ∧ t)
Writing the square, using the independence of the increments and the fact that
n n n n
W (ti ∧ t) − W (ti−1 ∧ t) ∼ N (0, ti ∧ t − ti−1 ∧ t),
34
we can eliminate the cross- terms and we obtain that
2
πn
h i
P n n 4 n n 2
E QV (t) − t = i≥1 E |W (ti ∧ t) − W (ti−1 ∧ t)| − (ti ∧ t − ti−1 ∧ t)
If a r.v. Y ∼ N (0, t) then E(Y 4 ) = 3t2 (Exercise! which is also useful to prove the a.s α- Holder
continuity of W (·) when α ∈ (0, 21 )). Thus
2
πn
X n n 2 n
E QV (t) − t =2 (ti ∧ t − ti−1 ∧ t) ≤ 2t|π | → 0.
i≥1
In fact, if the sequence of partitions has a mesh satisfying that π n converge fast
enough to 0 (for example if the mesh is given by the dyadic partition tni := i2−n) it is
n
possible to prove a.s. convergence for the whole trajectory QV π (·) (see [7]).
Before passing to the third identity, let us first discuss some important consequences
of the above result: First, from the inequality
πn n n
X
QV (t) ≤ max |W (tj ∧ t) − W (tj−1 ∧ t)| |W (ti ∧ t) − W (ti−1 ∧ t)|,
j≥1
i≥1
35
n
we see, by the continuity of the W (·), that V π (t) := |W (tni ∧ t) − W (tni−1 ∧ t)|
P
i≥1
n
satisfies V π (t) → ∞ a.s. This implies that a.s. the total variation of W (·) is +∞.
Therefore, it is not possible to define (see [21])
Z T
f (t)dW (t, ω) for all f ∈ C[0, T ].
0
36
[The augmented canonical filtration and the Blumenthal zero-one law] Clearly, by
the continuity of the trajectories, the natural filtration associated with W (·), given by
Ft = σ (W (s) ; s ≤ t), is left-continuous. However, it is easy to see that is not right-
continuous (think again on the event of having a local maximum). Therefore, it does not
satisfy the natural conditions. Now we are going to complete the filtration and then we
will show that the new filtration is right-continuous.
Proof: We will show that F0+ is independent to it self, which clearly implies the
result. Let 0 < t1 < ... < tk and g : Rk → R bounded continuous. For A ∈ F0+, by
continuity we get
For ε small enough, we obtain the independence between the above increments and Fε,
37
and so the increments are also independent of F0+. Thus
E(IA g(W (t1 ), . . . , W (tk )) = limε→0 P(A)E [g(W (t1 ) − W (ε), . . . , W (tk ) − W (ε)]
= P(A)E(g(W (t1 ), . . . , W (tk )).
Using the usual argument based in the monotone class theorem, we obtain that F0+
is independent of σ (W (s) ; 0 < s ≤ t) = σ (W (s) ; 0 ≤ s ≤ t) by continuity of
W (·), and so we get the independence to itself.
We define
Proof: The left continuity is clear. Clearly F 0 ⊆ F 0+ and by the above result
F 0+ ⊆ N (F ) ⊆ F0, which shows the right-continuity at zero. At the other times the
same argument applies by using the independence of the increments.
38
[Almost surely nowhere differentiable] As we have seen before the Brownian motion
has α-Holder continuous paths, for every α ∈ [0, 1/2). What happens for α ∈ [1/2, 1)?
Define the G(α, c, ε) as the set of all ω ∈ Ω such that for some s ∈ [0, 1] we have
α
|W (s, ω) − W (t, ω)| ≤ c|s − t| for all t with |s − t| ≤ ε. (11)
The proof of the following theorem is taken from the nice book [22].
Theorem 7. If α > 21 , then G(α, c, ε) has probability 0 for all 0 < c < ∞ and ε > 0.
Proof: First divide [0, 1] in intervals of length 1/n. When you are at k/n consider
m-blocks to the right (where m will be fixed and be chosen later!) . For m/n ≤ ε, set
j+1 j
Xn,k (ω) := max W ,ω −W ,ω ; k ≤j <k+m .
n n
If ω ∈ G(α, c, ε), then at least one block Bn,k̄ := [k̄/n, (k̄ + m)/n] is such that, for
the neighborhood where (11) holds, we have that Bn,k̄ ⊆ [s − ε, s + ε]. By the triangular
39
inequality and (11) we get that Xn,k̄ (ω) ≤ 2c(m/n)α. Therefore, we have
α
G(α, c, ε) ⊆ ω; min Xn,k ≤ 2c(m/n) .
0≤k≤n−m
On the other hand, by the properties of the increments Brownian motion W (t)
(independence and stationarity), we get:
α 1 α m
P (Xn,k ≤ 2c(m/n) ) = P W n ≤ 2c(m/n) m
1
−2
= P n |W (1)| ≤ 2c(m/n)α .
√
Using the trivial bound P(|B1| ≤ x) ≤ 2x/ 2π , we easily obtain
m
√
1
α α −α
P (Xn,k ≤ 2c(m/n) ) ≤ 4cm n 2 / 2π .
40
Thus,
m
4mα
1
α 1+m( 2 −α)
P(G(α, c, ε)) ≤ nP (Xn,k ≤ 2c(m/n) ) ≤ √ n .
2π
As a corollary, we get
Corollary 2. With probability one every path of the Brownian motion is not differentiable
at any 0 ≤ t ≤ 1.
41
[Anecdotic remark] Weierstrass in 1872 constructed the following function which is
continuous and nowhere differentiable
X cos(3nt)
f (t) := .
n≥1
2n
Thus we have answered to the question of the α-Holder property for α ∈ ( 21 , 1).
What happens with the limit case α = 12 ? The following important theorem gives, as a
corollary, a negative answer for α = 12 .
Theorem 8. [Law of the iterated logarithm] For the Brownian motion W (t) we have,
with probability 1, that
W (t) W (t)
lim sup q = 1 and lim inf q = −1.
t↓0
t↓0 2t log[log( 1t )] 2t log[log( 1t )]
42
[The strong Markov property] We recall that by definition of W (t) the process
W (t + ·) − W (t) is independent of W (t). Our aim now is to generalize this property
to the case of stopping times.
Theorem 9. Let τ be a stopping time. Consider the process
(τ )
W (t) := W (τ + t) − W (t).
Proof: Note that, by the monotone class argument, it is enough to prove that
h i
(τ ) (τ )
E IAg(W (t1), . . . , W (tp)) = P(A)E [g(W (t1), . . . , W (tp))]
for all 0 ≤ t1 < . . . , tp and any continuous bounded g . Let [τ ]n be the smallest real
number of the form k2−n which is greater than τ . Clearly
h i h i
(τ ) (τ ) ([τ ]n ) ([τ ]n )
E IAg(W (t1), . . . , W (tp)) = lim E IAg(W (t1), . . . , W (tp)) .
n→∞
43
But
h i
([τ ]n ) ([τ ]n )
E IA g(W (t1 ), . . . , W (tp ))
h i
P −n −n −n −n
= k≥0 E IA g W (k2 + t1 ) − W (k2 ), . . . , W (k2 + t1 ) − W (k2 ) ,
k
where Ak := A ∩ {(k − 1)2−n < τ ≤ k2−n}. Since τ is a stopping time, we have that
Ak is Fk2−n -measurable. Thus, by the independence of the increments of W , we get
h i
([τ ]n ) ([τ ]n ) P
E IA g(W (t1 ), . . . , W (tp )) = E g W (t1 ), . . . , W (tp ) k≥0 P(Ak ),
= P(A)E g W (t1 ), . . . , W (tp ) ,
44
Stochastic Integral
Let (Ω, F , {Ft}t≥0, P) be a filtered probability space satisfying the usual conditions.
We have that L2F ([0, T ]; R) is a Hilbert space. For simplicity, when the context is clear,
we write only L2F .
45
M2([0, T ]; R) is the set of Ft-square integrable (i.e. in L2F ([0, T ]; R)) martingales
with right-continuous paths starting from 0. We set
Theorem 10. M2([0, T ]; R) endowed with (12) is a Hilbert space and M2c ([0, T ]; R)
is closed in M2([0, T ]; R).
Using this theorem, in particular the closeness property of M2c ([0, T ]; R), we will define
the stochastic integral w.r.t. W (t),
46
Proof of theorem 10: The fact that k · kM 2 is a norm is easy. We now prove that the
space is complete. Let Xn ∈ M2 be a Cauchy sequence. Choosing a subsequence nk such that
kXnk+1 − Xnk kM 2 ≤ 2−3k , Doob inequality (I) yields
!
−k 2k 2 −k
P sup |Xnk+1 (t) − Xnk (t)| > 2 ≤2 E |Xnk+1 (T ) − Xnk (T )| ≤2 .
t∈[0,T ]
P
By the Borel-Cantelli lemma (if i P(Ai ) < +∞ then P(∩i ∪j≥i Ai ) = 0) we get the existence of
Ω0 ∈ F and X right-continuous, such that for all ω ∈ Ω0 we have supt∈[0,T ] |Xnk (t) − X(t)| → 0.
By dominated convergence and Doob inequality (II), we get that Xnk (t) → X(t) in L2 (Ω) for all
t ∈ [0, T ]. By passing to the limit in the martingale equality we get that X ∈ M2 . Finally, if Xnk is
continuous we have that X is continuous, because of the a.s. uniform convergence of the subsequence.
47
Define the stochastic integral If (ω, ·) : [0, T ] → R as
X
If (ω, ·) := fi(ω) [W (ω, · ∧ ti+1) − W (ω, · ∧ ti)] .
i≥0
(i) For all ω ∈ Ω, the continuity of W (ω, ·) yields the continuity of If (ω, t).
hR i
2 T 2
(iii) [Ito isometry] kIf (·)kM 2 = E(If (T ) ) = E 0
|f (t)| dt = kf kL2 .
48
Proof:
h i
2 P 2 2
E(If (T ) ) = i≥0 E |f (t
i )| (W (ti+1 ) − W (ti ))
P
+2 i<j E f (ti )f (tj )(W (ti+1 ) − W (ti ))(W (tj+1 ) − W (tj )) .
Conditioning by Fti each term of the first sum and by Ftj each term of the second sum we get the result.
Properties (i)-(iii) imply that If ∈ Mc2 and that the linear application f ∈ L2F →
If ∈ Mc2 is an isometry. Therefore we can extend If to f ∈ clos(L0) (the closure of L0
in L2F ). But... what is clos(L0)? Note that if f ∈ L2F is bounded and a.s. continuous,
we have that
n −1
2X
iT 2
fn(·) := f I iT (i+1)T
(·) → f (·) in LF .
i
2n 2n
,
2n
49
Therefore, for every f ∈ L2F we have defined its stochastic integral If (·) ∈ M2c ,
which is usually denoted by
Z ·
If (·) =: f (s)dW (s).
0
Proof of lemma 5:
(ii) If f is bounded progressively measurable we can approximate first by the following continuous
adapted (and thus progressively measurable) process
Z t
k
F (t) := k f (s)ds., (13)
0∨(t− 1 )
k
(iii) If f is bounded adapted, it has a progressive measurable modification and the analogous process
to (13) is progressive measurable. Using the argument based on Fubini theorem (see [14]) we can prove that
F k (t) is adapted and then can proceed as in (ii).
50
(iv) If f is only measurable and adapted we truncate.
(i) One can think to approach a progressive process by sums as in the usual Lebesgue
theory (see for example the sum in (1)). However, it is clear that it does not work since
an event like
−n −n
{(t, ω) ; f (t, ω) ∈ [j2 , (j + 1)2 }
belongs only to B([0, T ]) × FT and thus we cannot construct elements in L0 of this
kind, which implies the loose of the isometry property, the martingale property, etc.
(ii) Note that in the construction of the integral, we can avoid the use of the
normal distribution of the increments as follows: We have used this in the Itô isometry
property for the simple process. But the same result can be obtained, using that
W (t)2 − t = W (t)2 − [W, W ](t) is a martingale (exercise!). This remark allows to
define the stochastic integral for arbitrary continuous square-integrable martingales M (t).
In fact, there is a deep result called the Doob-Meyer decomposition that says that for such
51
a martingale M the exists a process [M, M ](t) (the quadratic variation of M ) such that
2
M (t) − [M, M ](t) is a martingale.
where "Z #
T
2
kxkM,2 := E |X(t)| d[M, M ](t) .
0
It can be shown that if the measure induced by [M, M ](·) is absolutely continuous
w.r.t. the Lebesgue measure (which is the case for the Brownian motion), then we can
approximate with simple functions any element of L2M . However, for the general case we
can approximate only any progressive element of L2M . For more information on this see
the books [12, 14].
52
Fundamental properties of the stochastic integral Consider f , g ∈ L2F and s < t
2 hR i
Rt t 2
E f (r)dW (r) |Fs
s
= E s
|f (r)| dr|Fs ,
R · 2 R ·
0
f (r)dW (r) − 0 |f (r)|2dr is a martingale.
Proof: The first identity follows analogously to the derivation for the unconditioned Ito isometry. In fact,
the same proof yields the identity for simple process and then we can pass to the limit in the conditional
expectation.
R· The second identity comes from a direct computation using the first identity. In fact, using
that 0 f (r)dW (r) is a martingale we get that
"Z
t 2 # "Z
t 2 # Z s 2
E f (r)dW (r) |Fs =E f (r)dW (r) |Fs − f (r)dW (r) ,
s 0 0
53
(ii) [Optional Sampling] For stopping times σ ≤ τ
hR i
t∧τ
E s∧σ f (r)dW (r)|Fσ = 0,
2 hR i
R t∧τ t∧τ 2
E s∧σ
f (r)dW (r) |Fσ = E s∧σ |f (r)| dr|Fσ ,
hR R i hR i
t∧τ t∧τ t∧τ
E s∧σ
f (r)dW (r) s∧σ
g(r)dW (r) |Fσ = E s∧σ f (r)g(r)dr|Fσ
Proof: The first two identities are a direct consequence of the optional sampling theorem and (i). For the
third identity, note that by the second identity we have
"Z
t∧τ 2 # Z t∧τ
2
E [f (r) + g(r)]dW (r) |Fσ = E [f (r) + g(r)] dr|Fσ ,
s∧σ s∧σ
developing the squares and using the second identity again, we get the result.
(ii) [Consistency] For any stopping time σ and f ∈ L2F , let fˆ(t, ω) :=
f (t, ω)Iσ(ω)≥t. Then
Z t∧σ Z t
f (s)dW (s) = fˆ(s)dW (s).
0 0
54
Proof: The proof of this fact follows easily from our construction (Exercise!).
Now, we sketch the construction of the integral for a more general kind of process.
Define
( Z T )
2,loc 2
LF := X : [0, T ] × Ω → R ; X is Ft adapted and |X(t)| dt < ∞, P − a.s.
0
n o
2,loc 2,loc
Mc := X ∈M ; X is continuous .
55
argument. In fact, for every j ≥ 1 define the stopping time
Z t
2
σj (ω) := inf t ∈ [0, T ] ; |f (s, ω)| ds ≥ j .
0
Define fj (t) := f (t)It≤σj . By definition of the stopping times we have that fj ∈ L2F .
We define the stochastic integral of f as
Z t Z t
f (s)dW (s) := fj (s)dW (s) if t ∈ [0, σj ]
0 0
We have to check that is well defined, because if t ≤ σi(ω) ≤ σj (ω) we have to verify
that Z t Z t
fj (s)dW (s) = fi(s)dW (s).
0 0
To see this, note that by the property of consistency, we have
Z t∧σ Z t Z t Z t
i
fj (s)dW (s) = fj (s)Is≤σi dW (s) = f (s)Is≤σi dW (s) = fi (s)dW (s).
0 0 0 0
56
R·
It can be proved that 0
f (s)dW (s) is a local martingale, but in general is not a
martingale.
Note that for r = 1 the first inequality is trivial and the second inequality is Doob
inequality (II) with p = 2.
57
Ito’s formula
Note that if “standard” differential calculus were valid, when consider the stochastic
integral one would expect that
Z t
2
W (t) = 2 W (s)dW (s).
0
Let us check that this formula is wrong! In fact, evaluating in t 6= 0 and taking the
expectation we obtain t = 0!. Let us find the correct formula with the definition of the
stochastic integral. Let us consider a sequence of partitions 0 = tn0 < tn1 < . . . < tnn = t
such that maxi≥1 |tni − tni−1| → 0 as n ↑ ∞. Since W (·) has continuous trajectories
Rt
we can approximate 2 0 W (s)dW (s) as the limit of
Pn h i Pn h 2 i P h i2
2 n
i=1 2W (ti − 1) W (ti ) − W (ti−1 ) = i=1 W (ti ) − W (ti−1 ) − i=1 W (ti ) − W (ti−1 )
Pn h i2
= 2
W (t) − i=1 W (ti ) − W (ti−1 ) .
58
By taking the L2 limit we get
Z t
2
2 W (s)dW (s) = W (t) − t.
0
Now we provide the following change of variable formula better known as Itô’s formula.
Now, we comeback to the setting where the Brownian motion is m-dimensional.
It is important to fix now the following notation that will be used throughout the
0
notes. For a smooth function f : R+ × Rm → R (m0 ∈ N), we set
2 2
∂tf (t, x) := Dtf (t, x), Df (t, x) := Dxf (t, x), D f (t, x) := Dxxf (t, x)
59
For the proof of this important result see [12]. It is important to note that, since the
result holds with probability one, then we can replace t by any stopping time τ .
Now we extend the result to the so-called Itô’s process. For adapted measurable
processes b(t) and σ(t) with values in Rn and Rn×m respectively, satisfying that
Z T Z T
2
|b(s)|ds + |σ(s)| dt < ∞
0 0
A process of this type is called an Itô’s process. For these process we have the following
Itô’s formula:
Theorem 13. Let f : R+ × Rn → R be C 1,2([0, T ]; Rn). Then, with probability one,
we have
Z t Z th
> 1 Tr σ(s)σ(s)> D 2 f (s, X(s)) dt.
h ii
f (t, X(t)) = f (0, 0) + Df (s, X(s)) dX(s) + ∂t f (s, X(s)) + 2
0 0
60
Stochastic Differential Equations (SDEs)
We say that an adapted, continuous process X(t) is a solution of (14) if a.s. we have
61
(Note that by the lemma 6 below the process b(t, X(·)) and σ(t, X(·)) are adapted)
Lemma 6. Let b ∈ A(R). Let (Ω, F , {F }t≥0, P) be given satisfying the usual
conditions and let X be continuous and progressive. Then the process (t, ω) →
b(t, X(·, ω)) is progressive.
Proof: We consider the application (t, ω) → Φ(t, ω) = (t, X(·, ω)) ∈ [0, T ]×W.
Therefore,
b(t, X(·, ω)) = b ◦ Φ(t, ω).
Since b ∈ A(R) we have that b−1(B(R)) ⊆ B([0, t]) × Bt+(W). Thus, since
X −1(Bt+(W)) ⊆ Ft+ = Ft (note that it is enough to verify it on the cylinders defined
in (8)) we obtain immediately the result.
62
(H) [Lipschitz conditions for the coefficients] There exists a constant L > 0 such that
for all t ∈ [0, ∞), and x, y in W.
kx(· ∧ j̄) − y(· ∧ j̄)kC[0,j] ≤ kx(· ∧ j̄) − y(· ∧ j̄)kC[0,j̄] for all j ∈ N.
63
where " #
` `
kxk`,∞ := E sup |X(t)| .
t∈[0,T ]
It is easy to show that L`F (Ω, C([0, T ]; Rn)) is a Banach space (exercise!).
Theorem 14. [Existence and uniqueness] Under the assumption (H), for all ` ≥ 1 and
ξ0 ∈ L`(Ω, Rn), there exists a unique solution X(t) of (14). Moreover, for all T > 0
we have that X ∈ L`F (Ω, C([0, T ]; Rn)).
Proof: We will use a fixed point technique. First, let us fix a fixed deterministic time
τ , to be chosen later, and consider the space Sτ := L`F (Ω, C([0, τ ]; Rn)). We define
the application T : Sτ → Sτ by
Z t Z t
T (x)(t) := ξ0 + b(s, X(·))ds + σ(s, X(·))dW (s) for all t ∈ [0, τ ].
0 0
If the application is well defined and is contractive, then we get our solution in [0, τ ] and
of course using the same argument we can extended to [τ, 2τ ], by taking the logical initial
condition in order to get continuity, etc. Therefore, it is enough to prove that
64
(i) T is well defined . We easily obtain the existence of a constant L1 > 0 such that
`
` !
Z τ Z t
` `
|T (x)(t)| ≤ L1 |ξ0| + |b(s, X(·))|ds + sup σ(s, X(·))dW (s) .
0 t∈[0,τ ] 0
(15)
The BDG inequality gives
" Z t `
# Z τ `
2
2
E sup σ(s, X(·))dW (s) ≤ K` E |σ(s, X(·))| ds
t∈[0,τ ] 0 0
" Z t `
# Z τ `
2 `
2 `
E sup σ(s, X(·))dW (s) ≤ L2 |σ(s, 0)| ds + τ 2 kxk`,∞ < ∞
t∈[0,τ ] 0 0
65
Similarly, there exists L3 > 0 such that
"Z ` # "Z ` #
τ τ ` `
E |b(s, X(·))|ds ≤ L3 |b(s, 0)|ds + τ 2 kxk`,∞ < ∞.
0 0
Thus, by taking the supremum and then the expectation in (15) we obtain that T is well
defined.
66
By (H) we obtain the existence of L6 > 0 such that
`
" #
Z t ` `
E sup [σ(s, X(·)) − σ(s, y(·))] dW (s) ≤ L6τ 2 kx − yk`,∞
t∈[0,τ ] 0
Therefore, by taking the supremum in (16), we finally obtain the existence of L8 > 0 such
that ` `
kT (y) − T (x)k`,∞ ≤ L8τ 2 kx − yk`,∞.
2
Letting τ < min{1, ( L1 ) ` }, we get the result.
8
67
[Markov solutions] Suppose that we are given b : [0, ∞) × Rn → Rn and
σ : [0, ∞) × Rn → Rn×m such that
(H’) [Lipschitz conditions for the coefficients (bis)] There exists a constant L > 0 such
that for all t ∈ [0, ∞), and x, y in W.
Clearly, this is a special case of the one studied before. Moreover, it is clear that (H’)
implies (H) when b and σ are viewed as elements in A(Rn) and A(Rn×m) respectively.
This particular structure is very important since the solutions are Markovian, i.e. satisfy
the Markov property.
68
Theorem 15. Under (H’) there exists a unique solution X of (17). Moreover, X(t)
is a Markov process, i.e.
Moreover, X(t) is a strong Markov process, i.e. for every stopping time τ , we have
Idea of the proof: We only need to prove the Markov and strong Markov property. We
will argue in a formal way (see [14] for a rigorous proof). Note that
Z t+h Z τ +t
X(t + h) = X(t) + b(s, X(s))ds + σ(s, X(s))dW (s).
t τ
69
form
Z h Z h
0 0 0 0 0 0 0 0 0 0 0 0
X (h) = X (0) + b (h , X (h ))dh + σ (h , X (h ))dW (h ).
0 0
Since W 0(h) is a Brownian motion w.r.t. the filtration Fh0 (and independent of Fs by the
Markov property), the above SDE is well posed and its solution depend only on X 0(0) .
By path uniqueness of the solutions this implies the result. For the strong Markov property
the previous ideas work in the same way if we have autonomous coefficients, noting that
the new W 0(h) := W (τ + h) − W (τ ) is a Brownian motion independent of Fτ by the
strong Markov property. Otherwise, we can add an artificial variable (dXn+1 = 1dt), to
treat the time as a part of X .
70
Proposition 9. [Important stability estimates] Suppose that (H) holds true. Then,
for all T > 0 there exists a constant KT such that the unique solution X(t) of (14)
satisfies:
(i)
" #
` `
E sup |X(s)| ≤ KT 1 + E(|ξ| ) .
0≤s≤T
(ii)
`
`
`
E |X(t) − X(s)| ≤ KT 1 + E(|ξ| ) |t − s| 2 .
(iii) If ξ̂ ∈ L`F0 (Ω; Rn) is another r.v. and X̂(t) is the corresponding solution, then
!
` `
E sup |X(s) − X̂(s)| ≤ KT E |ξ − ξ̂| .
0≤s≤T
71
Lemma 7. [Gronwall lemma] Let f : [a, b] → R be piecewise continuous, satisfying
Z t
f (t) ≤ α + β f (s)ds, for all t ∈ [a, b], (18)
a
β(t−a)
f (t) ≤ αe .
72
Proof of proposition 9: We will prove the result for ` ≥ 2. The proof for the general case can be
found in [18].
Proof of (i): Given T > 0 and t ∈ [0, T ], arguing as in the first part of the proof of theorem 14 we
easily obtain
Z t ` Z s `
!
` `
|X(t)| ≤ LT |ξ| + |b(s, X(r))|ds + sup σ(r, X(r))dW (r)
0 0≤s≤t 0
!`
" # !
Z s ` Z T 2 Z t
2 `
sup σ(r, X(r))dW (r) ≤C |σ(s, 0)| ds + sup |X(r)| ds ,
E E
0≤s≤t 0 0 0 0≤r≤s
for some C > 0. In what follows C will always denote a generic constant. Clearly,
!` Z
"Z
t ` # Z T t
!
`
E |b(s, X(s))|ds ≤C |b(s, 0)|ds + E sup |X(r)| ds .
0 0 0 0≤r≤s
73
Therefore, we finally obtain
! " ! #
Z t
` ` `
E sup |X(s)| ≤ C 1 + E |ξ| + E sup |X(r)| ds ,
0≤s≤t 0 0≤r≤s
Using this fact, majoring X(r) by sup0≤t≤T X(r) in (19) and using (i), we easily obtain the result.
74
Stochastic equations with random coefficients
(i) For every ω ∈ Ω, we have that b(·, ·, ω) ∈ An(Rn) and b(·, ·, ω) ∈ An(Rn×m).
75
(iii) There exists a constant L > 0 such that for all t ∈ [0, ∞), ω ∈ Ω, and x, y in W.
Theorem 16. [Existence and uniqueness for SDEs with random coefficients] Under
the above assumptions there exists a unique solution of (20). Moreover, the obvious
estimates (uniforms in ω ), of proposition 9 hold true.
Proof: It is a direct adaptation of the proof for the non-random coefficient case (Exercise!).
76
Connections with PDEs
Itô’s formula implies that A is well defined for every C 1,2-function with bounded derivatives
and h i
> 1 > 2
Af = b(t, x) Df (t, x) + 2 Tr σσ (t, x)D f (x, t) . (21)
77
Proposition 10. Assume that (t, x) → v(t, x) := E g X (T ) is C 1,2([0, T ] ×
t,x
∂tv + Av = 0,
v(T, ·) = g(·).
Proof: If the function v has bounded derivatives the proof below simplifies considerably (exercise).
To treat the case when only v ∈ C 1,2 , we use a typical localization in space technique in order
to be able to eliminate the diffusion term by taking expectation. In fact, define the stopping time
τ1 := T ∧ inf{s > t ; |X t,x (s) − x| ≥ 1}. Itô’s formula gives
t,x
R s∧τ
v s ∧ τ1 , X (s ∧ τ1 ) = v(t, x) + t 1 (∂t v + Av) (r, X t,x (r))dr
R s∧τ
+ t 1 Dv(s, X t,x (r))> σ(r, X t,x (r))dW (r).
78
But, by the strong Markov property we obtain
s∧τ1 ,X t,x (s∧τ1 )
t,x
v s ∧ τ1 (ω), X (s ∧ τ1 (ω), ω) = E g X (T ) |Fτ1 (ω).
Therefore,
h i
t,x x,t
v s ∧ τ1 (ω), X (s ∧ τ1 (ω), ω) = E g X (T ) |Fτ1 (ω).
79
We suppose that: (i) b and σ are continuous, Lipschitz in x uniformly in t and |b(·, x)|,
|σ(·, x)| belong to L2([0, T ], (ii) the function k is uniformly bounded from below, (iii)
the function f has quadratic growth in x uniformly in t.
"Z #
T
t,x t,x t,x t,x
v(t, x) = E β (s)f s, X (s) .s + β (T )g(X (T )) ,
t
t,x
Rs
where β (s) := exp{− t
k(r, X t,x(r))dr}, with X x,t(s) being the solution of
80
Proof: Define the sequence of stopping times
n o
t,x
τn := T ∧ inf s > t ; |X (s) − x| ≥ n .
By the continuity of the paths of X t,x (s) it is clear that a.s. we have τn ↑ T as n ↑ ∞. Using that v is
smooth, we can apply Itô’s formula to β t,x (s)v(s, X x,t (s)) obtaining that
h i
t,x x,t t,x x,t
d β (s)v(s, X (s)) = β (s) [−kv + ∂t v + Av] s, X (s) ds
t,x x,t x,t
+β (s)Dv s, X (s) σ s, X (s) dW (s).
Using that v solves (22) and taking the expected value, we obtain
h i h R τn t,x i
t,x x,t x,t
E β (τn )v(τn , X (τn )) − v(t, x) = E − t β (s)f s, X (s) ds
hR i
τn t,x x,t x,t
+E t β (s)Dv s, X (s) σ s, X (s) dW (s) .
Since before τn , X x,t (s) is bounded, the last term in the above expression is zero. Therefore, we have
Z τn
t,x x,t t,x x,t
v(t, x) = E β (s)f s, X (s) ds + β (τn )v(τn , X (τn )) .
t
81
The result easily follows by letting n ↑ ∞ by Lebesgue theorem. In order to verify that the integrand is
dominated we use the quadratic growth property of f , the estimates for the second moment of X x,t (s) and
the fact that k is bounded by below.
82
Stochastic control: Problem formulation
Excellent references for what follows are the books [11, 20, 24] and the lecture notes
[23].
Let (Ω, F , {F }t≥0, P) be a given probability space satisfying the usual conditions.
Let W be a Brownian motion defined on this space. We suppose that the filtration is given
by the canonical augmentation of the one generated by W (t). For (t, x) ∈ [0, T ] × Rn,
let us consider the following controlled stochastic SDE
83
where U ⊆ Rd is closed. For u ∈ U we denote by X t,x[u] for the unique solution (if it
exists!) of (23). Given a function ` : [0, T ]×Rn ×Rd → R and g : Rn → R, we consider
the cost (which is well defined under the assumptions below) J : [0, T ] × Rn × U → R
"Z #
T
t,x t,x
J(t, x, u) := E `(s, X [u](s), u(s))ds + g(X [u](T )) .
t
The stochastic optimal control problem at (t, x) is to calculate the following value function
84
(i) [Lipschitz assumption] There exists a constant L > 0 such that for φ(t, x, u) =
b(t, x, u), σ(t, x, u), `(t, x, u), h(x), we have
Under these assumptions, for each u ∈ U , equation (23) admits a unique solution
X t,x[u] and the function J is well defined. For simplicity, we have considered these
assumptions. However, most of the results presented later can be extended to more general
settings as:
(iii’) Instead of a uniform bound over b(t, 0, u) and σ(t, 0, u) you only ask for linear
growth for b(t, x, u) and σ(t, x, u) .
(iv’) Instead of a uniform bound over `(t, 0, u), you only ask for quadratic growth for
85
`(t, x, u) and g(x).
For these assumptions see the lecture notes of N. Touzi [23] and the books [11, 20].
[An interesting reduction of the set of admissible controls] Let us consider the set
Ut := u ∈ U ; u|[t,T ] is independent of Ft .
We have:
Proposition 11. [A restriction of the set of admissible controls] The value function can
be calculated as
v(t, x) = inf J(t, x, u).
u∈Ut
Proof: Clearly, v(t, x) ≤ inf u∈Ut J(t, x, u). Now, let u ∈ U and for simplicity suppose that
` = 0. We know by Dynkin theorem that u can be written as
86
Therefore, for s > t we have
0
u(s) = h (s, W (· ∧ s)) = h s, W (· ∧ t) + W (· ∧ s)
h i
E g(X [u](T )) = C([0,T ]) g(X t,x [u](T, w))dµ(w)
t,x R
J(t, x, u) =
t,x
[u](T, w, w0 ))dµ(w)dµ(w0 )
R R
= C([t,T ]) C([0,t]) g(X
t,x
[u](T, w, w0 ))dµ(w)dµ(w)
R R
= C([0,t]) C([t,T ]) g(X
R
≥ C([0,t]) inf u∈Ut J(t, x, u)dµ(w) = inf u∈Ut J(t, x, u),
which yields the result.
Some properties of the value function: As we will see later v is the unique solution
(in a weak sense to be defined) of a second order HJB equation. However, we can deduce
some interesting properties of v without appealing to this PDE (which in turns a posteriori
prove these properties for the solution of the PDE).
87
1
Theorem 18. [Linear growth for v . Lipchitz property of v w.r.t. x and local 2 Holder
w.r.t. to t] There exists K > 0 such that the value function satisfies
|v(t, x)| ≤ K(1 + |x|), ∀ (t, x) ∈ [0, T ] × Rn ,
1
|v(t, x) − v(t̂, x̂)| ≤ K |x − x̂| + (1 + |x| ∨ |x̂|) |t − t̂| 2 ∀ t, t̂ ∈ [0, T ], x, x̂ ∈ Rn .
Proof: Since our assumptions are uniform on u, the proof is a direct consequence of our estimates for
solutions of SDE (with ` = 1).
88
where Xεt,x[u](s) is the solution of (24) associated with u. The function vε is defined as
We suppose, uniformly in ε, that the data of the family of problems satisfies the same kind
of assumptions than the data for the original problem. For ε = 0, we consider our original
data.
Proposition 12. [A stability result] Suppose that uniformly in (t, u) ∈ [0, T ] × U and
x ∈ K , for some compact K , we have
where φε = bε, σε, `ε, gε. Then vε(t, x) → v(t, x) uniformly over compact sets.
Proof: Fix (t, x) ∈ [0, T ] × Rn and u ∈ Ut . For notational convenience, we do not write the
dependence on time for the data of the problems. Let us set Xε (·) := Xεt,x [u](·), X0 (·) := X t,x [u](·),
δε b(X) := bε (Xε )−b(X), δε σ(X) := σε (Xε )−σ(X). By Itô’s formula, we have for every s ∈ [t, T ],
Z sn h io Z s
2 >
|Xε (s)−X(s)| = 2(Xε − X)δε b(x) + Tr δε σ(X)δε σ(X) dr+2 (Xε −X)δε σ(X)dW (r).
t t
89
From now on we denote by K > 0 a generic constant. Thus,
" #
2
E sup |Xε (r) − X(r)| ≤ K(I1 + I2 ),
t≤r≤s
where,
hR h i i
s Rs >
I1 = E t |Xε (r) − X(r)||δ ε b(X)|dr + t Tr δε σ(X)δε σ(X) dr ,
Rr
I2 = E supt≤r≤s t (Xε − X)δε σ(X)dW (r 0 ) .
Let us first estimate I1 . Since |δε b(X)| ≤ K (|Xε − X| + |bε (X) − b(X)|) with a similar
expression for δε σ , we easily obtain
( "Z # )
s Z s
0 0 2 2 2
I1 ≤ K E sup |Xε (r ) − X(r )| dr + E |bε (X) − b(X)| + |σε (X) − σ(X)| dr .
t t≤r 0 ≤r t
90
By Young inequality, with the obtain that
" # Z s
2 2 2
KI2 ≤ 21 E sup |Xε (r) − X(r)| + 2K E |δε σ(X)| dr .
t≤r≤s t
Therefore we get,
h i n hR i
2 s 0 0 2
E supt≤r≤s |Xε (r) − X(r)| ≤ K E t supt≤r0 ≤r |Xε (r ) − X(r )| dr
hR io
s 2 2
+E t |bε (X) − b(X)| + |σε (X) − σ(X)| dr ,
By the uniform convergence of the parameters we can find η : [0, ∞) × [0, ∞) → [0, ∞), continuous,
non-decreasing, with η(0, R) = 0 for all R ≥ 0, such that
n
|φε (t, x, u) − φ0 (t, x, u)| ≤ η(ε, |x|) ∀ (t, x, u) ∈ [0, T ] × R × U.
91
Thus, we get that
" # Z T h i
2 2
E sup |Xε (s) − X(s)| ≤K E η(ε, |X(s)|) ds
t≤s≤T t
Now, we estimate the r.h.s. of the above equation. By our assumptions, we obtain
92
Therefore,
h
2
i
2 (1 + |x|4 )
sup E η(ε, |X(t)|) ≤ η(ε, R) + K
t∈[0,T ] R2
which gives !
2
1 + |x|
sup E [η(ε, |X(t)|)] ≤ K η(ε, R) + . (25)
t∈[0,T ] R
In particular, letting first ε ↓ 0 and then R ↑ ∞, we get
" #
E sup |Xε (s) − X(s)| → 0.
t≤s≤T
Thus,
|Jε (t, x, u) − J(t, x, u)| ≤ KE sups≤t≤T |Xε (s) − X(s)| + |gε (X(T )) − g(X(T ))|
RT i
+ t |`ε (t, X(T ), u) − `(t, X(T ), u)| dt .
93
Which implies that Jε (t, x, u) → J(t, x, u) uniformly in [0, T ] × Ut and x in a compact set. This fact
implies the result.
and the result follows directly from the expression of |Jε(t, x, u) − J(t, x, u)|.
94
[Semiconcavity of the value function under stronger assumptions] Let us recall that
a function φ is said to be semiconcave if ∀ x, y ∈ Rn, λ ∈ [0, 1]
2
λφ(x) + (1 − λ)φ(y) − φ(λx + (1 − λ)y) ≤ Kλ(1 − λ)|x − y| .
The trivial and useful semiconcave and non-concave example is |x|2 thanks to the identity
2 2 2 2
λ|x| + (1 − λ)|y| = |λx + (1 − λ)y| + λ(1 − λ)|x − y| .
95
Let us assume that g(·) and `(t, ·, u) are semiconcave uniformly on (t, u). Moreover,
let us suppose that b(t, ·, u), σ(t, ·, u) are differentiable and its derivatives are Lipschitz
uniformly in (t, u).
Proposition 14. [Semiconcavity of v ] Under the above assumptions, the function v(t, ·)
is semiconcave, uniformly in t ∈ [0, T ].
Sketch of the proof: Take two points x1 , x2 ∈ Rn and set xλ := λx1 + (1 − λ)x2 . Fix u ∈ Ut
such that
J(t, xλ , u) ≤ v(t, xλ ) + ε.
Given this fact, it is enough to prove the semiconcavity of J(t, ·, u) with an associated constant independent
of u. Let us denote by
Due to the non linearity of b and σ we have in general Xλ 6= λX1 + (1 − λ)X2 and we cannot apply
directly the semiconcavity of `(t, ·, u) an g(·). So we have to work a little... In order to simplify the
notation, suppose that ` ≡ 0
96
Let us define X λ (t) := λX1 + (1 − λ)X2 . In order to use the semiconcavity of g we need to estimate
" #
λ
E sup |X (s) − Xλ (s)|
s∈[t,T ]
As usual these will depend on the difference of the coefficients. Let us write the coefficient of the dt part
This is easily checked using that b(s, ·) has a Lipschitz derivative. Analogously, we obtain
λ 2
λσ(s, X1 (s)) + (1 − λ)σ(s, X2 (s)) − σ(s, X (s)) ≤ Kλ(1 − λ)|X1 (s) − X2 (s)| .
Then use the standard procedure: write |X λ (s) − Xλ (s)|2 , apply the BDG inequality and then use the
Gronwall inequality to get the result (Exercise!).
97
The Hamilton-Jacobi-Bellman equation
Now, we do the link with the PDEs. In fact, for the moment, we prove that if
v ∈ C 1,2([0, T ] × Rd) then it satisfies a second order Hamilon-Jacobi-Bellman (HJB)
equation (compare with proposition 10). Let us define the function
Ĥ : [0, T ] × Rn × U × Rn × Rn×n → R
H : [0, T ] × Rn × Rn × Rn×n → R
as
98
where
h i
> 1 > 2
A[u]f (t, x) = b(t, x, u) Df (t, x) + 2 Tr σσ (t, x, u)D f (x, t) .
“An optimal policy has the property that whatever the initial
state and initial decision are, the remaining decisions must
constitute an optimal policy with regard to the state
resulting from the first decision.”
R. Bellman [4]
The following property will be crucial in order to characterize the value function as the
solution of a PDE.
99
Theorem 19. [Dynamic programming principle] For any (t, x) ∈ [0, T ] × Rn we have
"Z #
t̄
t,x t,x
v(t, x) = inf E `(t, X [u](s), u(s))ds + v(t̄, X [u](t̄)) .
u∈Ut t
Dynkin theorem 3 implies that each time that we freeze the W (·) up to time t̄ (i.e. we condition w.r.t.
Ft̄ and we evaluate at ω ), there exists an admissible control ûω ∈ Ut̄ such that ûω (s) = u(s, ω) for all
s ∈ [0, t̄]. Conditioning w.r.t. Ft̄ we get
hR i
t̄ t,x t,x
J(t, x, u) =E `(t, X [u](s), u(s))ds + J(t̄, X
[u](t̄), û)
hRt i
t̄ t,x t,x
≥E t `(t, X [u](s), u(s))ds + v(t̄, X [u](t̄)) ,
100
which implies that
"Z #
t̄
t,x t,x
v(t, x) ≥ inf E `(t, X [u](s), u(s))ds + v(t̄, X [u](t̄)) .
u∈Ut t
The other inequality is rather difficult to prove, specially for more general problems when the value
function v is not a priori continuous. There are different types if proofs: El Karoui [9] used delicate
measurability selections theorems. Yong and Zhou [24] consider a weak form of the formulation of stochastic
optimal control problems (i.e. the probability space is a part of the control). B. Bouchard and N. Touzi [6]
established a general “weak form” of the dynamic programming principle, that allows to establish the HJB
equation (26) in a very general framework.
101
Proof: For simplicity we will suppose that the derivatives ∂t v , Dv and D 2 v are bounded. Otherwise,
we have to use a localization argument with the stopping times (as in the proof of proposition 10). Given
u ∈ U define a control u ∈ Ut as u(t, ω) ≡ u. By the dynamic programming principle we have
"Z #
t̄
x,t x,t
v(t, x) ≤ E `(s, X [u](s), u)ds + v(t̄, X [u](t̄)) . (27)
t
2
∂t v(t, x) + Ĥ(t, x, u, Dv(t, x), D v(t, x)) ≥ 0.
102
By taking the infimum over u ∈ U we get
2
∂t v(t, x) + H(t, x, Dv(t, x), D v(t, x)) ≥ 0.
Again, using Itô’s formula, we get (we simplify the notation since the context is clear)
hR n oi
t̄ 2
ε(t − t̄) ≥ E t ∂t v(s, X[uε,t̄ ](s)) + Ĥ s, X[uε,t̄ ], uε,t̄ , Dv s, X[uε,t̄ ] , D v s, X[uε,t̄ ]
hR i
t̄ 2
≥ E t ∂t v(s, X[uε,t̄ ](s)) + H s, X(s), Dv s, X[uε,t̄ ], D v(s, X[uε,t̄ ] .
Using the uniform continuity of the functions, we can divide by t̄ − t and to pass to the limit to get the
result.
103
(ii) We have the following verification argument for open-loops controls. An admissible
pair (x̄(·), ū(·)) is optimal for v(x, t) iff
2
∂tv(x, t) + Ĥ(t, x̄(t), ū(t), Dv(t, x̄(t)), D v(t, x̄(t)) = 0.
(iii) We have the following verification argument for Markov or feed-back controls.
Suppose that for each (t, x) there exists a ū(t, x) ∈ U such that
2 2
Ĥ t, x, ū(t, x), Dv(t, x), D v(t, x) = H t, x, Dv(t, x), D v(t, x) .
dX(s) = b (s, X(s), ū(s, X(s))) ds + σ(s, X(s), ū(s, X(s)))dW (s)
X(t) = x
(28)
admits a unique solution for all (x, t). If s ∈ [t, T ] → ū(s, X(s)) ∈ U is admissible,
then it is optimal.
104
Proof: Let u ∈ Ut with associate trajectory x(·) with x(t) = x, then we have
hR i
T
E(g(x(T )) = v(x, t) + E t {∂t v(s, x(s)) + Av(s, x(s))}
hR n
T
≥ v(x, t) + E t ∂t v(s, x(s)) + H(s, x(s), Dv(s, x(s)), D 2 v(s, x(s)))
−`(s, x(s), u(s))}] ,
which, using the equation (26), proves (i). For (ii), the inequality becomes equality, from which follows
easily the result. Finally, (iii) is a direct consequence of (ii).
105
Viscosity solutions
As we have seen, if the value function is regular enough, then it solves equation (26).
This equation can be written in the abstract form
2
F (t, x, v(t, x), ∂tv(t, x), Dv(t, x), D v(t, x)) = 0 plus limit conditions.
(29)
In general this equation does not admit a classical solution. However we can expect to
define a weak type of solutions such that (29) is well posed, i.e. it has a unique solution.
The correct manner is the one of f viscosity solutions (introduced in this context by Lions
in [15, 16])
However, we need to pay attention to the following fact: One of the crucial assumptions
in the theory of viscosity solutions is that F (t, x, v, q, p, P ) is monotone w.r.t. P . The
usual convention for the type of monotonicity is that F is not increasing w.r.t. P . This is
106
also called ellipticity condition. In order to follow this convention, we have to write (26) as
= supu∈U −Ĥ(t, x, u, p, P ).
107
[Definition of viscosity solutions] (i) We say that v ∈ C([0, T ] × Rn) is a viscosity
subsolution of (30) if
v(T, x) ≤ g(x),
and for any φ ∈ C 1,2([0, T ] × Rn), whenever v − φ attains a local maximum at (t, x)
we have
2
−∂tφ(t, x) + G(t, x, Dφ(t, x), D φ(t, x)) ≤ 0.
v(T, x) ≥ g(x),
and for any φ ∈ C 1,2([0, T ] × Rn), whenever v − φ attains a local minimum at (t, x)
we have
2
−∂tφ(t, x) + G(t, x, Dφ(t, x), D φ(t, x)) ≥ 0.
108
[Some important remarks] The following properties are proposed as an exercise! (in
increasing order of difficulty)
(i) We can suppose that at the test point (x, t) we have v(x, t) = φ(x, t).
(ii) We can change “local maximum”, “local minimum” by “local maximum strict” and
“local minimum strict”.
(iii) We can change “local strict maximum”, “local strict minimum” by “global strict
maximum” and “global strict minimum” .
In our stochastic framework, it will be convenient to work with the last notion in order
to avoid localization arguments with stopping times (see the proof of theorem 22).
Proof: We essentially repeat the proof for the regular case, but taking into the account the test
functions in order to be able to differentiate. Let φ ∈ C 1,2 ([0, T ] × Rn ) and (t, x) such that v − φ has
a global maximum at (t, x). Let u ∈ U and define u(·) ∈ Ut defined as u(t, ω) ≡ u. Let us denote by
109
x(·) the associated state, with x(t) = x. We have, for any t̄ > t,
"Z #
t̄
E {v(s, x(s)) − φ(s, x(s)) − [v(t, x) − φ(t, x)]} ds ≤ 0.
t
Therefore, "Z #
t̄
E {−∂t φ(s, x(s)) − `(s, x(s), u) − A[u]φ(s, x(s))} ds ≤0
t
By taking the supremum w.r.t. u ∈ U we obtain
"Z #
t̄ n o
2
E −∂t φ(s, x(s)) + G(s, x(s), Dφ(s, x(s)), D φ(s, x(s)) ds ≤ 0.
t
110
Dividing by t − t̄ and taking the limit t̄ → t, yields the subsolution property.
For the supersolution property, let φ ∈ C 1,2 ([0, T ] × Rn ) and (t, x) such that v − φ has a global
minimum at (t, x). Thus, for t̄ > t, we have for every adapted x(·) starting at x,
"Z #
t̄
E {v(s, x(s)) − φ(s, x(s)) − [v(t, x) − φ(t, x)]} ds ≥ 0. (31)
t
Using the dynamic programming principle, chose uε,s̄ ∈ Ut , with associated state xε,s̄ , such that
"Z # "Z #
t̄ n o t̄
E v(t, x) − v(s, xε,t̄ (s)) ds ≥ E ` s, xε,t̄ (s), uε,t̄ (s) ds − ε(t̄ − t).
t t
111
which implies that
"Z #
t̄ n o
2
E −∂t φ(s, xε,t̄ (s)) + G(s, xε,t̄ (s), Dφ(s, xε,t̄ (s)), D φ(s, xε,t̄ (s)) ds ≥ −ε(t̄ − t).
t
[Basic results] We first provide a proposition that implies that viscosity solutions
qualify as generalized solutions:
Proposition 15. [Equivalence of notions under regularity] Let v ∈ C 1,2([0, T ] × Rn.
Then v is a viscosity solution of (30) iff it is a classical solution.
Proof: The proof for the subsolution (supersolution) property follows very easily from the first and
second order optimality conditions for a maximum (minimum) of v − φ at some (x, t), together will the
ellipticity of G.
In fact, in order to satisfies the equation point-wisely, we can ask, instead of v ∈ C 1,2,
that v admits a “first order expansion in t” and a second order expansion at x.
112
Lemma 8. [Relation with “second order expansions in x”] Suppose that for all (t, x)
there exists X (for notational convenience we denote D 2v(t, x) = X ) such that
2
v(s, y) = v(t, x) + ∂t v(t, x)(s − t) + Dv(x, t)(y − x) + 21 hX(y − x), y − xi + o(|s − t| + |x − y| ).
2
−∂tv(t, x) + G(t, x, Dv, D v(t, x)) ≤ (≥)0.
Proof: For the subsolution case, it is enough to test using the quadratic function
2
φε,δ (s, y) := v(t, x) + (∂t v(t, x) + ε)(s − t) + Dv(x, t)(y − x) + 12 h[D v(x, t) + δI](y − x), y − xi
In fact, we have that v(t, x) − φε,δ (s, y) has a local maximum at (t, x). We thus find that
2
−(∂t v(t, x) + ε) + G(t, x, Dv, D v(t, x) + δI) ≤ 0.
and the result follows by letting ε, δ → 0. An analogous argument applies for the supersolution property.
113
Now we turn our attention to an important and rather surprising stability result.
Proposition 16. [Stability result] Let vε be a solution of
where, for every ε > 0, the functions Gε and gε are continuous. Suppose that as
ε ↓ 0, we have Gε → G, hε → h and vε → v uniformly over any compact set. Then
v is a viscosity solution of
Proof: The proof, as well as a several of arguments in this theory, is based on the the basic lemma 9
below. Let us prove that v is a subsolution of (33). Let φ ∈ C 1,2 and (t, x) such that v − φ has a strict
maximum at (t, x). Using the lemma below, we have that the existence of (tn , xn ) → (t, x) such that
vn − φ has a local maximum at (tn , xn ). Using that vn is a subsolution of (32), we have
2
−∂t φ(tn , xn ) + Gε (tn , xn , Dφ(tn , xn ), D φ(tn , xn )) ≤ 0.
114
By passing to the limit, we obtain the result. The supersolution property follows with the same procedure.
Proof of the lemma: There exists a δ > 0 such that v(x0 ) > v(x) for x ∈ Bδ (x0 ). Therefore,
since vn → v uniformly on B δ (x0 ), after some n0 , we will have that any maximum xn of vn in B δ (x0 )
belongs to Bδ (x0 ). In order to prove that xn → x0 , let x̄ be a limit point of xn . Then, because
of the uniform convergence, vn (xn ) → v(x̄) = v(x0 ) (uniform convergence implies convergence of the
maximum of the functions, to see this note that vn (xn ) ≥ vn (z) for all z ∈ Bδ (x0 ) and then pass to the
limit). Therefore, because x0 is the unique maximum of v in Bδ (x0 ) we get that x̄ = x0 , and the proof is
completed.
Now we state the important issue of uniqueness of a viscosity solution for (30).
Theorem 23. [Uniqueness theorem] The value function is the unique solution of (30).
115
As a important corollary we get the an error estimate for the vanishing viscosity
approximation of (30). In fact, equation (30) can be degenerate and in general does not
have regular solutions. A natural way of regularizing the equation is to consider
We will not prove exactly the uniqueness theorem. However, we will prove a comparison
principle for a model second order HJB equation. Up to important technical matters (see
[24]), the same procedure allows to prove the uniqueness result of a viscosity solution of
(30).
116
Let O ⊆ Rn be an open and bounded set. Let us consider the equation
2
H v(x), Dv(x), D v(x) = 0, x ∈ O. (35)
where H : Rn × R × Rn × S n → R. We assume:
(i) H is continuous.
(ii) [Ellipticity condition] We suppose that H is non-increasing w.r.t. the last variable, i.e.
for every A, B ∈ S n, with A ≤ B we have
n
H(v, p, A)≥H(v, p, B) for all (v, p) ∈ R × R .
The definition of subsolutions and supersolutions is of course analogous to the one given
for the parabolic case. The result that we want to prove is the following:
117
Theorem 24. Let v and v̂ be a subsolution and a supersolution, respectively, of (35).
If v ≤ v̂ in ∂O , then v ≤ v̂ in O .
Proof for the regular case: Let us assume that v, v̂ ∈ C 2 (Ω) (or more generally that the admit a
second order expansion). We argue by contradiction. Suppose that M := supx∈O [v(x) − v̂(x)] =
v(x0 ) − v̂(x0 )> 0, for some x0 ∈ O . Then,
2 2
Dv(x0 ) = Dv̂(x0 ) and D v(x0 ) ≤ D v̂(x0 ). (36)
By combining (36) and (36) and using the ellipticity property, we obtain
2 2
H v(x0 ), Dv̂(x0 ), D v̂(x0 ) − H v̂(x0 ), Dv̂(x0 ), D v̂(x0 ) ≤ 0,
For the proof of the general case, we will try to “mimic” the above proof. We will
need need to approximate v and v̂ in such a way that they become almost two times
118
differentiable. The good approximation is by semiconvex functions and one way to do it is
to use the so-called inf-convolutions and sup-convolutions.
The following arguments and the proof of the uniqueness result are based on the notes
[8].
Why semi-convex functions are the good regularization? We provide below two
theorems that answer this question:
Theorem 25. [Alexandrov theorem] Let w be a semiconvex function with constant M
over the open set O . Then, for a.a. x ∈ O , the function w admits a second order
expansion, i.e. there exist X ∈ Sn (which depends on x) such that
2
w(y) = w(x) + Dw(x)(y − x) + 12 hX(y − x), y − xi + o |y − x| .
119
Theorem 26. [Jensen maximum principle] Let w : O → R be a semi-convex function
with an strict local maximum at some x0 ∈ O . More precisely, set α := w(x0) −
max∂Br (x0) w, where r > 0. For δ > 0, define
n
Eδ := x ∈ Br (x0 ) ; ∃ p ∈ R , |p| ≤ δ, w(y) − hp, y − xi ≤ w(x) ∀ y ∈ Br (x0 ) .
We will use both theorems in the following way: If a point x ∈ Eδ and it admits a
second order expansion then
2
|Dw(x)| ≤ δ, D w(x) ≤ 0, and w(y) ≤ w(x) + hDw(x), y − xi ∀ y ∈ Br (x0).
120
Proposition 18. Consider a semiconvex function w with a local strict maximum at x0.
Then, there exists a sequence xn → x0 such that w has a second order expansion at
x0 and
2
Dw(xn) → 0 and D w(xn) ≤ 0.
Proof: It suffices to consider points xn in E 1 such that w has a second order expansion at xn . Thus
n
the properties announced follow for the above remark, except for the convergence of xn to x. But
which, by passing to the limit and using that x0 is a strict local minimum, imply that every limit point of
xn is equal to x0 . The result follows.
Now we state a version of the above propositions that do not ask for a strict local
maximum.
Proposition 19. Consider a semiconvex function w with a local maximum at x0. Then
there exists a matrix X ∈ Sn and a sequence xn → x0 such that w has a second order
121
expansion at x0 and
2
Dw(xn) → 0 and D w(xn) → X ≤ 0.
Proof: Apply proposition 18 to the semiconvex function wk (x) := w(x) − k1 |x − x0 |2 and use a
diagonal procedure (exercise!).
122
Note that vα = −(−v)α, which allow us to extend properties of the sup-convolution
to the inf-convolution.
Lemma 11. [A convergence result for v α] For all α > 0, we have that v α ≥ v . Also,
we have that v α(x) → v(x) for all x ∈ K . Moreover,
α
lim v (xα) = v(x). (38)
α↓0, xα →x
Proof: Taking y = x in the definition of v α we get that v α (x) ≥ v(x). Now, we prove
123
lim supxα →x v α (xα ) ≤ v(x). In fact, by compactness we have the existence of yα such that
α 1 1
v (xα ) = v(yα ) − |xα − yα | ≤ M − |xα − yα |,
α α
1 α
|xα − yα | ≤ M − v (xα ) ≤ M − v(xα ) ≤ 2M.
α
Consider a subsequence such that lim v αn (xαn ) = lim supxα →x v α (xα ). We have
α
lim v n (xαn ) ≤ lim v(yαn ) = v(x).
α α
v(x) = lim v(xα ) ≤ lim inf v (xα ) ≤ lim sup v (xα ) ≤ v(x).
.
124
then v α is a subsolution of the same equation but on the open set
" #1
2
α,k α
O := x ∈ O ; v > −k, d(x, ∂O) > α(sup v + k)
O
Finally, we will need another simple lemma whose easy proof is left to the reader.
α
lim max w = max w.
α↓0 K
125
Proof of the comparison principle: We argue by contradiction. If the result is not true, since we
have v ≤ v̂ in ∂O , we have the existence of x̂ ∈ O such that v(x̂) > v̂(x̂). Let us denote by
M = maxx∈O {v(x) − v̂(x)} > 0. As for the proof of the comparison principle for first order equations
(see e.g. [3]) we double variables, by introducing, for ε > 0,
1 2
wε (x, y) := v(x) − v̂(y) − 2 |x − y| .
ε
which we know that “approximates well” the function v − v̂ (see e.g. [3]). By continuity and convergence
we have maxÔ wε > 0 for ε small enough. Also, there exists θ > 0 (independent of ε) such that for every
maximum point (xε , yε ) of wε we have
d(xε , ∂O) > θ ; d(yε , ∂O) > θ. for all ε small enough.
Now fix such an ε. To complete the proof we follow the following steps:
(i) We modify the regularization wε to
α 1 2
wε,α (x, y) = v (x) − v̂α (y) − 2 |x − y| for all α > 0. (39)
ε
This allows to obtain functions that have a.e. a second order expansion and to apply optimality conditions
at an optimal (xε , yε ). In this way we try to proceed as if originally the functions were differentiable. To
put v α it is natural because we now that it is a subsolution over a smaller domain O α . Analogously, to put
v̂α is natural since it is a super-solution over a smaller domain Oα .
126
(ii) We construct O α and Oα in such a way that any maximum (xα , yα ) of wε,α satisfies that xα ∈ O α
and yα ∈ Oα . In fact, to do this we need the following lemma (whose easy proof is left to the reader):
Lemma 14. Let f α and f be u.s.c. over a compact set K . Suppose that
α α
f ≥f and that lim sup f (xα ) = f (x).
α↓0, xα →x
Then, for every ε > 0, there exists α0 > 0 such that: For all α ∈ (0, α0 ) and xα ∈ argmax f α
there exists x ∈ argmax f such that |x − xα | ≤ ε. Moreover
α
lim max f = max f.
α↓0 K K
Using the above lemma we see that for α small enough, we have that (xα , yα ) is uniformly close to
(xε , yε ) ∈ argmax wε . In particular, d(xα , ∂O) > θ/2 and d(yα , ∂O) > θ/2. The same lemma
implies that
lim max wα,ε = wε .
α↓0 O×O
From the definition in terms of v and v̂ and the fact that
α 1 2
v (xα ) − v̂α (yα ) − 2 |xα − yα | = max wε,α > 0.
ε
127
we readily have that v α (xα ) and v̂α (yα ) are bounded by a constant independent of (ε, α). Therefore
there exists a constant k such that
i1
( )
h
xα ∈ O := x ∈ O ; v (x) > −k ; d(x, ∂O) > α(supO v + k) 2
α α
,
i1
( )
h
yα ∈ Oα := x ∈ O ; v̂α (x) < k ; d(x, ∂O) > α(inf O v̂ + k) 2
(iii) We fix α and a maximum (x̄, ȳ) ∈ O α × Oα . We obtain information from this optimality thanks to
the semiconvexity and proposition 19. In fact, there exists a sequence (xn , yn ) → (x̄, ȳ) and a matrix
A ∈ S 2n , with A ≤ 0, such that
2
Dwε,α (xn , yn ) → 0 ; D wε,α (xn , yn ) → A. (40)
(iv) In view of the particular structure of wε,α (a decoupled function plus a C ∞ function), we have an
important information: the fact that wε,α admits a second order expansion at (xn , yn ) implies that v α
admits a second order expansion at xn and v̂α admits a second order expansion at yn .
128
Moreover, we have
α 2 2
Dwε,α (xn , yn ) = Dv (xn ) − 2 (xn − yn ), −Dv̂α (yn ) + 2 (xn − yn )
ε ε
! (41)
D 2 v α (xn )
2 0 I n −I n
D wε,α (xn , yn ) = 2 − 22
0 −D v̂α (yn ) ε −I n I n
α 2 2
Dv (xn ) → 2 (x̄ − ȳ), Dv̂α (yn ) → 2 (x̄ − ȳ) (42)
ε ε
and
X 0 2 In −In
A= − 2 (43)
0 −Y ε −In In
n n
Testing with (z, z) ∈ R × R we get that X ≤ Y .
(iv) Intuitively, we are almost done, because if we do the analogy with the regular case (recall that x̄ should
be like ȳ ), we have that (42) is like Dv(x̄) = Dv̂(x̄) and (43) is like D 2 v(x̄) ≤ D 2 v̂(x̄). Let us finish
the proof in the correct manner. Since v α is a subsolution of (35) in O α and it has a second order expansion
at every xn , we get
α α 2 α
H(v (xn ), Dv (xn ), D v (xn )) ≤ 0.
The continuity of H implies
α 2
H v (x̄), 2 (x̄ − ȳ), X ≤ 0.
ε
129
Analogously, since v̂α is a subsolution of (35) in Oα and it has a second order expansion at every yn , we
get
2
H v̂α (ȳ), 2 (x̄ − ȳ), Y ≥ 0.
ε
Subtracting the inequalities we get
α 2 2
H v (x̄), 2 (x̄ − ȳ), X − H v̂α (ȳ), 2 (x̄ − ȳ), Y ≤ 0.
ε ε
130
References
[1] R.B. Ash. Basic probability theory. Wiley, NY, 1970.
[2] R.B. Ash. Real Analysis and Probability. Academic Press, NY, 1972.
[3] M. Bardi and I. Capuzzo Dolcetta. Optimal control and viscosity solutions of Hamilton-Jacobi-Bellman
equations. Birkauser, 1996.
[4] R. Bellman. Dynamic Programming. Princeton Univ. Press, Princeton, New Yersey, 1957.
[5] P. Billingsley. Convergence of Probability Measures. Wiley, NY, 1968.
[6] B. Bouchard and N. Touzi. Weak dynamic programming principle for viscosity solutions. SIAM J.
Control Optim., 49-3:948–962, 2011.
[7] L. Breiman. Probability. Addison-Wesley Publishing Company, Reading, MA, 1968.
[8] P. Cardaliaguet. Solutions de viscosité d‘équations elliptiques et paraboliques non linéaires. Lecture
Notes for the DEA program at Rennes, 2004.
[9] N. ElKaroui. Les aspects probabilistes du contrôle stochastique. Lecture Notes in Math. 876, 1981.
[10] L.C. Evans and R.F. Gariepy. Measure theory and fine properties of functions. CRC Press, Boca Ratón,
FL, 1992. Studies in Advanced Mathematics.
[11] W.H. Fleming and H.M. Soner. Controlled Markov processes and viscosity solutions. Springer, New
York, 1993.
[12] N. Ikeda and S. Watanabe. Stochastic differential equations and diffusion processes. Second Edition,
North-Holland Publishing Co., Amsterdam, 1989.
131
[13] R. Jensen. The maximum principle for viscosity solutions of fully nonlinear second order partial
differential equations. Arch. Ration. Mech. Anal., 101-1:1–27, 1988.
[14] I. Karatzas and S.E. Shreve. Brownian Motion and Stochastic Calculus. Second Edition Springer-Verlag,
New York, 1991.
[15] P.-L. Lions. Optimal control of diffusion processes and Hamilton-Jacobi-Bellman equations. I. The
dynamic programming principle and applications. Comm. Partial Differential Equations, 8(10):1101–
1174, 1983.
[16] P.-L. Lions. Optimal control of diffusion processes and Hamilton-Jacobi-Bellman equations. Part 2:
viscosity solutions and uniqueness. Comm. Partial Differential Equations, 8:1229–1276, 1983.
[17] P.A. Meyer. Probability and Potentials. Blaisdell Publishing Company, Waltham, Mass., 1966.
[18] L. Mou and J. Yong. A variational formula for stochastic controls and some applications. Pure and
Applied Mathematics Quarterly, 3:539–567, 2007.
[19] K.R. Parthasarathy. Probability Measures on Metric Spaces. Academic Press, New York, 1967.
[20] H. Pham. Optimisation et contrôle stochastique appliqués à la finance, volume 61 of Mathématiques
& Applications. Springer, Berlin, 2007.
[21] P. Protter. Stochastic integration and differential equations. Springer-Verlag, Berlin, 2nd edition,,
2004.
[22] J. Steele. Stochastic calculus and financial applications. Springer-Verlag, New York, 2001.
[23] N. Touzi. Optimal Stochastic Control, Stochastic Target Problems, and Backward SDEs. Lecture
Notes at the Fields Institute, 2010.
132
[24] J. Yong and X.Y. Zhou. Stochastic controls, Hamiltonian systems and HJB equations. Springer-Verlag,
New York, Berlin, 2000.
133