Brownian Motion 22/23 Warwick
Brownian Motion 22/23 Warwick
2
CONTENTS 3
X: Ω → S .
Definition 1.1.1. Let I ⊆ [0, ∞) be an index set. Then a stochastic process over I with
values in S is a collection {Xt }t∈I of random variables. Namely, for every t ∈ I, Xt is a
measurable map
Xt : Ω → S .
If I = N we refer to {Xn }n∈N as a discrete time stochastic process. If I = [0, ∞) we refer
to {Xt }t∈[0,∞) as a continuous time stochastic processes.
For example, let {ηi }i∈N be a collection of i.i.d. real valued random variables. Then
the random walk {Sn }n∈N
Xn
Sn = ηi
i=0
is a discrete time stochastic process. Of course, it can also be seen as a continuous time
stochastic processes, by defining St = S⌊t⌋ for any t > 0.
πt ((sr )r∈I ) = st .
Then the space S I is a measurable space, when enriched with the product σ-field S I . This
is the smallest sigma-field that makes finitely many coordinate maps measurable:
S I = σ(πt : t ∈ I) .
4
1.1. STOCHASTIC PROCESSES 5
Let us denote with T the set of ordered times t = (t1 , · · · , tn ) with n arbitrary and
t1 < · · · < tn . Then we can equivalently characerise S I as the smallest σ-field that
contains all cylinder sets of the form
Ct,A = {(st )t∈I ∈ S I : (st1 , · · · , stn ) ∈ A} , (1.1)
for any choice of A ∈ S t and any t ∈ T .
Theorem 1.1.2 (Dynkin Systems). Let D be a collection of subsets of a nonempty set Ω
such that
1. Ω ∈ D.
2. A, B ∈ D and B ⊆ A imply A \ B ∈ D.
S∞
3. {An }∞
n=1 ⊆ D and A1 ⊆ A2 ⊆ · · · An ⊆ · · · imply n=1 Ai ∈ D.
If C is any other collection of subsets and C ⊆ D, then also σ(C) ⊆ D.
Exercise 1.1.3. Any stochastic process can be viewed as a measurable map X : Ω → S I
(use cylinder sets and the Dynkin System theorem).
Proof. Let D be the collection of subsets A ⊆ S I such that X −1 (A) is measurable. Then
D is a Dynkin system containing cylinder sets.
Now consider a stochastic process X and a set of times t = (t1 , · · · , tn ) such that
t1 < · · · < tn and tk ∈ I for all k ∈ {1, · · · , n}. To any such choice of t we can associate the
finite dimensional distributions (FDDs) of X a times t. Namely, we obtain the probability
measure µt on S n characterised by
µt (A1 × · · · × An ) = P(Xt1 ∈ A1 , · · · , Xtn ∈ An ) , ∀Ai ⊆ S .
The family {µt }t∈T of FDDs of a stochastic process is consistent, in the following sense.
Definition 1.1.4. A collection {µt }t∈T of probability measures such that µt is a measure
on S n (where t = (t1 , · · · , tn )) is said to be consistent if for any t = (t1 , · · · , tn ) ∈ T and
any k ∈ {1, · · · , n} it holds that
µt̂k (A1 × · · · × Ak−1 × Ak+1 × · · · × An ) = µt (A1 × · · · × Ak−1 × S × Ak+1 × · · · × An ) ,
Then there exists a unique measure µ : S → [0, 1] such that µ = µ when restricted to C.
6 CHAPTER 1. CONSTRUCTION OF BROWNIAN MOTION
Proof of Theorem 1.1.5. For any t ∈ T and A ∈ S I , let Ct,A be the Cylinder set as in
(1.1). Then define
µ(Ct,A ) = µt (A) .
Since there can exists (t, A), (t′ , A′ ) such that Ct,A = Ct′ ,A′ , the definition of µ makes
sense only if for any such couple we have that µt (A) = µt (A′ ). It suffices to prove the
claim for product sets, namely assuming that A = A1 × · · · × An and A′ = A′1 × · · · × A′n ,
we must show that
µt (A1 × · · · × An ) = µt′ (A′1 × · · · A′n′ ) .
This follows from the consistency of the measures and the observation that if t ∈ t \ t′
(when viewed as sets of times), then necessarily At = Rd , since Ct,A = Ct′ ,A′ .
Exercise 1.1.7. Prove that the collection of cylinder sets {Ct,A }t,A is an algebra, namely
that it is closed under taking complements, finite unions and finite intersections.
In view of this exercise, the existence of the measure µ (that is the extension of its
definition to the entirety of S I ) follows from Theorem 1.1.6, if we show that the measure
is countably
S additive. Namely, let Cn , n ∈ N be a disjoint collection of sets in C such that
C = n∈N Cn ∈ C. We would like to prove that
X
µ(C) = µ(Cn ) .
n∈N
and let us find a contradiction. Let us assume in addition that for each n ∈ N there exist
times (t1 , · · · , tn ) and a set An ∈ B(Rd×n ) such that
Qn = {(st1 , · · · , stn ) ∈ An } .
Exercise 1.1.8. How can we always reduce ourselves toTthis setting,T up to choosing a
slightly different sequence of decreasing sets Q′n satisfying n∈N Q′n = n∈N Qn ?
If the sets An were compact we could now complete the proof. Indeed, since µ(Qn ) > ε
for every n ∈ N the sets Qn cannot be empty and we would find a point (snt1 , · · · , sntn ) ∈ Qn .
Since the sets are also decreasing, the sequence (snt1 )n∈N ∈ A1 , which is compact, so there
exists a limit point rt1 ∈ A1 . Similarly for (snt1 , snt2 )n∈N , which must admit a limit point
T of
the form (rt1 , rt2 ) ∈ A2 . Iterating this procedure we can construct a point in p ∈ n Qn
satisfying
πtk (p) = rk ,
T
which contradicts the fact that n Qn = ∅.
In general the sets An are not compact, but this problem can be overcome as follows.
For each n ∈ N one can find a compact subset Kn ⊆ An such that µtn (An \Kn ) ⩽ 2εn (Borel
1.2. GAUSSIAN PROCESSES 7
• Definition 1.2.1 is stronger than the statement that each Xi is Gaussian. For exam-
ple, let X ∼ N (0, 1) and
+X , if |X| ≤ 1
Y = .
−X , if |X| > 1
• Suppose X = (X1 , . . . , Xn ) is Gaussian. Then for any matrix A ∈ Rm×n the random
vector Y = AX ∈ Rm is Gaussian.
The previous proposition allows us to define Gaussian stochastic processes based solely
on their mean and covariance structure.
Definition 1.2.3. A continuous-time stochastic process (Xt )t⩾0 with state space S = R
is a Gaussian process if all FDDs are Gaussian, i.e. if for all n ∈ N and all t1 , . . . , tn ≥ 0
the random vector X(t1 ), . . . , X(tn ) has a multivariate Gaussian distribution.
The definition extends analogously to discrete time stochastic processes. From Propo-
sition 1.2.2 we deduce that the following holds.
Corollary 1.2.4. Let Xt be a Gaussian process. Then its law on (Rd )[0,∞) is uniquely
determined by its mean and covariance functions:
µ(t) := E X(t) and σ(s, t) := Cov X(s), X(t) , s, t ≥ 0 . (1.3)
Moreover, for any two functions t 7→ µt and (s, t) 7→ σs,t , such that for any t = (t1 , . . . , tn ) ∈
T the matrix (σti ,tj )i,j is positive semidefinite there exists a unique Gaussian process with
mean function µ and covariance function σ.
3. (Variance and mean) For any t ⩾ 0 we have Var(Bt ) = t, and E[Bt ] = 0. Or better
Bt ∼ N (0, t).
Observe that it is yet unclear that Brownian motion exists, as we have no tools to
prove that the fourth property holds. But the tools we have introduced so far allow us to
construct a unique (in law, on R[0,∞) ) stochastic process satisfying properties (1) − (3).
For the previous definition to make sense (in view of Proposition 1.2.2 and Kol-
mogorov’s consistency criterion), we observe that for ordered times 0 ⩽ t1 < . . . < tn the
FDDs of a stochastic process (Xt )t⩾0 at times t = (t1 , . . . , tn ) are fully characterized by the
initial distribution of X0 and distribution of the increments ∆(ti−1 , ti ) := X(ti ) − X(ti−1 )
(with the convention that t0 = 0), since we can write:
n
X
X(t1 ), . . . , X(tn ) = X(0) + ∆(t0 , t1 ), ∆(t0 , t1 ) + ∆(t1 , t2 ), . . . , ∆(ti−1 , ti ) . (1.4)
i=1
Proposition 1.2.7. The following are equivalent for a stochastic process X(t) : t ≥ 0
on R with a fixed initial condition X(0) = x.
(a) X(t) : t ≥ 0 has stationary independent increments with
X(t) − X(s) ∼ N (0, t − s) for all t > s ≥ 0 .
(b) X(t) : t ≥ 0 is a Gaussian process with constant mean E[X(t)] = x and covariance
Cov[Xs , Xt ] = s ∧ t (:= min{s, t}) . (1.5)
While the FDDs fix the distributional properties of a process via its law on the product
space (S I , S I ), we might be interested in path properties of Brownian motion. For example
we might wonder whether the typical path t 7→ Bt (ω) is continuous (or Hölder continu-
ous, or even differentiable). For all these purposes, the infinite produce space R[0,∞) is
insufficient and we have to consider the law of B on the space of continuous functions.
P Bk (t) = B(t) = P[τk ̸= t] = 1 for all t ≥ 0 , (1.7)
due to the continuous distribution of the τk . Since via Theorem 1.1.5 the FDDs uniquely
determine the law of a process on (R[0,∞) , B(R[0,∞) )), we must have P[Ck ] = P[C] for all
k ≥ 1 where Ck = {Xk is continuous}. Since the Ck ⊆ Ω0 are mutually disjoint this
implies P[C] = 0.
Instead, we shall prove that it is always possible to modify Brownian motion in such
a way that the FDDs are not affected, but also so that its sample paths are continuous in
time.
Definition 1.3.2. Y is a modification of X, if P X(t) = Y (t) = 1 for all t ≥ 0.
Proof. Since a finite union of null-sets is still a null-set, we have that for any t1 < t2 <
· · · < tn
P(Yt1 = Xt1 , Yt2 = Xt2 , . . . , Ytn = Xtn ) = 1 .
The result follows.
Before we prove this result, let us consider a few simple properties of Brownian motion
(assuming that the process exists).
1.3. CONSTRUCTION OF BROWNIAN MOTION 11
1
(c) X(t) := √ B(ct) for any fixed c > 0 , (scaling)
c
0 , for t = 0
(d) X(t) := , (time inversion)
tB(1/t) , for t > 0
Note that this implies the following asymptotic behaviour: For a ∈ R we have
B(t)
+∞ , a > 0
B(t) + at = t +a → almost surely as t → ∞ .
t −∞ , a < 0
Proof of Theorem 1.3.7. To prove this result, let us fix T = 1 without loss of generality,
and introduce dyadic points on [0, 1]:
∞
k [
Dn = : k = 0, . . . , 2n , D= Dn .
2n
n=1
We write tnk ∈ Dn for the point tnk = 2kn . We will then study the cointinuity properties
of X on the set of points D and construct a modification X e of X that coincides with the
latter on D.
Step 1: Continuity on D. For two successive points tnk , tnk+1 ∈ Dn we can use Markov’s
inequality to bound
and similarly
−γn
P max |Xtnk − Xtnk+1 | ⩾ 2 ⩽ C2−(β−γα)n .
k=0,...,2n
with
−γn
An = max |X tn
k
−X tn
k+1
|<2 .
k=0,...,2n
In words, all but finitely many of the collection of events {An }n∈N take place, almost
surely. Let us therefore define the set
∞ \
[ ∞
Λ= Am .
n=1 m=n
Next, we claim that for every ω ∈ Λ there exists a C(ω) > 0 such that
|Xt − Xs |
sup γ
⩽ C(ω) .
s,t∈D,s̸=t |t − s|
We will prove (1.8) by induction. Suppose that m = n + 1 and w.l.o.g. consider the case
s < t. Then necessarily s = tm m m −m+1 ,
k , t = tk+1 for some k = 0, . . . , 2 − 1, since t − s < 2
and the claim follows from the definition of Am .
Next, suppose (1.8) is true for m and let us prove it for m + 1. There exist k < l ∈
{0, . . . , 2m+1 } such that s = tm+1
l , t = tm+1
k . Then we consider points s ⩽ s < t ⩽ t such
that in s, t ∈ Dm , constructed as follows
( (
tm+1
l−1 if l is odd, tm+1 if k is odd,
s = m+1 t = k+1
tl if l is even, tm+1
k if k is even.
1.3. CONSTRUCTION OF BROWNIAN MOTION 13
We can bound
|Xt − Xs | ⩽ |Xt − Xt | + |Xt − Xs | + |Xs − Xs |
Xm
−(m+1)γ
⩽ 22 +2 2−kγ
k=n+1
m+1
X
⩽2 2−kγ ,
k=n+1
as required. Now we can conclude that for any ω ∈ Λ there exists an n0 (ω) such that for
any m > n ⩾ n0 we have
4 8
|Xt − Xs | ⩽ −γ
2−γn ⩽ |t − s|γ , ∀t, s ∈ Dm : 2−n−1 ⩽ |t − s| ⩽ 2−n .
1−2 1 − 2−γ
We can conclude that for every ω ∈ Λ
|Xt − Xs |
sup γ
<∞. (1.9)
s,t∈D |t − s|
Step 2: Construction of X.
e Now we define
et (ω) = 0 , ∀ω ̸∈ Λ ,
X et (ω) = Xt (ω) , ∀ω ∈ Λ , t ∈ D .
X
And finally, for arbitrary t ∈ [0, 1], consider a sequence {tk }k∈N ⊆ D such that tk → t
as k → ∞. By (1.9), we have that limk→∞ Xtk (ω) exists for every ω ∈ Λ and does not
depend on the particular choice of converging sequence {tk }. Therefore, set
X
et (ω) = lim Xt (ω) ,
k
∀ω ∈ Λ .
k→∞
Interpolating linearly between the integer points we can define the continuous time process
St = S⌊t⌋ + t − ⌊t⌋ S⌊t⌋+1 − S⌊t⌋ , ∀t ⩾ 0 , (2.2)
which is a random function on the space of continuous paths C[0, ∞). We rescale space
and time in analogy to the CLT and set
Snt
Stn = √ for all t ⩾ 0 .
n
This defines a sequence of paths S 1 , S 2 , . . . ∈ C[0, ∞). Our aim in this chapter will be to
prove that this sequence converges to a Brownian motion in law (see the section below for
the definition of the concepts used in this result).
This result is also called a functional central limit theorem, since it is a functional
analogue of the CLT. The term ’invariance principle’ pertains to the fact that BM as the
scaling limit does not depend on the details of the original RW (apart from the normali-
√
sation of mean and variance), and is itself invariant (in law) under scaling: Bt = λBt/λ .
In other words, BM is a universal continuous limit for normalised RW’s.
Taking the increments Xk to be normalised is not a restriction. pFor a general RW Sn the
result would hold for the normalised RW Ŝn := Sn − nE[Xk ] / nVar[Xk ] in full analogy
to the standard CLT.
14
2.1. RANDOM WALKS AND WEAK CONVERGENCE 15
Exercise 2.1.2. Endowed with the metric d, the space C([0, ∞)) is complete and separable.
Proof. If A is closed, than Ct,A is closed and therefore lies in B(C([0, ∞))). If follows
that σ(Ct,A : t, A) ⊆ B(C([0, ∞))). For the converse inequality (say that we look only
at times in the interval [0, 1]), it suffices to show that an open ball Bf (δ), for any δ > 0
lies in σ(Ct,A : t, A). Here we observe that since we are considering continuous paths it
suffices to condition on all rational times.
Definition 2.1.4 (Canonical space). We say that (C([0, ∞)), B, P), with P being the law
of Brownian motion, is the canonical proability space for Brownian motion. We also refer
to P as the Wiener measure on the space of paths C([0, ∞); R).
For simplicity, we will later consider convergence only on a finite time horizon, that is
in the space C([0, 1]).
Definition 2.1.5. Let E be a metric space and {Xn }n∈N , X E-valued random variables.
d
Then Xn converges to X in distribution (or in law, or weakly), written Xn −→ X, if
E g(Xn ) → E g(X) as n → ∞
Although we will not make use of this, weak convergence is associated to a topology
on the space of probability measures, the so-called weak-∗ topology, and therefore there
is exists a natural concept of compactness. For our purposes we will only need sequential
compactness, therefore we define the following.
16 CHAPTER 2. UNIVERSALITY AND FUNCTIONAL CLTS
Definition 2.1.6. Let (S, ϱ) be a metric space and Π a collection of probability measures
on (S, B(S)).
1. We say that Π is compact, if for any sequence {Pn }n∈N ⊆ Π there exists a subse-
quence {nk }k∈N and a probability measure P ∈ Π such that
Pn → P weakly for n → ∞ .
2. We say that Π is tight if for every ε ∈ (0, 1) there exists a compact set Kε ⊆ S such
that
inf P(Kε ) ⩾ 1 − ε .
P∈Π
Then let {xi }i∈N be a countable dense set in R. We can find a subsequence {n1k }k such
that
lim Fn1 (x1 ) → F (x1 ) ,
k→∞ k
{nℓk }k ⊆ {nℓ−1 1
k }k ⊆ · · · ⊆ {nk }k ,
for some xℓ ∈ R. Choosing the diagonal sequence {nkk }k∈N we obtain that
By construction the limiting F is defined only on {xℓ }ℓ∈N and is increasing. It can therefore
be extended in a unique way to a cadlag function on the entire line R.
Our aim is to prove that the sequence converges weakly to a probability measure, so
we have to make sure that
lim F (x) = 1 .
x→∞
From tightness we know that for every ε > 0 there exists a xε ∈ R such that
F (x) ⩾ 1 − ε , ∀x ⩾ xε .
Proof. We only prove that the condition is sufficient. Since the space C([0, 1]) is a metric
space it suffices to check sequential compactness. Consider a sequence {ω n }n∈N ⊆ A. We
must prove that there exists a limit point ω ∈ C([0, 1]) for the sequence. By assumption
we find a constant C > 0 such that
sup |ωs | ⩽ C . (2.4)
0⩽s⩽1
From (2.4) we have that for every rational t ∈ Q ∩ [0, 1] the sequence ωtn is bounded. Let
{ti }i∈N = Q ∩ [0, 1] be an enumeration of the rationals.
Then there exists a subsequence {n1k }k∈N such that
n1
ωt1k → ωt1 ∈ R ,
for some ωt1 . We can refine the subsequence to obtain convergence also at t2 : there exists
{n2k }k ⊆ {n1k }k such that also
n2
ωtik → ωti , for i = 1, 2 .
Iterating the procedure we find nested sequences
{nℓk }k ⊆ {nℓ−1 1
k }k ⊆ · · · ⊆ {nk }k ,
such that
nℓ
ωtik → ωti , ∀i = 1, . . . , ℓ .
Choosing a diagonal sequence we obtain that
nk
ωtik → ωti , for k → ∞ , ∀i ∈ N .
Therefore, ω can be extended to a continuous function in C([0, 1]), and it can be checked
k
from the calculations above that the convergence ω nk → ω holds in C([0, 1]).
18 CHAPTER 2. UNIVERSALITY AND FUNCTIONAL CLTS
We can use this result to obtain a characterisation of tightness for measures on the
space of paths.
Theorem 2.1.10. A sequence {Pn }n∈N of probability measures on (C, B(C)) is tight if
and only if
lim sup Pn (ω : |ω0 | > λ) = 0 ,
λ↑∞ n∈N
Proof. As above, we prove only that the condition is sufficient, and we show the result
only on a compact time interval, say on C([0, 1]). Let us write m(ω, δ) for m1 (ω, δ). For
any ε > 0 we have to find a compact set Kε ⊆ C([0, 1]) such that inf n∈N Pn (Kε ) ⩾ 1 − 2ε.
Now by assumption, for any k ∈ N we can find a δk > 0 such that
P(A) ⩾ 1 − ε ,
sup m(ω, δk ) ⩽ k −1 ,
ω∈A
so that
lim sup m(ω, δ) = 0 .
δ↓0 ω∈A
In addition choose λε > 0 such that the set B = {|ω0 | > λ} satisfies
sup Pn (B) ⩽ ε .
n∈N
Building on the previous result, we can for example obtain a criterion for tightness
that is linked to the Kolmogorov–Chentsov continuity criterion of Theorem 1.3.7.
Proposition 2.1.11. Suppose that a sequence of probability measures {Pn }n∈N on C([0, 1])
satisfies for some α, β, γ, C > 0
Proof. Our aim is to apply Theorem 2.1.10. We therefore have to prove that
Indeed (2.5) holds true, then for any ε > 0 there exists n(ε, δ) ∈ N, δε ∈ (0, 1) such that
sup P m(S k , δε ) > η ⩽ ε .
k⩾nε (ε,δε )
Therefore
lim sup P (m(S n , δ) > η) ⩽ ε + lim sup P m(S k , δ) > η ⩽ ε ,
δ↓0 n δ↓0 k⩽n(ε,δε )
so that the claim follows. Therefore, it remains only to check (2.5). This will now follow
from Lemma 2.2.3 below. Indeed, recall that
tn
n 1 X
m(S , δ) = sup |Stn − Ssn | ⩽√ sup | Xj | ,
s,t∈[0,1], ,|t−s|⩽δ n 0⩽s⩽t⩽1, ,|t−s|⩽δ
j=sn
where
tn = ⌊nt⌋ + 1 , sn = ⌊sn⌋ .
The last quantity is furthermore upper bounded by
k+ℓ
1 X
√ sup | Xj | .
n k⩽n+1,ℓ⩽⌊nδ⌋+1
j=k
Proposition 2.2.2. For any t = (t1 , . . . , tm ) ∈ T we have that the following convergence
holds true weakly:
(Stn1 , . . . , Stnm ) → (Bt1 , . . . , Btm ) .
20 CHAPTER 2. UNIVERSALITY AND FUNCTIONAL CLTS
⌊nti ⌋
1 X
∆nti ,ti−1 =√ Xj
n
j=⌊nti−1 ⌋+1
1 1
+ √ (nti − ⌊nti ⌋)X⌊nti ⌋+1 − √ (nti−1 − ⌊nti−1 ⌋)X⌊nti−1 ⌋+1 .
n n
⌊nti ⌋
1 X
√ Xj → N (0, ti − ti−1 )
n
j=⌊nti−1 ⌋+1
we see that ∆nti ,ti−1 → ∆ti ,ti−1 = Bti − Bti−1 weakly. Since
⌊nti ⌋ ⌊ntk ⌋
1 X 1 X
√ Xj is independent of √ Xj
n n
j=⌊nti−1 ⌋+1 j=⌊ntk−1 ⌋+1
(Stni , ∆nt2 ,t1 , . . . , ∆ntn ,tn−1 ) → (Bt1 , ∆t2 ,t1 , . . . , ∆tn ,tn−1 ) ,
Proof of Theorem 2.1.1. Our aim is to prove that Pn → P weakly (where P is the law of
Brownian motion on C[0, 1]). Suppose that this is not the case, then, since by Proposi-
tion 2.2.1 the sequence is tight, and hence by Theorem 2.1.7 precompact, there must exist
a law Q ̸= P and a subsequence {nk }k∈N such that Pnk → Q as k → ∞ (weakly). But the
law of a stochastic process is uniquely characterised by its finite-dimensional distributions.
By Proposition 2.2.2, Q must have the FDDs of a Brownian motion, therefore we have
found a contradiction.
Lemma 2.2.3. Under the assumptions of Theorem 2.1.1 we have that for every ε > 0
√
lim lim sup P max |Sj+k − Sk | > ε n = 0 .
δ↓0 n→∞ 0⩽j⩽⌊δn⌋+1
0⩽k⩽n+1
2.2. DONSKER’S INVARIANCE PRINCIPLE 21
Proof. We split up the integer times up to n in bits of length ⌊nδ⌋ + 1. In particular, there
exists an m(δ, n) such that
And in addition, since j ⩽ ⌊nδ + 1⌋, either of the following two cases hold true
Assume that j, k assume the maximum in the event whose probability we are attempting
to bound. In the first case we must have that either of the following two hold:
1 √ 1 √
|Sj+k − S(⌊nδ⌋+1)(p−1) | > ε n , or |Sk − S(⌊nδ⌋+1)(p−1) | > ε n .
2 2
In the second case, one of the following three has to hold:
1 √
|Sj+k − S(⌊nδ⌋+1)p | > ε n , or
3
1 √
|Sk − S(⌊nδ⌋+1)(p−1) | > ε n , or
3
1 √
|S(⌊nδ⌋+1)p − S(⌊nδ⌋+1)(p−1) | > ε n .
3
In particular, we can conclude that
m(n,δ)+1
√ √
[
max |S − Sk | > ε n ⊆ max |Sj+(⌊nδ⌋+1)(p−1) − S(⌊nδ⌋+1)(p−1) | > ε n .
0⩽j⩽⌊δn⌋+1 j+k 0⩽j⩽⌊δn⌋+1
0⩽k⩽n+1 p=1
Therefore, our aim is now to prove (2.6). The idea is to use the CLT, which gaurantees
that (roughly)
√ √ ε2
lim P S⌊nδ⌋+1 ⩾ ε n ⩽ P(|Z| ⩾ δε) ≲ e− 2δ , (2.7)
n→∞
22 CHAPTER 2. UNIVERSALITY AND FUNCTIONAL CLTS
ε2
so that δ −1 e− δ → 0 as δ → 0. Of course, we first have to deal with the maximum over
the variable j. Let us introduce the stopping time
√
τ = inf{j : |Sj | > ε n} ,
η ∈ (0, ε) .
E |S⌊δn⌋+1 − Sj |2 |τ = j
√
P |S⌊δn⌋+1 | < n(ε − η)|τ = j ⩽
nη 2
⌊nδ⌋ + 1 1
⩽ E|X1 |2 2
⩽ ,
nη 2
where we could bound the expecation by using the independence of increments and the
last bound follows for n large by fixing
√
η = 3δ .
Px (·) := P[·|X(0) = x]
the law of the process started in x, and write Ex for the corresponding expectation.
Definition
3.1.1. [Filtration]A filtration on a probability space (Ω, F, P) is a family
Ft t⩾0 of σ-algebras such that F(s) ⊆ F(t) ⊆ F for all 0 ≤ s ≤ t.
The proof is immediate from the definition of F 0 (t). Intuitively, F 0 (t) contains all the
information from observing the process up to time t.
By analogy with Markov chains, we say that X possesses the Markov property if the
state of the process at time t ⩾ s, given its entire history up to time s, depends only on the
state Xs at time s. More precisely, an adapted process X : Ω → S is a Markov process
if
P[X(t) ∈ A | F(s)] = P[X(t) ∈ A | X(s)]
for all 0 ≤ s ≤ t and A ∈ S. Equivalently, we can define Markov property in terms of
transition kernels:
23
24 CHAPTER 3. BM AS A STRONG MARKOV PROCESS
Remarks.
• P (s, t, x, A) is the probability that the process takes a value in A at time t, provided
it is in state x at time s. Comparing the two equivalent definitions of a Markov
process, we find
P (s, t, x, A) = P X(t) ∈ A X(s) = x , t ≥ s ≥ 0 . (3.3)
• Note that for processes with continuous paths we have from (3.3)
1 , x∈A
P (t, t, x, .) = δx i.e. P (t, t, x, A) = .
0 , x ̸∈ A
• A Markov process w.r.t. its natural filtration is simply called a Markov process.
Definition 3.1.4. Two processes X and Y are called independent, if all finite dimen-
sional observations are independent, i.e. for all t1 , . . . , tn , n ∈ N and s1 , . . . , sm , m ∈ N
X(t1 ), . . . , X(tn ) ⊥ Y (s1 ), . . . , Y (sm ) .
Remarks.
• For a BM B we have (S, S) = (R, B) and Px is the law of the process t 7→ x + B(t),
called BM started in x.
Remark 3.1.6. The Gaussian transition kernel density (or just density) is translation
invariant, i.e.
p(t, x, y) = p(t, 0, y − x) for all t ≥ 0, x, y ∈ R . (3.5)
Proof. We know that the time-shifted process is still a Brownian motion, and independence
of B(u) : 0 ≤ u ≤ s follows directly from independence of increments of Brownian
motion, which proves (ii).
The full proof of part (i) can be based on finite-dimensional approximations of BM.
Here is a skeleton of a more direct proof: for any t, s ≥ 0, A ∈ B,
Z
0 0
P[B(t + s) ∈ A | F (s)] = E[1A (Bt+s ) | F (s)] = E[1B(t+s)∈A δx (B(s)) | F 0 (s)]dx
Z R
The equality (a) is due to the measurability of B(s) w. r. t. F 0 (s); (b) is due to the
independence of the increment B(t + s) − B(s) from F 0 (s); (c) is due to B(t + s) − B(s) ∼
N (0, t); (d) follows from (ii). The last equality proves that BM is a Markov process,
the penultimate inequality yields the transition kernel. To make the above rigorous one
needs to approximate δx with a sequence of bounded continuous functions and analyze
the sequence of corresponding expectations.
26 CHAPTER 3. BM AS A STRONG MARKOV PROCESS
Example 3.1.7. Brownian motion with drift t 7→ X(t) := B(t) + ct for some c ∈ R
is a also a time-homogeneous Markov process. Using the Markov property of BM we have
X(t+s)|F (s) = B(t+s)|B(s) +c(t+s) = B(t+s)−B(s) B(s) +ct+X(s) ∼ X(s)+N ct, t ,
1 −(y−x−ct)/(2t)
p(t, x, y) = √ e .
2πt
Now, the following generalisation of Theorem 3.1.5 is useful for finding transition
kernels explicitly:
The proof follows the same steps as the one of Theorem 3.1.5.
\
F + (t) := F 0 (s) ⊇ F 0 (t) .
s>t
1. The augmented filtration is larger (or finer) than the natural one, and F + (t) contains
infinitesimal information about the future, such as the event
ω : t 7→ X(t, ω) is differentiable at t .
2. By Theorem 3.1.5, for every s ≥ 0 the process B(t + s) − B(s) : t ≥ 0 is indepen-
dent of F 0 (s). The next result shows that this independence can be extended to the
augmented filtration.
Theorem 3.2.3. For every s ≥ 0 the process B(t + s) − B(s) : t ≥ 0 is independent of
F + (s).
Extending this result to all measurable integrable functions (measure theory details omit-
ted), we get
B(t1 + s) − B(s), . . . , B(tm + s) − B(s) ⊥ F + (s) ,
0 (t) : t ≥ 0 ,
Definition 3.2.4. For a process X on (Ω, F, P) with natural filtration F
F + (0) = t>0 F 0 (t) is called the germ σ-algebra, and the tail σ-algebra is defined as
T
\
T := σ X(s) : s ≥ t .
t≥0
Consider then the time inversion bijection ψ : C0 → C0 defined by ψω(t) = tω(1/t), t > 0
and ψω(0) = 0. In Proposition 1.3.5 we have shown that the process B◦ψ is again BM, and
by definition tail events are mapped to germ events and vice versa, so that ψT = F + (0).
So tail events for B are germ events for B ◦ ψ , and can therefore only have probability 0
or 1 according to Blumenthal’s 0-1 law.
• We will see in an exercise sheet that for tail events A ∈ T of Brownian motion,
Px [A] is in general independent of x, whereas for a germ event A ∈ F + (0) it may
depend on x.
As a first application we show that BM takes positive and negative values and zeros
in every small time interval to the right of 0, and we look at properties of local and global
extrema.
Theorem 3.2.8. For a Brownian motion B define τ := inf t > 0 : B(t) > 0 and
λ := inf t > 0 : B(t) = 0 . Then
P0 [τ = 0] = P0 [λ = 0] = 1 .
so by Theorem 3.2.5 we just have to show that the event in question has positive proba-
bility. This follows as
P0 [τ ≤ t] ≥ P0 [B(t) > 0] = 1/2
for all t > 0. Hence P0 [τ = 0] ≥ 1/2 and we have shown the first part. The same argument
works by replacing B(t) > 0 with B(t) < 0, and this implies also that P0 [λ = 0] = 1 using
continuity of paths and the intermediate value property.
Theorem 3.2.9. For a Brownian motion B, almost surely the following hold:
(a) Every local maximum is a strict local maximum.
(b) The set of times where the local maxima are attained is countable and dense.
Conditioning on the values of m1 − B(b1 ) and m2 − B(a2 ) the right-hand side is constant
while the left-hand side is a continuous random variable (recall that a2 > b1 ), which implies
that P[m1 = m2 ] = 0. Applying this result to a countable set of intervals with rational
endpoints, we find
P any two non-overlapping non-degenerate compact time intervals
with rational endpoints have different maxima = 1 . (3.6)
1
The second and the third variables are not Gaussian, as times of maxima are not stopping times, see
below.
3.3. THE STRONG MARKOV PROPERTY 29
(a) If BM had non-strict local maxima, then there would exist two such intervals where
BM has the same maximum, which is almost surely not the case.
(b) Theorem 3.2.8 implies that the maximum over any non-degenerate compact time
interval with rational endpoints is not attained at an endpoint. So each such interval
contains a local maximum, and the set of times ∆ ⊆ [0, 1] where they are attained is
dense. Since each local maximum is strict, ∆ has at most the cardinality of the set of such
intervals and is therefore countable.
(c) Due to (3.6), almost surely, for any q ∈ [0, 1] ∩ Q the maxima on [0, q] and [q, 1]
are different. If the global maximum were attained at two points t1 < t2 one could find
t1 < q < t2 for which the maxima on [0, q] and [q, 1] would agree, which has probability
zero.
Definition 3.3.1. A random variable T with values in [0, ∞], defined on a filtered prob-
ability space (Ω, F, [F(t)], P) is called a stopping time with respect to the filtration F(t) :
t ≥ 0 , if {T ≤ t} ∈ F(t) for all t ≥ 0.
• On the other hand, if T is a stopping time w.r.t. F(t) : t ≥ 0 define for any n ≥ 1
i.e. we approximate the value of T by the smallest dyadic integer k2−n > T . Then
Tn are stopping times w.r.t. F(t) : t ≥ 0 since, for t ∈ [m/2n , (m + 1)/2n )
∞
[
Tn ≤ t = T < m/2n = T ≤ m/2n − 1/q ∈ F(t) .
q=1
• Every stopping time T w.r.t. the natural filtration F 0 (t) : t ≥ 0 is also a stopping
time w.r.t. the augmented filtration F + (t) : t ≥ 0 , since F 0 (t) ⊆ F + (t) for all
t ≥ 0.
30 CHAPTER 3. BM AS A STRONG MARKOV PROCESS
An important property distinguishing the augmented from the natural filtration is the
right-continuity of the former:
Lemma 3.3.3. Consider a process with natural filtration F 0 (t) : t ≥ 0 . Then the
As usual we want to replace quantifiers by countable unions and intersections and write
∞
\ [ [ n 1 1 o
{TH ≤ t} = B(s) ∈ x − , x + ∈ F 0 (t) ,
n
n n
n=1 s∈Q∩[0,t] x∈Q∩H | {z }
∈F 0 (s)
H n = {x ∈ R : (x − 1/n, x + 1/n) ∩ H ̸= ∅} .
Finally, suppose that G is bounded and that the starting point x ̸∈ Ḡ. Then we can
fix a path γ : [0, t] → R with γ(0, 0
t) ∩ Ḡ = ∅ and γ(t) ∈ ∂G. Then the σ-algebra F (t)
contains no non-trivial subset of B(s) = γ(s) for all 0 ≤ s ≤ t , i.e. no subset other than
the empty set and the set itself. If we had {TG ≤ t} ∈ F 0 (t) then
B(s) = γ(s) for all 0 ≤ s ≤ t, TG = t
From now on, we will silently assume that a stopping time is defined with respect
to a right continuous filtration such as an augmented filtration. Moreover, we will be
checking that T is a stopping time either using strict or non-strict inequalities, as for a
right continuous filtration {T < t} ∈ F(t) ⇐⇒ {T ≤ t} ∈ F(t).
Definition 3.3.5. For any stopping time T we define the σ-algebra
This means that the part of A that lies in {T ≤ t} should be measurable w.r.t. the
information available at time t. Heuristically, this is the collection of events that hap-
pened before the stopping time T . One can also check that the path B(t) : 0 ≤ t ≤ T
is F + (T )-measurable. Now we can state the strong Markov property for BM, which was
rigorously established by Hunt (1956) and Dynkin (1957).
Theorem 3.3.6 (Strong Markov property). For every almost surely finite stopping time
T , the process
of the process B(T + u) − B(T ) : u ≥ 0 are independent and normally distributed
with mean zero and variance s. Since its paths are obviously almost surely continuous the
process is BM. Moreover, all increments (3.8), and hence the process itself, are independent
of F + (T ).
is Brownian Motion.
Proof. The hitting time Ta = inf{t ≥ 0 : B(t) = a is a stopping time (cf. Proposi-
tion 3.3.4). Then let B ∗ (t) : t ≥ 0 be BM reflected at Ta . The event {Mt > a} can be
written as a disjoint union:
M (t) > a = B(t) > a ∪ B(t) ≤ a, M (t) > a
3.5. MARKOV PROCESSES RELATED TO BROWNIAN MOTION 33
The second event coincides with B ∗ (t) > a , and hence the first statement follows from
the reflection principle. For the second statement we have the disjoint union
|B(t)| > a = B(t) > a ∪ B(t) < −a
Remark 3.4.4. This result implies that the PDF fM for the maximum M (t) of BM up
to time t is given by
2 −x2 /(2t)
fM (x) = √ e for x > 0 .
2πt
In combination with Gaussian tail estimates, we obtain
√
2t 2
P0 M (t) > a ≤ √ e−a /(2t) for all a, t > 0 .
a π
Ta < ∞ , ∀a ∈ R .
since Mt is an increasing process, and therefore {Mt > a} an increasing sequence of sets.
Now
√
P(Mt > a) = P(|B1 | ⩾ a/ t) → 1 ,
as t → ∞. The nullset can be chosen uniformly over all a, by considering a countable set
of a’s. As for the second claim, we observe that
1 − (y−x)2 (y+x)2
p(t, x, y) = √ e 2t + e− 2t for x, y ⩾ 0 .
2πt
Next, we observe that the sequence of stopping times (Ta )a∈R forms itself a Markov process.
34 CHAPTER 3. BM AS A STRONG MARKOV PROCESS
Theorem 3.5.1. Consider a Brownian Motion B(t) : t ≥ 0 and define Ta := inf t ≥
0 : B(t) = a for any a ≥ 0. Then the process Ta : a ≥ 0 is an increasing Markov
process with transition kernel density
a a2
p(a, t, s) = p exp − 1 for all a > 0 .
2π(s − t)3 2(s − t) {s>t}
By the strong Markov property for BM this event is independent F + (Tb ), and therefore
in particular of (Td : d ≤ b) ∈ F + (Tb ). This implies the Markov property for Ta : a ≥ 0 ,
• The transition densities are homogeneous (notice that ’time’ and ’space’ for the pro-
cess T are interchanged
in comparison with Brownian Motion). Therefore the process
Ta : a ≥ 0 has stationary and independent increments. However, it’s paths are not
continuous.
• Note that for t = 0 this result implies a heavy tail (with ’index’ 1/2)
a 2a
p(a, 0, s) ≃ s−3/2 √ P a, 0, [s, ∞) ≃ s−1/2 √
and as s → ∞ .
2π 2π
So the expected hitting time for BM is E[Ta ] = ∞ for all a ̸= 0 (using symmetry for
negative a). Intuitively, even though Ta < ∞ almost surely for all a, large positive
or negative excursion of BM lead to a diverging contribution to the expectation.
• Another simple but useful consequence for later is that monotonicity of the process
and independence of increments implies that Ta → ∞ as a → ∞ almost surely. We
summarize all useful direct consequences of Theorem 3.5.1 in the next result.
3.5. MARKOV PROCESSES RELATED TO BROWNIAN MOTION 35
Corollary 3.5.3. Consider BM B(t) : t ≥ 0 and let Ta := inf t ≥ 0 : B(t) = a for
a ∈ R. Then
which can be established using the reflection principle (see Q. 2.8 of the assignment for
the main computational step) and time reversibility of BM.
Remark 3.5.5. It is worth noticing that the maximum process itself is not Markov.
Chapter 4
|B(t + h) − B(t)|
lim sup sup ≤1. (4.1)
h↘0 0≤t≤1−h ϕ(h)
p
Our first result shows that one can choose c > 0 so small that c h log(1/h) fails to be a
modulus of continuity.
√
Theorem 4.1.1. For every constant c < 2, almost surely, for every ϵ > 0 there exists
0 ≤ h ≤ ϵ and t ∈ [0, 1 − h] such that
p
B(t + h) − B(t) ≥ c h log(1/h) .
√
Proof. Let c < 2 and fix ϵ > 0. We have to show that P[Aϵ ] = 1 with
p
Aϵ := ∃h ≤ ϵ ∃t ∈ [0, 1 − h] : B(t + h) − B(t) ≥ c h log(1/h) .
For integers k, n ≥ 0 define the events
n √ o
Ak,n := B (k + 1)e−n − B ke−n ≥ c n e−n/2 .
and in the following we can use a lower bound on P[Ak,n ] to show the the left-hand side
has probability 1. Using stationarity of increments, scaling and the Gaussian tail estimate
we get independently of k ≥ 0
√
−n √ −n/2 √ c n 1 2
√ e−c n/2 .
P[Ak,n ] = P B e > c ne = P B(1) > c n ≥ 2
c n + 1 2π
36
4.2. VARIATION OF BROWNIAN PATHS 37
√
Since c < 2 this implies that en P[Ak,n ] → ∞ as n → ∞. Therefore, using 1 − x ≤ e−x
for all x ∈ R, the probability that for given n none of the Ak,n occurs satisfies
n −1⌋
⌊e\
en
Ack,n = 1 − P[A0,n ] ≤ exp − en P[A0,n ] → 0
P as n → ∞ ,
k=0
Theorem 4.1.2. There exists a constant C > 0 such that, almost surely, for all sufficiently
small h > 0 and all 0 ≤ t ≤ 1 − h,
p
B(t + h) − B(t) ≤ C h log(1/h) . (4.3)
The proof is omitted. One can determine the optimal constant c for the best possible
modulus of continuity which achieves
p equality in (4.1). √
The above lower bound on the
increments implies that ϕ(h) ≥ 2h log(1/h), and indeed 2 can be shown to be optimal.
|B(t + h) − B(t)|
Theorem 4.1.3. Almost surely, lim sup sup p =1.
h↘0 0≤t≤1−h 2h log(1/h)
Proof. See pp. 16-17 in [?]. A stronger statement with a lim replacing lim sup in the l. h.
s. of (4.1) can also be proved.
the p-variation sum for Π of a function f : [0, ∞) → R. The supremum over all finite
partitions is called the p-variation
Note that the limit may in principle depend on the choice of the partition.
38 CHAPTER 4. MORE PATH PROPERTIES OF BROWNIAN MOTION
Remark 4.2.2. Note that the total variation (4.4) is always defined, and if (4.5) exists
VARp (f ; t) ≥ varp (f ; t). If f is continuous it is no restriction to require t0 = 0 and tn = t,
and we can focus on partitions with ti ∈ Q or any other dense set for i < n. Indeed, for
all ϵ > 0 and every partition Π = (ti : P 0 ≤ i ≤ n) we can find Π′ = (qi : 0 ≤ i ≤ n) with
points from the dense subset such that ni=0 f (ti ) − f (qi ) < ϵ. Thus
′
SpΠ (f ; t) − SpΠ (f ; t) ≤ cp ϵ
Remark 4.2.4. The 2-variation var2 (B; t) := limn→∞ S2Πn is called quadratic variation
of BM. The above theorem says that the sequence S2Πn (B; t) converges to t in L2 (hence
in probability), for any sequence of finite partitions with vanishing mesh.
Proof. Let Πn = {t0 = 0 < t1 < . . . < tn = t} be some partition of [0, t], then we have
n n
X 2 X
E S2Πn (B; t) =
E B(tj ) − B(tj−1 = (tj − tj−1 ) = t .
j=1 j=1
Therefore,
n
2 X 2
E S2Πn (B; t) − t = Var S2Πn (B; t) =
Var B(tj ) − B(tj−1 ) ,
j=1
2
where we used the independence of increments. Since B(tj ) − B(tj−1 ) ∼ B(tj − tj−1 )2
with mean tj − tj−1 , we get with the scaling property
2
= (tj − tj−1 )2 E B(1)4 − 1 = (tj − tj−1 )2 (3 − 1) ,
Var B(tj ) − B(tj−1 )
where the last identity uses the value of the 4-th Gaussian moment (see problem sheet 1).
This implies
n
Πn 2 X
E S2 (B; t) − t ≤ 2|Πn | (tj − tj−1 ) = 2|Πn |t → 0
j=1
Corollary 4.2.5. Almost surely, Brownian motion paths are of infinite total variation.
In fact VARp (B; t) = ∞ almost surely for all p < 2.
4.2. VARIATION OF BROWNIAN PATHS 39
Proof. Let p = 2 − δ for some δ > 0 and Πn any sequence of finite partitions of [0, t] with
|Πn | → 0 as n → ∞. Then
X 2 δ
X 2−δ
B(tj ) − B(tj−1 ≤ max B(tj ) − B(tj−1 ) B(tj ) − B(tj−1
tj−1 ,tj ∈Πn
tj−1 ,tj ∈Πn tj−1 ,tj ∈Πn
δ
≤ max B(tj ) − B(tj−1 ) VAR2−δ (B; t) .
tj−1 ,tj ∈Πn
By Theorem 4.2.3 the left-hand side converges along a subsequence almost surely to t (a
sequence that converges in probability admits a subsequence that converges almost surely).
On the other hand,
δ
max B(tj ) − B(tj−1 ) →0 almost surely as n → ∞ ,
tj−1 ,tj ∈Πn
by uniform continuity of Brownian paths on [0, t]. This implies VAR2−δ (B; t) = ∞ almost
surely.
Under additional assumptions on the sequence of partitions one can establish almost
sure convergence for the quadratic variation (without passing to subsequences).
Theorem 4.2.6 (Almost sure convergence of quadratic variation). P∞Let B(t) : t ≥ 0 be
BM and (Πn : n ≥ 0) a sequence of finite partitions of [0, t] with n=1 |Πn | < ∞. Then
(n)
Let Aϵ = {ω : |S2Πn (B(ω); t) − t| > ϵ}. It follows from (4.6) and Borel-Cantelli lemma
that almost surely, only finitely many of the A(n)ε can kick in. Therefore, almost surely,
there exists Nϵ (ω): such that for any n > Nϵ (ω), we have that |S2Πn (B(ω); t) − t| ≤ ϵ. In
other words, P[Ωϵ ] = 1, where Ωϵ = {ω : lim supn→∞ |S2Πn (B(ω); t) − t| ≤ ϵ}.
The event of interest Ω0 := {ω : limn→∞ S2Πn (B(ω); t) = t} can be presented as the
limit Ω0 = ∩∞k=1 Ω1/k . Therefore,
∞
X
P[Ω0 ] = 1 − P[Ωc0 ] =1− P[∪∞ c
k=1 Ω1/k ] ≥1− P[Ωc1/k ] = 1.
k=1
It is natural to ask what the asymptotic smallest envelope for BM is, namely what is
a deterministic function ψ : (0, ∞) → R such that lim supt→∞ B(t)/ψ(t) = 1? Such ψ, if
it exists, determines the almost sure asymptotic growth of a Brownian Motion.
Theorem 4.3.2 (Law of the iterated logarithm). For BM B(t) : t ≥ 0 we have, almost
surely,
B(t)
lim sup √ =1.
t→∞ 2t log log t
Remark 4.3.3. By symmetry of BM, it follows that, almost surely,
B(t)
lim inf √ = −1 .
t→∞ 2t log log t
Proof. We have to prove the following two almost sure statements:
p
∀ϵ > 0 ∃t0 > 0 : ∀t ≥ t0 B(t) ≤ (1 + ϵ) 2t log log t , (4.7)
and p
∀ϵ > 0 ∀t0 > 0 ∃t ≥ t0 : B(t) ≥ (1 − ϵ) 2t log log t . (4.8)
We will establish (4.7) and (4.8), using Borel
√ Cantelli Lemmas and scaling along a geo-
metric sequence of times q n . Define ψ(t) := 2t log log t.
By Proposition 3.4.3, the maximum of BM up to time t has the same distribution as |B(t)|,
so we can write h |B(q n )| ψ(q n ) i
P[An ] = P √ n ≥ (1 + ϵ) √ n .
q q
n √ n
By scaling, |B(q )|/ q is standard Gaussian, and we can use the Gaussian tail estimate
2
from Lemma 4.3.4 (P[Z ⩾ x] ≤ x−1 e−x /2 ), to obtain
2 log log q n 2
P[An ] ≤ 2e−(1+ϵ) = 2(n log q)−(1+ϵ) .
This is summable in n, and the (first) Borel Cantelli Lemma implies that, almost surely,
only finitely many events An occur, i.e. P[An .i.o.] = 0. So there is Nω < ∞ a.s.: for all
n > Nω and t such that q n−1 ≤ t ≤ q n , we have
√ √
since ψ(t)/t = 2 log log t/ t is decreasing in t. This implies
B(t)
≤ (1 + ϵ)q for all t > q Nω +1 ,
ψ(t)
and since this holds for all q > 1, (4.7) is proved.
Lower bound. In order to use the (second) Borel Cantelli Lemma in the other direction,
we have to construct a relevant sequence of independent events. Again, fix q > 1, and
define
Dn := B(q n ) − B(q n−1 ) ≥ ψ(q n − q n−1 ) .
From the tail estimate for a standard Gaussian Z we have a constant C > 0 such that
2
P[Z ≥ z] ≥ Ce−z /2 /z for large enough z. Using this estimate and monotonicity of log we
get
n n−1
h ψ(q n − q n−1 ) i Ce− log log(q −q )
P[Dn ] = P Z ≥ p ≥p
q n − q n−1 2 log log(q n − q n−1 )
Ce− log(n log q) C̃
≥p > √
2 log(n log q) n log n
P
for a further constant C̃ and all n large enough. Therefore n≥1 P[Dn ] = ∞ and so, by
the independence of Dn ’s and the (second) Borel Catelli Lemma we have, almost surely,
for infinitely many n ≥ 1
Here the second inequality follows from applying the upper bound established above to
−B(t), which implies in particular that B(q n ) ≥ −2ψ(q n ) for all n large enough. Therefore
we get, almost surely, for infinitely many n
B(q n ) −2ψ(q n−1 ) + ψ(q n − q n−1 ) −2 q n − q n−1 2 1
≥ ≥ √ + =1− √ − . (4.9)
ψ(q n ) ψ(q n ) q qn q q
To obtain the second inequality we used for the first term
√ n
ψ(q n−1 ) ψ(q n−1 ) q 1 1
n
= p n √ ≤√ ,
ψ(q ) q n−1 ψ(q ) q q
√
since ψ(t)/ t is increasing in t for large enough t. For the second term we have an
analogous argument using that ψ(t)/t is decreasing in t. Now, (4.9) implies that a. s.
there is a sequence of times tn → ∞
B(tn ) 2 1
≥1− √ − ≥1−ϵ ,
ψ(tn ) q q
for any ϵ > 0 and q large enough. The lower bound (4.8) is proved.
The proof of the result above relies on the following Gaussian tail estimate:
Lemma 4.3.4. Let X ∼ N (0, 1). Then there exist 0 < c1 ⩽ c2 for all x > 1
1 2 1 2
c1 e−x /2 ≤ P[X > x] ≤ c2 e−x /2 .
x x
42 CHAPTER 4. MORE PATH PROPERTIES OF BROWNIAN MOTION
Proof. We have Z ∞
1 y2
P(X > x) = √ e− 2 dy .
x 2π
Then Z ∞
2 /2 2 /2 2 /2−λ2 /2−λx
e−y dy ⩾ e−(x+λ) λ = e−x λ.
x
The lower bound follows by choosing λ = 1/x. For the upper bound
Z ∞ Z ∞
2 y −y2 /2 1 2
e−y /2 dy ⩽ e dy ⩽ e−x /2 .
x x x x
Corollary 4.3.1 implies that the growth of B(t) as t → ∞ is almost surely bounded by
tα for any α > 1/2, while the LIL implies that this is sharp and the lim sup growth of B(t)
is faster than tα for any α ≤ 1/2, which we summarize in a corollary.
B(t) B(t)
almost surely, lim sup = +∞ and lim inf = −∞ .
t→∞ tα t→∞ tα
Proof. Since
B(t) 1/2−α B(t) p
= t √ log log t
tα t log log t
the statement for α = 1/2 implies the general case for α ≤ 1/2, and the latter follows
directly from the LIL and the fact that limn→∞ log log n = ∞.
In addition to the extreme case α = 1/2, the case α = 0 shows that BM paths them-
selves of course do not converge but keep taking values arbitrarily large in positive and
negative directions. Using the time inversion property we can use the LIL to understand
continuity properties of BM in typical time points, as opposed to the modulus of continu-
ity in Theorem 4.1.1 which concerns the corresponding lim sup on a compact time interval.
Corollary 4.3.6. Suppose B(t) : t ≥ 0 is a BM. Then for all t ≥ 0 we have, almost
surely,
|B(t + h) − B(t)|
lim sup p =1.
h↘0 2h log log(1/h)
The proof is left as an exercise.
Chapter 5
• Martingales are processes where the current state is always the best prediction for
the future, and their origin comes from modelling fair games. Martingales are not
necessarily Markov processes and vice versa.
• Standard +
BM B(t) : t ≥ 0 is a martingale w.r.t. F (t) : t ≥ 0 , since obviously
E |B(t)| < ∞ for any t ≥ 0 and for all 0 ≤ s ≤ t
Here we use independence of increments and E[B(s)|F + (s)] = B(s) since B(s) is
F + (s)-measurable.
43
44 CHAPTER 5. BROWNIAN MOTION AS A MARTINGALE
using independence
of increments and exponential Gaussian moments, which also
imply E |M θ (t)| < ∞ for any t ≥ 0.
The next two results are useful basic facts about continuous martingales which have
almost surely continuous paths, such as Brownian Motion. The first result provides con-
ditions under which the defining property of martingales can be extended to stopping
times.
Theorem 5.1.3. Let X(t) : t ≥ 0 be a continuous
martingale, and 0 ≤ S ≤ T stopping
times,
all w.r.t. a filtration F(t) : t ≥ 0 . If the stopped process X(t ∧ T ) : t ≥
by an integrable random variable X, i.e. X(t ∧ T ) ≤ X a.s. ∀t ≥
0 is dominated
0 and E |X| < ∞ , then
E X(T ) F(S) = X(S) almost surely .
Proof in a slightly simpler setting. We prove a simpler version where S = 0 and for sim-
plicity we will also assume that the process X itself is bounded, namely that
|X(t)| ≤ X, a.s. ∀t ≥ 0,
where X is an integrable random variable: E[X] < ∞.
First we establish that
E X(t ∧ T ) F(0) = X(0) for all t ≥ 0 , (5.1)
by using a discrete approximation of T for each n ∈ N,
T n = tnm := (m + 1)2−n if m2−n ≤ T < (m + 1)2−n for all m ≥ 0 .
These are also stopping times with T n ↘ T as n → ∞. Now fix n and t ≥ 0, and let
t̄n : min{tnm : tnm > t}. Then we can write
E X(t ∧ T n ) F(0) = E X(T n )1T n ≤t F(0) + E X(t)1Tn >t F(0)
X
E X(tnm )1T n =tnm F(0) + E X(t)1T n ≥t̄n F(0) .
=
m:tn
m <t̄
n
Using {T n = tnm } = {T n ≤ tnm } \ {T n ≤ tnm−1 }, we can set up a telescopic sequence for the
sum on the right-hand side
E X(tnm )1T n =tnm F(0) = E X(tnm )1T n ≤tnm F(0) − E X(tnm )1T n ≤tnm−1 F(0)
where we used the tower property and E X(tnm ) F(tnm−1 ) = X(tnm−1 ) in the second line.
= X0 ,
5.1. BASIC PROPERTIES OF CONTINUOUS MARTINGALES 45
by using the martingale property also to treat the last term. Now, since X has continuous
paths, X(t ∧ T n ) → X(t ∧ T ) almost surely. Next, since by assumption |X(t ∧ T n )| ≤ X,
by the dominated convergence theorem,
taking the n → ∞ limit leads to (5.1).
Finally, since X(t ∧ T ) : t ≥ 0 is also dominated by an integrable random variable
X we can use the dominated convergence theorem to take the limit t → ∞ to conclude
E X(T ) F(0) = X(0) . (5.2)
The proof of the full statement under the original set of assumptions can be found
in Moerters & Peres on pages 54 and 354 and is based on the theory of discrete-time
martingales.
Note that the condition on domination of the stopped process by an integrable random
variable could be replaced by requiring bounded stopping times, i.e. for some K > 0 we
have T ≤ K almost surely. This is particularly easy to see if E[X 2 (t)] < ∞ using Doob’s
maximal inequality, which we recall below.
Theorem 5.1.4 (Doob’s Lp maximal inequality). Suppose X(t) : t ≥ 0 is a continuous
Therefore
p p p
E sup |B(s)| ⩽ E sup B(s) + E sup (−B(s))
0≤s≤t 0≤s≤t 0≤s≤t
p
= 2E sup B(s) = 2E [|B(t)|p ] .
0≤s≤t
d
The last identity is due to sup0≤s≤t B(s) ∼ |B(t)| (reflection principle).
The next remark is so useful that we will state is as a lemma.
Lemma 5.1.5. Let X(t) : t ≥ 0 be a continuous martingale and T a stopping time
w.r.t. a filtration F(t) : t ≥ 0 . Assume that E[X(t)2 ] < ∞ for all t ≥ 0. Then the
stopped process
t 7→ X(t ∧ T ) is a martingale .
Proof. Since T is a stopping time, t 7→ X(t ∧ T ) is adapted to F(t) : t ≥ 0 . As
E |X(t ∧ T )| ≤ E sup |X(s)| ≤ E[( sup X(s))2 ]1/2 ≤ 4E |X(t)|2 < ∞,
0≤s≤t 0≤s≤t
for all t ≥ 0, where we used Cauchy-Schwarz inequality at the second step and Doob’s
maximal inequality at the third. Therefore, the process X(t ∧ T ) is integrable.
For s < t we have
E X(t ∧ T ) F(s) =E X(t ∧ T ) 1T ≤s + 1T >s F(s) = E X(T )1T ≤s F(s)
+ E X(t ∧ T )1T >s F(s) = X(T )1T ≤s + X(s)1T >s = X(s ∧ T ) ,
where we have used optional stopping to simplify the second term. The use of optional
stopping is justified as the stopping time T ∧ t is bounded, see the remark above.
46 CHAPTER 5. BROWNIAN MOTION AS A MARTINGALE
Lemma 5.2.1 (Wald’s first lemma). Let B(t) : t ≥ 0 be standard BM and T ⩾ S be
two stopping times such that either
Then we have
E B(T )|FS = BS .
Proof. Under condition (b) we can directly apply the optional stopping theorem to get
E[B(T )|FS ] = BS .
It remains to show that (a) implies (b). Suppose E[T ] < ∞ and define
⌈T ⌉
X
Mk := max B(t + k) − B(k) and M = Mk .
0≤t≤1
k=0
⌈T ⌉
X X∞ ∞ h
X i
E E 1T >k−1 Mk F + (k − 1) + E[M0 ]
E[M ] = E Mk = E 1T >k−1 Mk =
k=0 k=0 k=1
∞
X
= E[M0 ] 1 + P[T > k] ≤ E[M0 ] 1 + E[T ] < ∞ ,
k=0
where we used the tower property, Mk ⊥ F(k − 1) and Mk ∼ M0 for all k ≥ 1 to get to the
penultimate inequality. The second equality is due to the monotone convergence theorem.
To show that E[M0 ] is finite, let’s use reflection principle:
P⌊T ∧t⌋
Finally, B(t ∧ T ) ≤ k=0 max0≤t≤1 B(k + t) − B(k) ≤ M , so that (b) holds.
Corollary 5.2.2. Let S ≤ T be stopping times and (Bt )t a Brownian motion with respect
to a common filtration (Ft )t⩾0 , and assume that E[T ] < ∞. Then
Let us observe that we do not claim that any of the quantities above are finite.
5.2. WALD’S LEMMAS 47
Note that using the tower property, E[T ] < ∞ implies E T −S F + (S) < ∞ almost surely.
so that the middle term in (5.3) vanishes. Note that in order to use the tower property
(and hence Wald’s lemma) we must first make sure that
BS (BT − BS ) ∈ L1 ,
although proving that either the positive or the negative part lies in L1 would be sufficient.
To obtain the integrability we can follow the construction in the proof of Lemma 5.2.1,
and we would obtain two inependent positive random variables MS , MT such that
B S ⩽ MS , BT − BS ⩽ MT ,
Here as usual we imply that the martingale property holds with respect to the filtration
F = (F + (t) : t ≥ 0).
Proof. The process is adapted to F + (t) : t ≥ 0 and E |B(t)2 − t| < 2t < ∞ for all
t ≥ 0. Furthermore,
E B(T )2 = E[T ] .
Proof. Consider the quadratic martingale M2 (t) : t ≥ 0 and define the stopping time
Tn := inf t ≥ 0 : |B(t)| = n ,
by monotone convergence
E B(T )2 ≥ lim E B(T ∧ Tn )2 = lim E[T ∧ Tn ] = E[T ] ,
n→∞ n→∞
⌈t⌉
as the integrand P B(s) ∈ (a, b) for all s ∈ [0, t] ≤ maxx∈(a,b) Px [B(s) ∈ (a, b), ∀s ∈ [0, 1]] ,
which decays exponentially fast with time. (Use Markov property recursively to prove the
last estimate. ) Now, by Wald’s second Lemma 5.2.4 and (a) we obtain
b |a|
E[T ] = E B(T )2 = a2 + b2
= |a|b .
|a| + b |a| + b
5.3. FROM MARTINGALES TO HARMONIC FUNCTIONS 49
Remark 5.2.7. Using the exponential martingale M θ (t) : t ≥ 0 Wald’s lemmas can
also be used to infer the moment generating function for the law of exit times (this is on
a problem sheet).
1 −(y−x)2 /(2t)
p(t, x, y) = √ e ,
2πt
∂ 1 ∂2 1
p(t, x, y) = p(t, x, y) = ∆y p(t, x, y) for all x ∈ R , (5.4)
∂t 2 ∂y 2 2
This is also known as Dynkin’s formula or Kolmogorov forward equation, and the operator
1 2
2 ∆ on C0 (R, R) is also called the infinitesimal generator of Brownian motion.
More generally, in the following let (S, d) be a compact or locally compact, separable
metric space (e.g. S = Rd with L2 -metric), with Borel σ-algebra S induced by all open
sets w.r.t. the metric d. Denote by C(S, R) all continuous, real-valued functions on S that
f (xn ) → 0 for all sequences (xn : n ∈ N)
vanish at infinity in case S is not compact, i.e.
such that d(xn , 0) → ∞. Then C(S, R), ∥.∥∞ is a Banach space.
Proposition 5.3.1. Let X(t) : t ≥ 0 be a time-homogeneous Markov process with state
space S and transition kernel p(t, x, dy) as in Definition 3.1.3, and assume that for all
f ∈ C(S, R) also Pt f ∈ C(S, R), where
Z
(Pt f )(x) := f (y)p(t, x, dy) = Ex f (X(t)) for all t ≥ 0 .
S
If this is also continuous, i.e. ∥Pt f − f ∥∞ → 0 as t ↘ 0 for all f ∈ C(S, R), the process
is called a Feller Markov process. The infinitesimal generator L : D ⊆ C(S, R) → C(S, R)
of the process is then the operator
1
Lf (x) := lim Pt f (x) − f (x) with domain D ⊆ C(S, R) ,
t↘0 t
50 CHAPTER 5. BROWNIAN MOTION AS A MARTINGALE
Here we have used the Markov property and regularity assumptions to exchange expecta-
tion and integration. Now, using integration by parts and the diffusion equation (5.4) for
the transition kernel we get for the integrand of (∗)
Z
1 1
E ∆f (B(u)) = p(u, B(s), y)∆f (y) dy
2 B(s) 2 R
Z Z
1 ∂
= ∆y p(u, B(s), y)f (y) dy = p(u, B(s), y)f (y) dy .
2 R R ∂u
Using the regularity assumptions (5.6) to exchange integrals and the Markov property we
get
Z Z t−s
∂
(∗) = lim dy p(u, B(s), y)f (y)
ϵ↘0 R ϵ ∂u
Z Z
= dyp(t − s, B(s), y)f (y) − lim dyp(ϵ, B(s), y)f (y)
R ϵ↘0 R
= EB(s) f (B(t − s)) − f (B(s)) ,
which after cancellation of two terms implies E M f (t) F(s) = M f (s) as required.
Example 5.3.5. Using f (x) = x gives the obvious martingale M f (t) = B(t), and f (x) =
x2 the quadratic martingale M f (t) = B(t)2 − t. On a problem sheet we will see further
examples with monomials f (x) = xk .
Corollary 5.3.6. Suppose f ∈ C 2 (R, R) is harmonic, i.e. ∆f (x) ≡ 0. Then the process
t 7→ f (B(t)) is a martingale .
This follows immediately from Theorem 5.3.4. Harmonic functions in R are not too
exciting and simply linear of the form f (x) = a + bx, a, b ∈ R, the regularity assumptions
(5.6) are automatically fulfilled. But the same concept also applies in higher dimensions
where it can be applied in a useful way as we will see later.
Chapter 6
In this chapter we will discuss connections between Brownian Motion and harmonic func-
tions and recurrence and transience properties of Brownian Motion depending on the
dimension d.
d
X ∂2
∆u(x) := u(x) = 0 for all x ∈ U .
j=1
∂x2j
Remark 6.1.2.
• The latter conservation property is best appreciated using Green’s identity, which is
a special case of Stokes’ Theorem: For any (open) ball B(x, r) ⊆ U centered in x
with radius r, we have
Z Z
∂u
∆u(y) dy = (y) dσ(y) .
B(x,r) ∂B(x,r) ∂n
Here σ denotes the surface measure on the sphere ∂B(x, r) defined in the next bullet
∂u
point, and ∂n (y) = ∇u(y) · n(y) where n(y) is the unit outer normal vector (of
Euclidan length 1) at the point y ∈ ∂B(x, r).
52
6.1. HARMONIC FUNCTIONS AND BROWNIAN MOTION 53
π d/2
Z
dy = rd Vd
L B(0, r) = where Vd := L B(0, 1) = ,
B(0,r) Γ(d/2 + 1)
2π d/2
σ(∂B(0, r)) = drd−1 Vd = rd−1 Sd−1 where Sd−1 = dVd = ,
Γ(d/2)
• Examples of harmonic functions include (for the last cases the domains do not con-
tain the origin)
Here and in the following we simply write |x| := ∥x∥2 for the Euclidean norm of
x ∈ Rd .
(a) u is harmonic.
Z
1
(b) For any ball B(x, r) ⊆ U we have u(x) = u(y) dy.
L(B(x, r)) B(x,r)
Z
1
(c) For any ball B(x, r) ⊆ U we have u(x) = u(y) dσ(y).
σ(∂B(x, r)) ∂B(x,r)
Sketch of proof. (b)⇒(c): Assume u : U → Rd satisfies (b) and fix x ∈ U . Then there
exists R > 0 such that B(x, R) ⊆ U and we define
Z
1−d
ψ(r) := r u(y) dσ(y)
∂B(x,r)
54 CHAPTER 6. BROWNIAN MOTION IN HIGHER DIMENSIONS
Differentiating w.r.t. r gives ψ(r) = dVd u(x) for almost all 0 < r ≤ R. Since drd−1 Vd =
σ(∂B(x, r)) this implies (c) for almost all 0 < r ≤ R.
It remains to show that u is differentiable so that (c) holds for all 0 ≤ r < R and we can
infer (a) as well. To Rsee this, pick a smooth function g : [0, ∞) → [0, ∞) with compact
support in [0, R) and [0,∞)d g(|x|) dx = 1. Multiply (c) by g(r) and integrate over r to get
Z R Z R Z
u(x) σ(∂B(x, r))g(r) dr = g(r) dr u(y) dσ(y)
|0 R
{z } |
0 ∂B(x,r)
{z }
= g(|x|) dx=1
Rd
R
= Rd
u(y) g |y−x| dy
to obtain Z
u(x) = u(y) g |y − x| dy .
Rd
With appropriately chosen g, u can be written as a convolution of itself with a smooth
function for all x ∈ U , which implies that u ∈ C ∞ (U ) (using partial integration on the
r.h.s. to compute derivatives).
(c)⇒(b): Multiply (c) with r and integrate over 0 < r < R to obtain (c) for any R > 0
such that B(x, R) ⊆ U .
(c)⇒(a): We established above that (c) implies u ∈ C ∞ (U ). Suppose that ∆u ̸= 0 in U ,
so by continuity of ∆u there exists a ball B(x, r) ⊆ U such that either ∆u(y) > 0 or < 0
for all y ∈ B(x, r). But with the notation from above (c) implies that
Z Z
′ 1−d ∂u
0 = ψ (r) = r (y) dσ(y) = ∆u(y) dy ̸= 0, (6.1)
∂B(x,r) ∂n B(x,r)
which is a contradiction. Here we have used Green’s identity for the second equality, and
the polar coordinate representation of the integral to compute ψ ′ (r).
(a)⇒(c): Suppose u is harmonic and B(x, R) ∈ U . Then (6.1) implies that ψ ′ (r) = 0 for
all r < R, so ψ is constant and given by
If u is harmonic, the theorem applies to both u and −u, so that all extrema on bounded
domains is attained at the boundary. For d = 1 this is consistent with harmonic functions
being affine.
Therefore equality holds and u(y) = M for almost all y ∈ B(x, r), and by continuity of u
this implies that B(x, r) ⊆ V . So V is open. V ∈ U is both relatively closed and open,
V ̸= ∅. Since U is connected, V = U which implies that u ≡ M is constant on U .
(b) Since u is continuous on U which is closed and bounded, u attains its maximum on
Ū . With (a), the maximum has to be attained on ∂U .
so u1 (x) ≤ u2 (x) for all x ∈ U . The same works for u2 − u1 which is also harmonic and
vanishes on ∂U , and we get u1 (x) = u2 (x) for all x ∈ U .
x ∈ U and
τ = τ (∂U ) := min{t ≥ 0 : B(t) ∈ ∂U }
the first hitting time of the boundary. Let ϕ : ∂U → R be measurable, and such that the
function u : U → R with
u(x) := Ex ϕ(B(τ )) 1τ <∞ for all x ∈ U ,
Proof. For fixed x ∈ U choose a ball B(x, r) ∈ U and let τ̃ := inf t > 0 : B(t) ̸∈ B(x, r) .
Then the tower property and strong Markov property for BM imply
h i
u(x) = Ex Ex ϕ(B(τ )) 1τ <∞ F + (τ̃ ) = Ex u(B(τ̃ )) .
R
By symmetry of BM the last expectation is simply given by ∂B(x,r) u(y) dσ(y)/σ(∂B(x, r)).
This implies the mean value property in Theorem 6.1.3 (c) and hence u is harmonic.
While uniqueness of the solution to the Dirichlet problem follows directly from the
maximum principle for harmonic functions in Corollary 6.1.6, to ensure existence some
regularity of the domain is needed to avoid over-specification at the boundary.
Theorem 6.2.3. Let U ⊆ Rd be a bounded domain that satisfies the cone condition in
every point x ∈ ∂U , and let ϕ : ∂U → R be a continuous function. Let τ := τ (∂U ) =
inf{t ≥ 0 : B(t) ∈ ∂U }, which is an almost surely finite stopping time. Then u : U → R
given by
u(x) := Ex ϕ(B(τ )) for all x ∈ Ū (6.4)
is the unique solution to the Dirichlet problem (6.3).
Remark 6.2.4.
but we will not use any explicit representation in the following. The only important
property will be that Cx (α) ∩ B(x, 1) has positive Lebesgue measure for any α > 0.
• Most domains that come to mind satisfy the cone condition, examples include open
balls or any other regular shape. A typical example that does not satisfy the cone
condition is given below.
6.2. THE DIRICHLET PROBLEM 57
Example 6.2.5. Consider the solution v of the Dirichlet problem on the unit ball B(0, 1) ∈
Rd with constant boundary data ϕ(x) ≡ ϕ ∈ R, which implies that v(x) ≡ ϕ.
Now let U = B(0, 1) \ {0} be the punctured ball, which does obviously not satisfy the
cone condition at 0 ∈ ∂U . In dimension d = 1, U is not a domain (not connected), and
the Dirichlet problem happens to have a solution on U which is the linear interpolation
between ϕ and the boundary value ϕ(0) on both intervals [−1, 0] and [0, 1].
In higher dimensions d ≥ 2, U is a domain, and in general the Dirichlet problem on U
does not have a solution. Since BM in higher dimensions almost surely does not hit points
(see later) we have τ (∂U ) = τ (∂B(0, 1)) and (6.4) implies that the solution on U agrees
with u(x) = v(x) = ϕ for all x ̸= 0.
In order to prove Theorem 6.2.3 we require a technical Lemma, which asserts that if
you start Brownian Motion infinitesimally close to the base of a cone, you hit that cone
with probability 1 before you exit a ball of size 1.
Lemma 6.2.6. Let 0 < α < 2π and C0 (α) ⊆ Rd be a cone based at the origin with opening
angle α, and
a := sup Px τ (∂B(0, 1)) < τ (C0 (α)) .
x∈B(0,1/2)
Proof. It is easy to see from a picture (but a bit harder to prove, so we omit it) that a
lower bound for 1 − a is given by the probability of the event that BM started at x in the
interior of a cone of opening angle α/2, hits ∂B(x, 3/2) before leaving the cone, which has
positive probability. This implies that a < 1.
If x ∈ B(0, 2−k ) then by the strong Markov property
k−1
Y
Px τ (∂B(0, 2−k+i+1 )) < τ (C0 (α)) = ak ,
Px τ (∂B(0, 1)) < τ (C0 (α)) ≤ sup
−k+i )
i=0 x∈B(0,2
which follows from a nested argument with BM hitting spheres of increasing radius 2−k+i
and then restarting. Therefore, for any integer k ≥ 1 and h′ > 0, scaling of BM and
translation to z ∈ Rd implies the statement.
Proof of Theorem 6.2.3. Uniqueness of the solution follows directly from Corollary 6.1.6.
By Theorem 6.1.7 u is bounded and harmonic on U , and it remains to show that the
Poincaré cone condition implies that u is continuous on the boundary ∂U . For all x ∈ Ū
and z ∈ ∂U we have
h i
u(x) − u(z) = Ex ϕ(B(τ (∂U ))) − ϕ(z) ≤ Ex ϕ(B(τ (∂U ))) − ϕ(z) . (6.5)
To bound this from above, fix z ∈ ∂U , then there is a cone Cz (α) with Cz (α)∩B(z, h) ⊆ U c .
By Lemma 6.2.6, for any integer k ≥ 1 and h′ > 0, we have
for all x ∈ Rd with |x − z| < 2−k h′ . Also, since ϕ is continuous on ∂U , for given ϵ > 0
there exists 0 ≤ δ < h such that
Since ϵ > 0 and k ≥ 1 can be chosen arbitrarily, this implies continuity of u on Ū , which
finishes the proof.
Recall from Corollary 5.3.6 that harmonicity of u implies that u(B(t)) is a martingale, so
that in particular
u(x) = Ex u(B(t)) = Ex u(B(t))1t<τ (H) + Ex u(B(t))1t≥τ (h) .
The same relationship holds for u(y) when B(t) is replaced by B(t) and with (6.6)
u(x) − u(y) = Ex u(B(t))1t<τ (H) − Ey u(B(t))1t<τ (h) ≤ 2M Px t < τ (H) → 0
as t → ∞. This holds since by isotropy (see problem sheet) the projection of d-dimensional
BM along any direction is one-dimensional BM, so that with Theorem 3.5.1 the hitting
times of hyperplanes are finite almost surely. Thus u(x) = u(y) and since x and y were
chosen arbitrarily, u must be constant.
the inner boundary of A with τR analogously for the outer boundary, and τ = τr ∧ τR is
hitting time of ∂A; ϕ(r) and ϕ(R) are constants, which determine spherically symmetric
boundary condtions on the inner and outer boundaries of A correspondingly. We know
from Theorem 3.5.1 that Brownian motion in R almost surely hits every point in finite
time, so considering only the first coordinate of B(t) with x ∈ A we have τr , τR < ∞
almost surely, leading to the above representation of u(x). Hence, we obtain
ϕ(R) − u(x)
Px [τr < τR ] = for all x ∈ Ā ,
ϕ(R) − ϕ(r)
so that by solving the Dirichlet problem (6.8) we can get the following result.
(
1 , if d ≤ 2
Px [τr < ∞] = rd−2 .
|x|d−2
, if d ≥ 3
Proof. Proof of Theorem 6.3.1 We simply have to solve the Dirichlet problem
(6.8), and
with rotation invariant boundary data we make the ansatz u(x) = ψ |x|2 for some C 2
function ψ : [0, ∞) → R. Then we get
and with y = |x|2 > 0 this can be written as ψ ′′ (y) = −dψ ′ (y)/(2y), which in turn
reduces to
d 2
∂y log ψ ′ (y) = − ⇐⇒ ∂y log [(ψ ′ )− d (y)] = ∂y log (y)
2y
This is solved by ψ ′ (y) = const. y −d/2 leading to the radially symmetric harmonic functions
|x| , if d = 1
u(x) = C1 + C2 log |x| , if d = 2 ,
|x|2−d , if d ≥ 3
as we have already seen earlier. Fixing the constants to satisfy the boundary conditions
leads to (6.9).
• Point recurrent, if, almost surely, for all x ∈ Rd there exists a (random) sequence
Tn ↗ ∞ such that X(Tn ) = x for all n ∈ N,
60 CHAPTER 6. BROWNIAN MOTION IN HIGHER DIMENSIONS
• Neighbourhood recurrent, if, almost surely, for all x ∈ Rd and ϵ > 0 there exists a
(random) sequence Tn ↗ ∞ such that X(Tn ) ∈ B(x, ϵ) for all n ∈ N,
Note that the events in the above definition are tail events, so by the 0-1 law for
BM (which also holds in Rd ) it follows that BM is either transient or recurrent. Point
recurrence trivially implies neighbourhood recurrence.
P[Acn i.o.] = 0 .
That means that almost surely there exists N ∈ N such that for all n ≥ N , An occurs
which implies |B(t)| → ∞ as t → ∞.
Remark 6.3.5.
• It can be shown that the range of BM is a (random) fractal with Hausdorff dimension
2 in any space dimension d ≥ 3.
6.3. RECURRENCE AND TRANSIENCE OF BM 61
and note that R(x) is independent of Y for all x. Therefore we can compute
Z
h i 1 2
e−|x| /2 E R(x) dx ,
0 = E L2 B[0, 1] ∩ B[2, 3] = E R(Y ) =
2π R2
where we integrate w.r.t. the Gaussian distribution of B(2) − B(1). Therefore, for L2 -
almost all x ∈ R2 we have R(x) = 0 almost surely, and therefore by Fubini’s theorem
One can show (see Lemma 2.23 in Moerters & Peres) that this property
of translation
for general Borel sets implies that almost surely, either L2 B[0, 1] = 0 or L2 B[2, 3] =
0. Since both
are identically distributed and independent implies that both vanish and
L2 B[0, 1] = 0 almost surely.
62 CHAPTER 6. BROWNIAN MOTION IN HIGHER DIMENSIONS
where the second identity follows from symmetry of BM. It remains to show that by
excluding time t = 0 from therange this in fact holds for all x ∈ R2 . For any ϵ > 0 we
thus have, almost surely, PB(ϵ) y ∈ B[0, 1] = 0. Hence, for any x ∈ R2
Px y ∈ B[0, 1] = lim Px y ∈ B[ϵ, 1] = lim Ex PB(ϵ) y ∈ B[0, 1 − ϵ] = 0 ,
ϵ↘0 ϵ↘0
Then, almost surely, for any t > 0 the measure µt is absolutely continuous with respect to
the Lebesgue measure.
Proof. To prove absolute continuity it suffices to prove that P−almost surely the following
holds:
µt (Bx,ε )
lim sup < ∞ , for µt − almost all x ∈ R .
ε→0 L1 (Bx,ε )
Therefore it suffices to prove that
Z
µt (Bx,ε )
E lim sup dµt (x) < ∞ .
R ε→0 L1 (Bx,ε )
where in the last step we used scale invariance of BM together with the fact that the PDF
of a standard Gaussian is bounded from above by 1. Since
Z tZ t
1
p dr ds < ∞ , ∀t > 0 ,
0 0 |r − s|
We observe that
Z t Z Z t
νx,t (A) = Px (Bs ∈ A) ds = ps (x, y) ds dy ,
0 A 0
so that the expected occupation time measure admits an explicit density with respect to
the Lebesgue measure:
Z t
dνx,t
(y) = ps (x, y) ds .
dL1 0
Depending on the dimenision, we see different behaviors of the expected occupation mea-
sure as t → ∞.
Instead in d ⩾ 3 we have Z
lim νx,t (A) = g(x, y) dy ,
t→∞ A
with
Γ(d/2 − 1)
g(x, y) = d |x − y|2−d .
2π 2
The proof of this result is left as an exercise in a problem sheet. Next, we show that
the kernel g appearing in Theorem 6.3.10 is harmonic outside the diagonal x = y.
Proof. Let us fix a point x ̸= y and a ball Bx,ε ⊆ Rd \ {y}. Then it suffices to check the
mean value property: Z
1
g(x, y) = g(z, y) dz .
|Bx,ε | Bx,ε
Indeed, if we denote with
we find that
Z ∞ Z Z ∞
1
g(x, y) = Ex δy (Bt ) dt = Ez δy (Bt ) dt dz
τ |Bx,ε | Bx,ε 0
Z
1
= g(z, y) dz ,
|Bx,ε | Bx,ε
For completeness, we observe that one can prove that g is a fundamental solution to
the Poisson problem on Rd , for d ⩾ 3:
1
∆x g(x, y) + δy (x) = 0 , ∀x ∈ Rd .
2
∂
p(t, x, y) = ∆y p(t, x, y) with initial condition p(0, x, dy) = δx (dy) .
∂t
The same is true in higher space dimensions d > 1 (see problem sheet 4), and since for
the Gaussian transition kernel ∆y p(t, x, y) = ∆x p(t, x, y) an analogous backward heat
equation holds
∂ 1
p(t, x, y) = ∆x p(t, x, y) .
∂t 2
Multiplying this equation with f (y) and integrating over y implies that
Z
u(t, x) := Ex f (B(t)) = p(t, x, y)f (y) dy
Rd
R
solves the heat equation with initial data u(0, x) = Rd f (y) δx (dy) = f (x). This can be
extended to more general versions of the heat equation with a dissipation term, as is made
precise in the following.
h Z t i
u(t, x) := Ex f (B(t)) exp V (B(s)) ds (6.11)
0
solves the heat equation on Rd with dissipation rate V and initial condition f .
Proof. For V ≡ 0 this corresponds to the backward heat equation (5.4) discussed above,
and the easiest proof is a direct calculation. P Expanding the exponential in (6.11) in a
power series we get for the terms u(t, x) = ∞ n=0 an (t, x), a0 (t, x) = Ex [f (B(t))] and for
n≥1
Z t Z t
1 h i
an (t, x) := Ex f (B(t)) ··· V (B(t1 )) · · · V (B(tn )) dt1 . . . dtn
n! 0 0
h Z t Z t Z t i
= Ex f (B(t)) dt1 dt2 · · · dtn V (B(t1 )) · · · V (B(tn ))
0 t1 tn−1
Z Z Z t Z t n
Y n
Y
= dx1 · · · dxn dt1 · · · dtn V (xi ) p(ti − ti−1 , xi−1 , xi )
Rd Rd 0 tn−1 i=1 i=1
Z
dy p(t − tn , xn , y)f (y)
Rd
Z Z t
= dx1 V (x1 ) dt1 p(t1 , x, x1 )an−1 (t − t1 , x1 )
Rd 0
with the convention x0 = x and t0 = 0. Using the backward heat equation 12 ∆x p(t1 , x, x1 ) =
∂t1 p(t1 , x, x1 ) and integrating by parts we get
Z Z t
1
∆x an (t, x) = dx1 V (x1 ) dt1 ∂t1 p(t1 , x, x1 )an−1 (t − t1 , x1 )
2 Rd 0
Z Z t
=− dx1 V (x1 ) dt1 p(t1 , x, x1 )∂t1 an−1 (t − t1 , x1 ) − V (x)an−1 (t, x)
Rd 0
Z
+ V (x1 )p(t, x, x1 )an−1 (0, x1 ) dx1
R
= ∂t an (t, x) − V (x)an−1 (t, x) .
P
Since f and V are bounded the sum n an (t, x) is absolutely convergent for all t ≥ 0 and
x ∈ Rd and we can exchange differentiation with the summation. Summing up all terms
verifies the validity of the heat equation with dissipation for u(t, x). The initial condition
is of course
u(0, x) = Ex f (B(0)) = f (x) .
Chapter 7
The Hausdorff dimension and measure of a set are intrinsic notions for a metric space
(E, ϱ). For a set F ⊆ E we define its diameter to be
|F | = sup{ϱ(x, y) : x, y ∈ F } ,
and we denote with {Ei }i∈N a covering of E if
∞
[
E⊆ Ei .
i=1
Definition 7.0.1. For any α ⩾ 0 and any metric space E we define the α–Hausdorff
content of E as
(∞ )
X
α α
H∞ (E) = inf |Ei | : {Ei }i∈N is a covering of E .
i=1
66
67
Then we define
Hα (E) = sup Hδα (E) = lim Hδα (E) .
δ>0 δ↓0
Remark 7.0.4. The Hausdorff measure satisfies the following two properties:
Proof. We only prove the second claim (the first one is obvious). Fix any ε ∈ (0, 1) and
choose δ(ε) ∈ (0, 1) such that
! !
[ [
α α
H Ei ⩽ Hδ Ei + ε .
i∈N i∈N
! !
[ [
α α
H Ei ⩽ ε + Hδ Ei
i∈N i∈N
X X ε
⩽ε+ |Fik | ⩽ Hδα (Ei ) +
2i−1
i,k i
( )
X
⩽ Hα (Ei ) + 2ε ,
i
Next, we can connect the Hausdorff measure to the Hausdorff dimension through the
following result.
1. Hα (E) = 0 ⇐⇒ H∞
α (E) = 0.
α
H∞ (E) = 0 ⇒ Hα (E) = 0 .
Now, for any ε > 0 there exists a covering {Ei }i∈N of E such that
X
|Ei |α ⩽ ε .
i
1
It therefore follows that |Ei | ⩽ ε α for all i ∈ N so that also Hα1 (Ei ) < ε. Sending
εα
ε → 0 we obtain the desired result.
2. The second claim follows now directly from the definition of Hausdorff dimension
and the first claim.
Then the result follows since the right hand-side vanishes for δ → 0.
Definition 7.1.1. A function f : (E1 , ϱ1 ) → (E2 , ϱ2 ), for two metric spaces (Ei , ϱi ) is
α–Hölder continuous with constant C ∈ (0, ∞) (for some α ∈ (0, 1]) if
ϱ2 (f (x), f (y))
sup =C.
x̸=y∈E1 ϱ2 (x, y)
1
dim(f (E1 )) ⩽ dim(E1 ) .
α
7.2. UPPER BOUNDS ON THE HAUSDORFF DIMENSION 69
Proof. For any δ, ε > 0 let {Fkε,δ }k∈N be a δ-covering of E1 such that
Hδαβ (E1 ) ⩾
X ε,δ
|Fk |αβ − ε .
k
|f (Fkε,δ )| ⩽ C|Fkε,δ |α .
Therefore
β
|Gε,δ αβ
X
β β
HCδ α (f (E1 )) ⩽ k | ⩽ C (Hδ (E1 ) + ε) .
k
dim(B(A)) ⩽ 2dim(A) ∧ d ,
Theorem 7.2.1. Let d ⩾ 2 and (Bt )t⩾0 be d-dimensional Brownian motion. Then
dim(B[0, ∞)) = 2 .
It follows from Corollary 7.1.3 that dim(B[0, ∞)) ⩽ 2 for d ⩾ 2, so the only thing
left to prove is a lower bound on the dimension. To obtain a lower bound we use a so-
called “energy method”, which is reminiscent of Kolmogorov’s continuity criterion. We
start with a simple variant of this energy method, which is dubbed the “mass distribution
principle”.
Theorem 7.2.2 (Mass distribution principle). Suppose that (E, ϱ) is a metric space and
µ is a Borel measure on E such that for constants C, δ > 0 and α ⩾ 0 it holds that
and
µ(V ) ⩽ C|V |α .
for all closed sets V with |V | ⩽ δ. Then
µ(E)
Hα (E) ⩾ >0,
C
and hence dim(E) ⩾ α.
70CHAPTER 7. HAUSDORFF DIMENSION OF THE RANGE OF BROWNIAN MOTION
We will now find a more sophisticated criterion, which does not require the condition
µ(V ) ⩽ C|V |α , which is sometimes hard to check. We define the α-energy associated to a
measure µ by: Z
1
Iα (µ) = α
dµ(x) dµ(y) ∈ (0, ∞] .
E×E ϱ(x, y)
Then we can prove the following result.
Theorem 7.2.3 (Energy method). Let α ⩾ 0 and µ be a Borel measure such that
Now we want to compare the latter quantity with µ(E). Here by Cauchy–Schwartz we
have that
X X α µ(Ui )
µ(E) ⩽ µ(Ui ) = |Ui | 2 α
i i
|Ui | 2
!1 !1
X 2 X (µ(Ui ))2 2
⩽ |Ui |α .
|Ui |α
i i
We conclude that
X µ(E)2 µ(E)2
|Ui |α ⩾ P (µ(U ))2 ⩾ R R −α dµ(x) dµ(y)
,
ϱ(x,y)⩽ε ϱ(x, y)
i
i i |Ui |α
as required.
Now we are going to apply this method to study the range of Brownian motion in
d ⩾ 2. In particular, we prove the following proposition.
Proposition 7.2.4. For any d ⩾ 2 and (Bt )t⩾0 a d−dimensional Brownian motion we
have
dim(B[0, ∞)) ⩾ 2 .
7.2. UPPER BOUNDS ON THE HAUSDORFF DIMENSION 71
Proof. Of course, we must define an appropriate measure. We consider here the (random)
occupation measure Z t
µB,t (A) = 1A (Bs ) ds .
0
Clearly
t = µB,t (B[0, t]) ∈ (0, ∞) ,
so there is hope to apply Theorem 7.2.3. In order to apply the theorem it suffices to prove
that for any α ∈ (0, 2) we have
Z
1
E α
dµB,t (x) dµB,t (y) < ∞ .
Rd ×Rd |x − y|