0% found this document useful (0 votes)
79 views13 pages

1 Preliminaries: 1.1 Dynkin's π-λ Theorem

1. The document provides definitions and theorems related to stochastic processes and probability theory. 2. It defines stochastic processes as collections of random variables indexed by time, and introduces concepts like indistinguishability and modifications. 3. It proves that if two right-continuous stochastic processes are equal almost surely at each time point, then they are indistinguishable.

Uploaded by

vdoblin
Copyright
© Attribution Non-Commercial (BY-NC)
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
79 views13 pages

1 Preliminaries: 1.1 Dynkin's π-λ Theorem

1. The document provides definitions and theorems related to stochastic processes and probability theory. 2. It defines stochastic processes as collections of random variables indexed by time, and introduces concepts like indistinguishability and modifications. 3. It proves that if two right-continuous stochastic processes are equal almost surely at each time point, then they are indistinguishable.

Uploaded by

vdoblin
Copyright
© Attribution Non-Commercial (BY-NC)
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 13

1 Preliminaries

Collection of useful facts.


Lemma 1 (Jensens Inequality, Proposition 1.4.2 in [3]). Suppose that X is an integrable R-
valued random variable on (, T, ) and is a convex function. Then (E[X]) E[(X)].
Proposition 1 (Operations that preserve measurability, see here). Suppose that (f
n
), f and g
map from some measurable space to R and that R. Then
f +g, f, fg, f g, f g, limsup
n
f
n
, liminf
n
f
n
are all measurable. So if (f
n
) converges point-wise to f we have that f is measurable, since
f = limsup
n
f
n
= liminf
n
f
n
.
Theorem 1 (Functional Monotone Class Theorem, theorem 1.5 p.277 of [1]). Let / be a -
system (that is, closed under nite intersections) that contains and let 1 be a collection of
real-valued functions on that satises:
If A /, then 1
A
1.
If f, g 1, then f +g, cf 1 for any real number c.
If (f
n
) 1 is a a sequence of non-negative, increasing functions that tend pointwise to a
bounded function f, then f 1.
Then 1 contains all bounded functions that are measurable with respect to (/).
1.1 Dynkins - Theorem
Denition 1 ( and -systems). Let be a set. A -system of is a collection of subsets of
that is closed under nite intersections. A -system T is a collection of subsets of such that
T and A, B T such that A B implies that B A T and such that for any sequence
of increasing sets A
1
A
2
. . . contained in T, the union of the sets is also contained in T.
Lemma 2 (Dynkins - Theorem, Lemma 1.7.2 in [3]). Let ( be a -system on some set
and let T := ( be the smallest -system on such that ( T. Then T = (.
Corollary 1 (Lemma 1.7.3 in [3]). Let (, T) be a measurable space, and let ( be a -system
such that T = (. If two probability measures P and Q agree on (, that is, if we have that
P(A) = Q(A) for all A (, then P and Q are equal (that is P(A) = Q(A) for all A T.
1.2 Convergence Theorems
Denition 2. A collection of random variables X

on a probability space (, T, P) is said


to be uniformly integrable if c implies that
sup

_
|X

|c
[X
n
[dP 0.
Theorem 2 (Vitalis Convergence Theorem, Theorem 4 on page 188 of [2]). Let (X
n
) be a
sequence of R-valued integrable random variables. Suppose that (X
n
) is uniformly integrable and
that X
n
X almost surely. Then X is integrable and E[[X
n
X[] 0 as n (which, by
Jensens inequality, implies that E[X
n
] E[X] as n ).
1
1.3 Conditional expectation
Denition 3. Let (, T, ) be a probability space and let X be an R
n
-valued integrable (that
is, E[[X
i
[] < for all i) random variable and let ( be any sigma-algebra. Then E[X[(] is the
unique
1
(-measurable random variable that satises the relation E[1
A
X] = E[1
A
E[X[(]] for all
events A (.
Conditional expectations have the following handy properties:
Proposition 2 (Theorem 2.3.2. in [3]). Let X, Y be real valued integrable functions on (, T, ),
let (, 1 T be sigma-algebras, and let , R. Then
1. Linearity: E[X +Y [(] = E[X[(] +E[Y [(].
2. If X 0 almost surely, then E[X[(] 0 almost surely.
3. Tower property: E[E[X(][1] = E[X[1] almost surely. Note that this implies that
E[X] = E[X[, ] = E[E[X[(][, ] = E[E[X[(]].
4. If X is (-measurable and XY is integrable, then E[XY [(] = XE[Y [(] almost surely. Thus
E[X[(] = XE[1

[(] = X almost surely.


5. If 1 and X, ( are independent, then E[X[(, 1] = E[X[(] almost surely. Thus
E[X[1] = E[X[1, , ] = E[X[, ] = E[X] almost surely.
6. Monotone and dominated convergence theorems, Fatous lemma, and Jensens inequality
all hold if we replace in them E[X] with E[X[(].
1
Uniqueness is a consequence of the Radon-Nikodym Theorem, see [3].
2
2 Stochastic processes, ltrations and stopping times
2.1 Stochastic processes, indistinguishability and modications
I start these notes on stochastic calculus with the denition of a continuous time stochastic
process. Very simply, a stochastic process is a collection of random variables X
t

t0
dened
on a probability space (, T, P). That is, for each time t 0, X
t
() is a measurable
function from to the real numbers.
Remark 1 (by Georgy Lowther). Stochastic processes may also take values in any measurable
space (E, c) but, in these notes, I concentrate on real valued processes. I am also restricting
to the case where the time index t runs through the non-negative real numbers R
+
, although
everything can easily be generalized to other subsets of the reals.
A stochastic process X := X
t

t0
can be viewed in either of the following three ways.
As a collection of random variables, one for each time t 0. As a collection of paths
R
+
R,
t X
t
(),
one for each . These are referred to as the sample paths of the process.
As a function from the product space
R
+
R,
(t, ) X
t
().
As is often the case in probability theory, we are not interested in events which occur with zero
probability. The theory of stochastic processes is no dierent, and two processes X and Y are
said to be indistinguishable if there is an event A of probability one such that X
t
() = Y
t
()
for all A and all t 0. This is the same as saying that they almost surely (that is, with
probability one) have the same sample paths. Alternative language which is often used is that
X and Y are equivalent up to evanescence. In general, when discussing any properties of a
stochastic process, it is common to only care about what holds up to evanescence. For example,
if a processes has continuous sample paths with probability one, then it is referred to as a
continuous (resp., right-continuous and left-continuous) process and we dont care if it actually
has discontinuous (resp. non-right-continuous and non-left-continuous) paths on some event of
zero probability.
It is important to realize that even if we have two processes X and Y such that for each
xed t, the X
t
= Y
t
occurs with probability one, it is not necessarily the case that they are
indistinguishable. As an example, consider a random variable T uniformly distributed over the
interval [0, 1], and dene the process X
t
= 1
{t=T}
. For each time t, P(X
t
,= 0) = P(T = t) = 0.
However, X is not indistinguishable from the zero process, as the sample path t X
t
() always
has one point at which X takes the value 1. The problem here is the uncountability of non-
negative real numbers used for the time index. By countable additivity of measures, if X
t
= Y
t
almost surely, for each t, then we can infer that the sample paths of X and Y are almost surely
identical on any given countable set of times S R
+
, but cannot extend this to uncountable
sets.
This necessitates a further denition. A process Y is a version or modication of X if, for
each time t 0, P(X
t
= Y
t
) = 1. Alternative language sometimes used is that X and Y are
stochastically equivalent.
Whenever a stochastic process is dened in terms of its values at each individual time t, or
in terms of its joint distributions at nite times, then replacing it by any other version will still
satisfy the denition. It is therefore important to choose a good version. Right-continuous (or
left-continuous) versions are often used when possible, as this then denes the process up to
evanescence.
3
Lemma 3. Let X and Y be right-continuous processes (resp. left-continuous processes) such
that X
t
= Y
t
almost surely, at each time t 0. Then, they are indistinguishable.
Proof. We prove the lemma only for right-continuous processes, the proof for left continuous
processes is almost identical. Let A
S
:= : X
t
() ,= Y
t
(), for some t o. We want to
prove that P(A
R
0
) = 0. By countable additivity of measures we have that
P(A
Q
0
) =

tQ
0
P(( : X
t
() ,= Y
t
()) = 0.
Fix any and suppose there exists a t 0 such that X
t
() ,= Y
t
(). Then [X
t
() Y
t
()[ =
for some > 0. By right-continuity of both processes we can nd, for almost all , a s Q
0
such that [X
t
() X
s
()[ /3 and [Y
t
() Y
s
()[ /3. Then
= [X
t
()Y
t
()[ [X
s
()Y
s
()[+[X
t
()X
s
()[+[Y
t
()Y
s
()[ [X
s
()Y
s
()[+2/3.
Thus, [X
s
() Y
s
()[ /3 which implies that A
Q
0
. In other words
A
R
0
A
Q
0
N
where N is some null event. So we have that P(A
R
0
) = 0.
Viewing a stochastic process in the third sense mentioned above, as a function on the product
space R
+
, it is often necessary to impose a measurability condition. The process X is said to
be jointly measurable if it is measurable with respect to the product sigma-algebra B
[0,)
T.
2.2 Filtrations and adapted processes
In the previous subsection I started by introducing the concept of a stochastic process and
their modications. It is necessary to introduce a further concept, to represent the information
available at each time. A ltration T
t

t0
on a probability space (, T, P) is a collection of
sub-sigma-algebras of T satisfying T
s
T
t
whenever s t. The idea is that T
t
represents the
set of events observable by time t. The probability space taken together with the ltration
(, T, T
t

t0
, P) is called a ltered probability space.
Given a ltration, its right and left limits at any time and the limit at innity are as follows
T
t+
:=

s>t
T
s
, T
t
=
_
_
s<t
T
s
_
, T

=
_
_
tR
0
T
t
_
.
The left limit as dened here only really makes sense at positive times. Throughout these
notes, I dene the left limit at time zero as T
0
:= T
0
. The ltration is said to be right-continuous
if T
t
= T
t+
.
A probability space (, T, P) is complete if T contains all subsets of zero probability elements
of T. Any probability space can be extended to a complete probability space (its completion) in
a unique way by enlarging the sigma-algebra to consist of all sets A such that B A C for
B, C T satisfying P(C B) = 0. Similarly, a ltered probability space is said to be complete
if the underlying probability space is complete and T
0
contains all zero probability sets.
Often, in stochastic process theory, ltered probability spaces are assumed to satisfy the
usual conditions, meaning that it is complete and the ltration is right-continuous. Note that
any ltered probability space can be completed simply by completing the underlying probability
space and then adding all zero probability sets to each T
t
. Furthermore, replacing T
t
by T
t+
,
any ltration can be enlarged to a right-continuous one. By these constructions, any ltered
probability space can be enlarged in a minimal way to one satisfying the usual conditions.
Throughout these notes I assume a complete ltered probability space. However,
for the sake of a bit more generality, I dont assume that ltrations are right-continuous.
4
Remark 2 (by George Lowther). Many of the results can be extended to the non-complete case
without much diculty.
One reason for using ltrations is to dene adapted processes. A stochastic process process
X
t

t0
is said to be adapted if X
t
is an T
t
-measurable random variable for each time t 0.
This is just saying that the value X
t
is observable by time t. Conversely, the ltration generated
by any process X is the smallest ltration with respect to which it is adapted. This is given by
T
X
t
= (X
s
: s t), and referred to as the natural ltration of X.
As mentioned in the previous post, it is often necessary to impose measurability constraints
on a process X
t
considered as a map R
+
R. Right-continuous and left-continuous processes
are automatically jointly measurability. When considering more general processes, it is useful
to combine the measurability concept with adaptedness.
Denition 4. A process X is progressively measurable, or just progressive, if for each t 0,
the map
[0, t] R,
(s, ) X
s
()
is B([0, t]) T
t
-measurable.
In verifying that certain processes are progressive, we will nd the following lemma of use.
Proposition 3. If a process X
t

t0
on (, T, T
t
, P) is right-continuous and T
t
-adapted, then
it is progressively measurable.
Proof. Fix t 0 and consider the sequence of processes dened by
X
n
(s, ) := 1
{0}
(s)X(0, ) +
n

k=1
1

(k1)t
n
,
kt
n

(s)X
_
kt
n
,
_
.
Since X
t
is T
t
adapted, the above processes are measurable functions from ([0, t] , B
[0,t]
)
to (E, B
E
) Note that if s = 0, t, then X
n
(s, ) = X(s, ) and if s (0, t) then
d(X
n
(s, ), X(s, )) = d
_
X
_
tsn/t| +t
n
,
_
, X(s, )
_
.
Since X
t
is right-continuous, the above implies that for any (s, ) in [0, t], X
n
(s, ) X(s, )
as n . That is X
s
, when viewed as a mapping from ([0, t] , B
[0,t]
) to (E, B
E
), is the
pointwise limit of measurable functions, and so, by Proposition 1, it is measurable. Since t was
arbitrary we have that X
t
is progressively measurable.
2.3 Stopping times and their associated sigma algebras
Denition 5. A stopping time for the ltration T
t
is a random variable : [0, +] that
satises for every t 0
: () t T
t
.
This denition is equivalent to stating that the process X
t
() := 1
[0,()]
(t) is adapted.
Equivalently, at any time t, the event t that the stopping time has already occurred is
observable.
One common way in which stopping times appear is as the rst time at which an adapted
stochastic process hits some value. The Debut Theorem states that this does indeed give a
stopping time.
5
Theorem 3 (Debut Theorem, a proof can be found here). Let X be an adapted right-continuous
stochastic process dened on a complete ltered probability space. If K is any real number then
: R
+
dened by
() =
_
inf t R
+
: X
t
() K if t R
+
: X
t
() K , =
+ otherwise
is a stopping time.
The class of stopping times is closed under basic operations such as taking the maximum
or minimum of two times or, for right-continuous ltrations, taking the limit of a sequence of
times. We prove the rst statement, the proof for the second can be found here.
Proposition 4. If , are stopping times, then so are and .
Proof. Proposition 1 implies that all of the functions formed by the operations in the premise
are measurable. Next,
: ( )() t = : () t : () t
: ( )() t = : () t : () t
.
Since, both sets on the RHS belong to T
t
and T
t
is closed under nite intersections and inter-
sections, the LHS belongs to T
t
and so both and and are stopping times.
We say that a stopping time is bounded if there exists a constant C such that () C for
almost all . We say that a stopping time takes nitely many values if for all , () belongs to
some subset of [0, +] of nite cardinality. Since stopping times that take nitely many values
are particularly easy to analyse, it is handy to have a result that allows us to approximate more
general stopping times by nitely many valued ones.
Proposition 5. Suppose that f : [0, +] [0, +] is function such that f(t) t, f(R) =
f
1
(R) = R and its restriction to R is measurable mapping from (R, B
r
) to itself. Also, suppose
that is a stopping time. Then f is a stopping time. If is bounded, then there exists a
sequence
k
of stopping times that tends pointwise (that is, if () is nite,
k
() is as well
and
k
() (), otherwise
k
() = +) to f as k tends to innity. Furthermore, such that
for each xed k,
k
,
k
takes only nitely many values and any constant that bounds above
also bounds above
k
.
Proof. For the rst part we show that the -preimage of any Borel subset of [0, t] is contained
in T
t
. Then we show that the f-preimage of [0, t] is a Borel subset of [0, t]. Putting both results
together we have that the f -preimage of [0, t] is contained in T
t
, that is that f is a stopping
time.
Let (
t
be the sub-collection of Borel subsets B of [0, t] such that
1
(B) is in T
t
. The
stopping property of and the fact that T
s
T
t
for all s t implies that [0, s] is in (
t
for every
s t.
Since for any sets A and A
1
, A
2
, . . . ,
1
(A
c
) = (
1
(A))
c
and
1
(
i
A
i
) =
i

1
(A
i
) and
since both B
[0,t]
and T
t
are sigma-algebras, (
t
is a sigma-algebra as well. Since the Borel sets
are generated by intervals we have that
B
[0,t]
= [0, s] : 0 s t (
t
= (
t
.
Since by denition (
t
B
[0,t]
, the above gives us that (
t
= B
[0,t]
. Now, by our assumptions
on f, f
1
([0, t]) is a Borel subset of R. Furthermore, since f(t) t, f
1
([0, t]) is contained in
[0, t]. Thus f
1
([0, t]) is contained in B
[0,t]
or, equivalently, in (
t
. Hence we have the desired
(f )
1
([0, t]) T
t
, that is, that f is a stopping time.
6
Now for the last bit introduce
k
= f
k
() :=
k+1
k
(if () = set
k
() = +). Clearly,
f
k
(t) t (thus,
k
) and, for all ,
k
()
1
k
,
2
k
, . . . , f
k
(C), + (that is,
k
takes nitely
many values). Pick any xed , if () = +, then
k
() = () for all k. Otherwise, the
sequence (
k
()) is decreasing and bounded from below by (). Thus it converges to a limit
and an easy contradiction gives that the limit must be (). In other words, (
k
) converges
pointwise to .
Clearly for any k f
k
(R) = f
1
k
(R) = R. So to conclude that
k
is a stopping time for any
k 0 all that remains to be shown is that f
k
[
R
: (R, B
R
) (R, B
R
) is measurable. We have
that for every k 0
kx| = lim
n
n

i=1
i1
[i/k,(i+1)/k)
(x).
So by Proposition 1 we have that kx| is the pointwise limit of measurable functions (indicator
functions are measurable if and only if the set they measure is measurable). Hence applying
same proposition again we have that f
k
is measurable.
An T
t
-adapted stochastic process X can be sampled at stopping times. However, X

() :=
X((), ) is merely a random variable and not a stochastic process. It is natural to extend the
notion of adapted processes to random times and ask the following. What is the sigma-algebra of
observable events at the random time , and is X

measurable with respect to this? The idea is


that if a set A is observable at time then for any time t, its restriction to the set t should
be in T
t
. As always, we work with respect to a ltered probability space (, T, T
t

t0
, P). The
sigma-algebra at the stopping time is then,
T

= A T

: A t T
t
for all t 0
The restriction to sets in T

is to take account of the possibility that the stopping time can be


innite, and it ensures that A = A T

. From this denition, a random variable U


us T

-measurable if and only if 1


{t}
U is T
t
-measurable for all times t R
+
.
With these denitions, the question of whether or not a process X is T

-measurable at a
stopping time can be answered. There is one minor issue here though; stopping times can
be innite whereas stochastic processes in these notes are dened on the time index set R
+
.
We could just restrict to the set < , but it is handy to allow the processes to take values
at innity. So, for the moment we consider a processes X
t
where the time index t runs over

R
+
:= R
+
, and say that X is a predictable, optional or progressive process if it satises
the respective property restricted to times in R
+
and X

is T

-measurable.
Lemma 4. If X is a progressively measurable stochastic process and is a stopping time, then
X

is T

-measurable.
Proof. First we show that 1
<
X

is T

-measurable. From the denition we require that


1
<
1
t
X

is T
t
-measurable for all t [0, +]. This is clearly the case if t = + (since its
the zero function). Now suppose that t < + and pick any B B
R
. We need to show that
D := : X((), ) B : () t T
t
.
Since X is progressively measurable
A := (s, ) [0, t] : X(s, ) B B
[0,t]
T
t
.
Suppose that A is of the form A = A
1
A
2
where A
1
B
[0,t]
, A
2
T
t
. Then
C
A
:= : ((), ) A = : () A
1
A
2
.
Since is a stopping time, : () s T
t
for all s [0, t]. Since the intervales [0, s],
0 s t, generate B
[0,t]
it follows that C is the intersection of two sets in T
t
and thus C
A
is
7
in T
t
as well. Furthermore since the sets of the form A = A
1
A
2
where A
1
B
[0,t]
, A
2
T
t
generate B
[0,t]
T
t
, it is straightforward to extend the above argument to show that C
A
T
t
for any A B
[0,t]
T
t
(one needs to use the fact that the generate sigma-algebra operation
commutes with functions, see here or here). Comparing the denitions of D, A and C
A
it follows
that D = C
A
for some A B
[0,t]
T
t
. Since the t was arbitrary we have that 1
<+
X

is T

measurable.
It is easy to see that 1
=

is T

-measurable: 1
=+
1
t
X

is the zero function for all


t < , hence 1
=+
1
t
X

T
t
-measurable). If t = , then we have that 1
=+
1
t
X

= X

is T

-measurable since, by denition X

is T

-measurable.
2.4 Stopped processes
As well as simply observing the value at this time, as the name suggests, stopping times are
often used to stop the process. A process X stopped at the random time is denoted by X

,
X

(t, ) := X
(
t (), ).
It is important that stopping an adapted process at a stopping time preserves the basic
measurability properties. To show this, we rst need the following technical lemma.
Lemma 5. If X is jointly measurable and : R
+
is a measurable map (such a random
variable is called a random time) then X

is measurable.
Proof. Let 1 denote the set of all progressively measurable processes X : [0, ) such
that X

: R is measurable. It follows from Proposition 1 that 1 is closed under the


addition and multiplication operations and under pointwise limits. Pick any B B
[0,)
and
F T. Since is measurable,
1
((,)BF
= 1

1
(B)F
is measurable as well. Thus, 1
(t,)BF
1. Applying the Functional Monotone Class Theorem
(Theorem 1) we have that 1 contains all bounded progressively measurable functions. For any
progressive processes X, it follows from Proposition 1 that Y
n
(t, ) := nX(t, ) is a sequence
of bounded progressive processes that tends pointwise to X as n . Since 1 is closed under
pointwise limits, X 1 completing the claim.
Lemma 6. Let be a stopping time. If the stochastic process X satises any of the following
properties then so does the stopped process X

.
Left-continuous and adapted.
Right-continuous and adapted.
Predictable.
Optional.
Progressively measurable.
Proof. Lemma 5 states that X is jointly measurable and is any random time then X

is
measurable (see here). It follows from the decomposition
X

t
= 1
{t}
X
t
+ 1
{t>}
X

,
that X

is also jointly measurable. Now suppose that X is progressive and T 0 is any xed
time. By denition, X
T
is B(R
+
) T
T
-measurable and, if is a stopping time, then T is
T
T
-measurable. Then, by what we have just shown above, the stopped process
8
(X

)
T
= X
T
= (X
T
)
T
is B(R
+
) T
T
-measurable. This shows that X

is progressive.
9
3 Continuous time martingales
All processes discussed in this section, with the exception of Martingales, take values in (E, B
E
),
where E is a complete separable metric space and B
E
is the collection of Borel subsets of E. We
reserve the symbol d : E E [0, ) for the metric on E. The martingales discussed here all
take values in (R, B
R
) (since we need a total order on the space for the denition of a martingale
to make sense).
3.1 Martingales and stopping times
Denition 6. Let (, T) be a measurable space. A family T
t
T : t 0 of sub-sigma-
algebras of T is called a ltration if T
s
T for any s t.
With a stochastic processes X
t

t0
we associate its natural ltration T
X
t
:= X
s
: s
t. Note we sometimes write X(t, ) instead of X
t
().
Denition 7. Given a probability space (, T, P) and a ltration T
t
T indexed by t 0, a
stochastic process M
t

t0
is called a martingale relative to (, T
t
, P) if the following hold
For almost all , M
t
() has left and right limits at every t and is continuous from the
right, that is, M(t + 0, ) := lim
h0
M(t +h, ) = M(t, ).
For every xed t 0, M(t) is T
t
measurable and integrable.
For 0 s t
E[M
t
[T
s
] = M(s) a.e.
If instead we have that E[M
t
[T
s
] M(s) (or E[M
t
[T
s
] M(s)) we say that M
t
is a super-
matingale (respectively submartingale)
2
.
Remark 3. People often do not require the rst bullet point in the denition and call mar-
tingales that satisfy it cadlag martingales (that is, continue a droite, limite a gauche). Our
assumption that a martingale is cadlag from the get go is justied by the fact that, under some
mild conditions, every martingale has a cadlag modication see Theorem 3 here.
Proposition 6. Suppose that : R R is a convex function, M
t
is a martingale and (M
t
) is
integrable. Then, (M
t
) is a submartingale. In particular, [X(t)[ is a submartingale.
Proof. Because is convex it is also continuous, thus (M
t
) is also cadlag. For any 0 s t,
by Jensens inequality we have that E[(M
t
)[T
s
] (E[M
t
[T
s
]) = (M
s
).
3.2 Doobs Optional Stopping Theorem
Theorem 4 (Optional Stopping Theorem). If
1
,
2
are two bounded stopping times such that

1

2
and M
t
is a martingale, then
E[M

2
[T

1
] = M(
1
)
holds almost surely.
To prove the above, we need the following lemma.
Lemma 7. Let X be an integrable random variable on a probability space (, T, P) and X

:=
E[X[] where T is a sigma-algebra. The collection X

as varies over all sub-sigma-


algebras of T is uniformly integrable.
2
Easy way to remember which one is which: just remember they are named the opposite way any reasonable
human being would have named them
10
Proof. Applying Jensens inequality in the rightmost inequality and the tower property of con-
ditional expectation in the last equality we have that
P[[X

[ l] = E[1
{|X

|l}
] =
E[l1
{|X

|l}
]
l

E[[X

[1
{|X

|l}
]
l

E[[X

[]
l

E[E[[X[[]]
l
=
E[[X[]
l
.
(1)
By the denition of X

, : [X

()[ l . So
_
|X

|l
[X

[dP
_
|X

|l
E[[X[[]dP =
_
|X

|l
[X[dP,
where weve used Jensens inequality to obtain inequality and the denition of E[[X[[] to
obtain the equality. It follows from (1) that ([X[1
|X

|n
)
n
converges pointwise to a function with
support of measure 0 (that is, [X[1

n
{|X

|n}
). So by the Dominated Convergence Theorem we
have that
lim
n
_
[X[1
|X

|n
dP =
_
[X[1

n
{|X

|n}
dP = 0.
The above implies that for any we can nd an N [0, ) such that
_
|X

|n
E[[X[[]dP
for all n N and all sub-sigma-algebras of T.
Proof of Theorem 4. We do this in three steps. First we prove that if C is a bounded stop-
ping time that takes nitely many values, then M

is T

-measurable (and thus T-measurable),


integrable and that
E[M
C
[T

] = M

. (2)
holds. Next we show repeat the above for the case that C is just a bounded stopping time.
Lastly, we show that if
1
and
2
are as in the premise, equation displayed in the conclusion
holds.
Step 1: Suppose that is a stopping time bounded by C and that () is in some nite
set t
1
, t
2
, . . . , t
l
for all . For all i = 1, 2 . . . , l let E
i
:= :
k
() = t
i
, and note that
E
1
E
2
. . . E
l
. First note that
M

= 1
E
1
M
t
1
+ 1
E
2
M
t
2
+ + 1
E
l
M
t
l
.
So by Proposition 1 M

is T
t
l
-measurable (hence, T-measurable) and by the triangle rule in
L
1
(P) it is integrable. Next, pick any Borel set B. We already know that : X((), ) B
lies in T, but we need to verify that it also lies in T

. That is, that for all i


: M((), ) B E
i
T
t
i
.
But
: M((), ) B : () t
i
=
l
_
j=1
: () = t
j
, M
t
j
() B E
i
=
n
_
j=1
(E
j
E
j1
) M
1
t
j
(B) E
i
=
i
_
j=1
(E
j
E
j1
) M
1
t
j
(B).
Since M
t
is adapted to T
t
, is an T
t
-stopping time and T
t
is closed under nite intersections
and unions, the above lies in T
t
i
. In other words, M

is T

measurable. Next pick any A T

.
Since is a stopping time and it is bounded by C we have that A E
i
T
t
i
and
_
AE
i
M

dP =
_
AE
i
M
t
i
dP =
_
AE
i
E[M
C
[T
t
i
]dP =
_
1
AE
i
E[M
C
[T
t
i
]dP
=
_
E[1
AE
i
M
C
[T
t
i
]dP =
_
1
AE
i
M
C
dP =
_
AE
i
M
C
dP.
11
Summing of i then gives us that
_
A
M

dP =
_
A
M
C
dP.
Since A T

was arbitrary, the above and uniqueness of conditional expectation implies that
E[M
C
[T

] = M

(almost surely, of course).


Step 2: Suppose that C is a bounded stopping time.
-
To prove that M

is T-measurable, integrable and that (3) holds it is sucient to prove it


for some sequence of stopping times (
k
) such that
k
C almost surely and
k
(that is,
converges pointwise from above to ). Since M

is cadlag, for almost all ,


k
() () implies
that M(
k
(), ) M((), )
Then, using Proposition ?? again, we have that if A T

it is also true that A T

k
and
thus
_
A
M
C
dP =
_
A
E[M
C
[T

k
]dP =
_
A
M

k
dP.
That is, M

k
tends pointwise to M

. Furthermore, by Jensens inequality we have that for any


k E[[M

k
[] = E[[E[M
C
[T

][] E[E[[M
C
[[T

]] = E[[M
C
[] < (the last inequality follows from
the denition of a martingale). Hence applying Vitalis Convergence Theorem (Theorem 2) we
have that M

is integrable and that


_
A
M

dP = lim
k
_
A
M

k
dP = lim
k
_
A
M
C
dP =
_
A
M
C
dP.
Since the above holds for any A T

, uniqueness of conditional expectation implies establishes


(3).
Suppose that we prove that for any stopping time bounded by C, M

is T-measurable,
integrable and that
E[M
C
[T

] = M

. (3)
Then we would have that
E[M

2
[T

1
] = E[E[M
C
[T

2
][T

1
] = E[M
C
[T

1
] = M

1
where the second equality follows from the tower property and Proposition ??. To prove that
M

is T-measurable, integrable and that (3) holds it is sucient to prove it for some sequence
of stopping times (
k
) such that
k
C almost surely and
k
(that is, converges pointwise
from above to ). Since M

is cadlag, for almost all ,


k
() () implies that M(
k
(), )
M((), )
Then, using Proposition ?? again, we have that if A T

it is also true that A T

k
and
thus
_
A
M
C
dP =
_
A
E[M
C
[T

k
]dP =
_
A
M

k
dP.
That is, M

k
tends pointwise to M

. Furthermore, by Jensens inequality we have that for any


k E[[M

k
[] = E[[E[M
C
[T

][] E[E[[M
C
[[T

]] = E[[M
C
[] < (the last inequality follows from
the denition of a martingale). Hence applying Vitalis Convergence Theorem (Theorem 2) we
have that M

is integrable and that


_
A
M

dP = lim
k
_
A
M

k
dP = lim
k
_
A
M
C
dP =
_
A
M
C
dP.
Since the above holds for any A T

, uniqueness of conditional expectation implies establishes


(3).
12
Proposition 5 gives us an approximating sequence (
k
) as required above, such that for each
xed k,
k
() is in some nite set t
1
, t
2
, . . . , t
l
for almost all . Fix k, pick any A T

k
and
let E
i
:= :
k
() = t
i
. Since
k
is a stopping time we have that A E
i
T
t
i
and
_
AE
i
M

k
dP =
_
AE
i
M
t
i
dP =
_
AE
i
E[M
C
[T
t
i
]dP =
_
1
AE
i
E[M
C
[T
t
i
]dP
=
_
E[1
AE
i
M
C
[T
t
i
]dP =
_
1
AE
i
M
C
dP =
_
AE
i
M
C
dP.
Summing of i then gives us that
_
A
M

k
dP =
_
A
M
C
dP.
Since A T

k
was arbitrary, the above implies that E[M
C
[T

k
] = M

k
(almost surely, of
course).
Corollary 2. If is any stopping time and M
t
is a martingale with respect to (, T
t
, P), then
so is M
t
.
Proof. Proposition ?? establishes that for every xed t, t is a stopping time. Next Lemma
3 in here establishes that M
t
is cadlag and that for every xed t it is T
t
measurable.
References
[1] R Durrett. Probability: theory and examples. Cambridge university press, 3rd edition, 2010.
[2] AN Shiryaev. Probability. Graduate texts in mathematics. Springer, 2nd edition, 1996.
[3] Ramon van Handel. Stochastic Calculus, Filtering, and Stochastic Control. 2007.
13

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy