1 Preliminaries: 1.1 Dynkin's π-λ Theorem
1 Preliminaries: 1.1 Dynkin's π-λ Theorem
_
|X
|c
[X
n
[dP 0.
Theorem 2 (Vitalis Convergence Theorem, Theorem 4 on page 188 of [2]). Let (X
n
) be a
sequence of R-valued integrable random variables. Suppose that (X
n
) is uniformly integrable and
that X
n
X almost surely. Then X is integrable and E[[X
n
X[] 0 as n (which, by
Jensens inequality, implies that E[X
n
] E[X] as n ).
1
1.3 Conditional expectation
Denition 3. Let (, T, ) be a probability space and let X be an R
n
-valued integrable (that
is, E[[X
i
[] < for all i) random variable and let ( be any sigma-algebra. Then E[X[(] is the
unique
1
(-measurable random variable that satises the relation E[1
A
X] = E[1
A
E[X[(]] for all
events A (.
Conditional expectations have the following handy properties:
Proposition 2 (Theorem 2.3.2. in [3]). Let X, Y be real valued integrable functions on (, T, ),
let (, 1 T be sigma-algebras, and let , R. Then
1. Linearity: E[X +Y [(] = E[X[(] +E[Y [(].
2. If X 0 almost surely, then E[X[(] 0 almost surely.
3. Tower property: E[E[X(][1] = E[X[1] almost surely. Note that this implies that
E[X] = E[X[, ] = E[E[X[(][, ] = E[E[X[(]].
4. If X is (-measurable and XY is integrable, then E[XY [(] = XE[Y [(] almost surely. Thus
E[X[(] = XE[1
t0
dened
on a probability space (, T, P). That is, for each time t 0, X
t
() is a measurable
function from to the real numbers.
Remark 1 (by Georgy Lowther). Stochastic processes may also take values in any measurable
space (E, c) but, in these notes, I concentrate on real valued processes. I am also restricting
to the case where the time index t runs through the non-negative real numbers R
+
, although
everything can easily be generalized to other subsets of the reals.
A stochastic process X := X
t
t0
can be viewed in either of the following three ways.
As a collection of random variables, one for each time t 0. As a collection of paths
R
+
R,
t X
t
(),
one for each . These are referred to as the sample paths of the process.
As a function from the product space
R
+
R,
(t, ) X
t
().
As is often the case in probability theory, we are not interested in events which occur with zero
probability. The theory of stochastic processes is no dierent, and two processes X and Y are
said to be indistinguishable if there is an event A of probability one such that X
t
() = Y
t
()
for all A and all t 0. This is the same as saying that they almost surely (that is, with
probability one) have the same sample paths. Alternative language which is often used is that
X and Y are equivalent up to evanescence. In general, when discussing any properties of a
stochastic process, it is common to only care about what holds up to evanescence. For example,
if a processes has continuous sample paths with probability one, then it is referred to as a
continuous (resp., right-continuous and left-continuous) process and we dont care if it actually
has discontinuous (resp. non-right-continuous and non-left-continuous) paths on some event of
zero probability.
It is important to realize that even if we have two processes X and Y such that for each
xed t, the X
t
= Y
t
occurs with probability one, it is not necessarily the case that they are
indistinguishable. As an example, consider a random variable T uniformly distributed over the
interval [0, 1], and dene the process X
t
= 1
{t=T}
. For each time t, P(X
t
,= 0) = P(T = t) = 0.
However, X is not indistinguishable from the zero process, as the sample path t X
t
() always
has one point at which X takes the value 1. The problem here is the uncountability of non-
negative real numbers used for the time index. By countable additivity of measures, if X
t
= Y
t
almost surely, for each t, then we can infer that the sample paths of X and Y are almost surely
identical on any given countable set of times S R
+
, but cannot extend this to uncountable
sets.
This necessitates a further denition. A process Y is a version or modication of X if, for
each time t 0, P(X
t
= Y
t
) = 1. Alternative language sometimes used is that X and Y are
stochastically equivalent.
Whenever a stochastic process is dened in terms of its values at each individual time t, or
in terms of its joint distributions at nite times, then replacing it by any other version will still
satisfy the denition. It is therefore important to choose a good version. Right-continuous (or
left-continuous) versions are often used when possible, as this then denes the process up to
evanescence.
3
Lemma 3. Let X and Y be right-continuous processes (resp. left-continuous processes) such
that X
t
= Y
t
almost surely, at each time t 0. Then, they are indistinguishable.
Proof. We prove the lemma only for right-continuous processes, the proof for left continuous
processes is almost identical. Let A
S
:= : X
t
() ,= Y
t
(), for some t o. We want to
prove that P(A
R
0
) = 0. By countable additivity of measures we have that
P(A
Q
0
) =
tQ
0
P(( : X
t
() ,= Y
t
()) = 0.
Fix any and suppose there exists a t 0 such that X
t
() ,= Y
t
(). Then [X
t
() Y
t
()[ =
for some > 0. By right-continuity of both processes we can nd, for almost all , a s Q
0
such that [X
t
() X
s
()[ /3 and [Y
t
() Y
s
()[ /3. Then
= [X
t
()Y
t
()[ [X
s
()Y
s
()[+[X
t
()X
s
()[+[Y
t
()Y
s
()[ [X
s
()Y
s
()[+2/3.
Thus, [X
s
() Y
s
()[ /3 which implies that A
Q
0
. In other words
A
R
0
A
Q
0
N
where N is some null event. So we have that P(A
R
0
) = 0.
Viewing a stochastic process in the third sense mentioned above, as a function on the product
space R
+
, it is often necessary to impose a measurability condition. The process X is said to
be jointly measurable if it is measurable with respect to the product sigma-algebra B
[0,)
T.
2.2 Filtrations and adapted processes
In the previous subsection I started by introducing the concept of a stochastic process and
their modications. It is necessary to introduce a further concept, to represent the information
available at each time. A ltration T
t
t0
on a probability space (, T, P) is a collection of
sub-sigma-algebras of T satisfying T
s
T
t
whenever s t. The idea is that T
t
represents the
set of events observable by time t. The probability space taken together with the ltration
(, T, T
t
t0
, P) is called a ltered probability space.
Given a ltration, its right and left limits at any time and the limit at innity are as follows
T
t+
:=
s>t
T
s
, T
t
=
_
_
s<t
T
s
_
, T
=
_
_
tR
0
T
t
_
.
The left limit as dened here only really makes sense at positive times. Throughout these
notes, I dene the left limit at time zero as T
0
:= T
0
. The ltration is said to be right-continuous
if T
t
= T
t+
.
A probability space (, T, P) is complete if T contains all subsets of zero probability elements
of T. Any probability space can be extended to a complete probability space (its completion) in
a unique way by enlarging the sigma-algebra to consist of all sets A such that B A C for
B, C T satisfying P(C B) = 0. Similarly, a ltered probability space is said to be complete
if the underlying probability space is complete and T
0
contains all zero probability sets.
Often, in stochastic process theory, ltered probability spaces are assumed to satisfy the
usual conditions, meaning that it is complete and the ltration is right-continuous. Note that
any ltered probability space can be completed simply by completing the underlying probability
space and then adding all zero probability sets to each T
t
. Furthermore, replacing T
t
by T
t+
,
any ltration can be enlarged to a right-continuous one. By these constructions, any ltered
probability space can be enlarged in a minimal way to one satisfying the usual conditions.
Throughout these notes I assume a complete ltered probability space. However,
for the sake of a bit more generality, I dont assume that ltrations are right-continuous.
4
Remark 2 (by George Lowther). Many of the results can be extended to the non-complete case
without much diculty.
One reason for using ltrations is to dene adapted processes. A stochastic process process
X
t
t0
is said to be adapted if X
t
is an T
t
-measurable random variable for each time t 0.
This is just saying that the value X
t
is observable by time t. Conversely, the ltration generated
by any process X is the smallest ltration with respect to which it is adapted. This is given by
T
X
t
= (X
s
: s t), and referred to as the natural ltration of X.
As mentioned in the previous post, it is often necessary to impose measurability constraints
on a process X
t
considered as a map R
+
R. Right-continuous and left-continuous processes
are automatically jointly measurability. When considering more general processes, it is useful
to combine the measurability concept with adaptedness.
Denition 4. A process X is progressively measurable, or just progressive, if for each t 0,
the map
[0, t] R,
(s, ) X
s
()
is B([0, t]) T
t
-measurable.
In verifying that certain processes are progressive, we will nd the following lemma of use.
Proposition 3. If a process X
t
t0
on (, T, T
t
, P) is right-continuous and T
t
-adapted, then
it is progressively measurable.
Proof. Fix t 0 and consider the sequence of processes dened by
X
n
(s, ) := 1
{0}
(s)X(0, ) +
n
k=1
1
(k1)t
n
,
kt
n
(s)X
_
kt
n
,
_
.
Since X
t
is T
t
adapted, the above processes are measurable functions from ([0, t] , B
[0,t]
)
to (E, B
E
) Note that if s = 0, t, then X
n
(s, ) = X(s, ) and if s (0, t) then
d(X
n
(s, ), X(s, )) = d
_
X
_
tsn/t| +t
n
,
_
, X(s, )
_
.
Since X
t
is right-continuous, the above implies that for any (s, ) in [0, t], X
n
(s, ) X(s, )
as n . That is X
s
, when viewed as a mapping from ([0, t] , B
[0,t]
) to (E, B
E
), is the
pointwise limit of measurable functions, and so, by Proposition 1, it is measurable. Since t was
arbitrary we have that X
t
is progressively measurable.
2.3 Stopping times and their associated sigma algebras
Denition 5. A stopping time for the ltration T
t
is a random variable : [0, +] that
satises for every t 0
: () t T
t
.
This denition is equivalent to stating that the process X
t
() := 1
[0,()]
(t) is adapted.
Equivalently, at any time t, the event t that the stopping time has already occurred is
observable.
One common way in which stopping times appear is as the rst time at which an adapted
stochastic process hits some value. The Debut Theorem states that this does indeed give a
stopping time.
5
Theorem 3 (Debut Theorem, a proof can be found here). Let X be an adapted right-continuous
stochastic process dened on a complete ltered probability space. If K is any real number then
: R
+
dened by
() =
_
inf t R
+
: X
t
() K if t R
+
: X
t
() K , =
+ otherwise
is a stopping time.
The class of stopping times is closed under basic operations such as taking the maximum
or minimum of two times or, for right-continuous ltrations, taking the limit of a sequence of
times. We prove the rst statement, the proof for the second can be found here.
Proposition 4. If , are stopping times, then so are and .
Proof. Proposition 1 implies that all of the functions formed by the operations in the premise
are measurable. Next,
: ( )() t = : () t : () t
: ( )() t = : () t : () t
.
Since, both sets on the RHS belong to T
t
and T
t
is closed under nite intersections and inter-
sections, the LHS belongs to T
t
and so both and and are stopping times.
We say that a stopping time is bounded if there exists a constant C such that () C for
almost all . We say that a stopping time takes nitely many values if for all , () belongs to
some subset of [0, +] of nite cardinality. Since stopping times that take nitely many values
are particularly easy to analyse, it is handy to have a result that allows us to approximate more
general stopping times by nitely many valued ones.
Proposition 5. Suppose that f : [0, +] [0, +] is function such that f(t) t, f(R) =
f
1
(R) = R and its restriction to R is measurable mapping from (R, B
r
) to itself. Also, suppose
that is a stopping time. Then f is a stopping time. If is bounded, then there exists a
sequence
k
of stopping times that tends pointwise (that is, if () is nite,
k
() is as well
and
k
() (), otherwise
k
() = +) to f as k tends to innity. Furthermore, such that
for each xed k,
k
,
k
takes only nitely many values and any constant that bounds above
also bounds above
k
.
Proof. For the rst part we show that the -preimage of any Borel subset of [0, t] is contained
in T
t
. Then we show that the f-preimage of [0, t] is a Borel subset of [0, t]. Putting both results
together we have that the f -preimage of [0, t] is contained in T
t
, that is that f is a stopping
time.
Let (
t
be the sub-collection of Borel subsets B of [0, t] such that
1
(B) is in T
t
. The
stopping property of and the fact that T
s
T
t
for all s t implies that [0, s] is in (
t
for every
s t.
Since for any sets A and A
1
, A
2
, . . . ,
1
(A
c
) = (
1
(A))
c
and
1
(
i
A
i
) =
i
1
(A
i
) and
since both B
[0,t]
and T
t
are sigma-algebras, (
t
is a sigma-algebra as well. Since the Borel sets
are generated by intervals we have that
B
[0,t]
= [0, s] : 0 s t (
t
= (
t
.
Since by denition (
t
B
[0,t]
, the above gives us that (
t
= B
[0,t]
. Now, by our assumptions
on f, f
1
([0, t]) is a Borel subset of R. Furthermore, since f(t) t, f
1
([0, t]) is contained in
[0, t]. Thus f
1
([0, t]) is contained in B
[0,t]
or, equivalently, in (
t
. Hence we have the desired
(f )
1
([0, t]) T
t
, that is, that f is a stopping time.
6
Now for the last bit introduce
k
= f
k
() :=
k+1
k
(if () = set
k
() = +). Clearly,
f
k
(t) t (thus,
k
) and, for all ,
k
()
1
k
,
2
k
, . . . , f
k
(C), + (that is,
k
takes nitely
many values). Pick any xed , if () = +, then
k
() = () for all k. Otherwise, the
sequence (
k
()) is decreasing and bounded from below by (). Thus it converges to a limit
and an easy contradiction gives that the limit must be (). In other words, (
k
) converges
pointwise to .
Clearly for any k f
k
(R) = f
1
k
(R) = R. So to conclude that
k
is a stopping time for any
k 0 all that remains to be shown is that f
k
[
R
: (R, B
R
) (R, B
R
) is measurable. We have
that for every k 0
kx| = lim
n
n
i=1
i1
[i/k,(i+1)/k)
(x).
So by Proposition 1 we have that kx| is the pointwise limit of measurable functions (indicator
functions are measurable if and only if the set they measure is measurable). Hence applying
same proposition again we have that f
k
is measurable.
An T
t
-adapted stochastic process X can be sampled at stopping times. However, X
() :=
X((), ) is merely a random variable and not a stochastic process. It is natural to extend the
notion of adapted processes to random times and ask the following. What is the sigma-algebra of
observable events at the random time , and is X
t0
, P). The
sigma-algebra at the stopping time is then,
T
= A T
: A t T
t
for all t 0
The restriction to sets in T
-measurable at a
stopping time can be answered. There is one minor issue here though; stopping times can
be innite whereas stochastic processes in these notes are dened on the time index set R
+
.
We could just restrict to the set < , but it is handy to allow the processes to take values
at innity. So, for the moment we consider a processes X
t
where the time index t runs over
R
+
:= R
+
, and say that X is a predictable, optional or progressive process if it satises
the respective property restricted to times in R
+
and X
is T
-measurable.
Lemma 4. If X is a progressively measurable stochastic process and is a stopping time, then
X
is T
-measurable.
Proof. First we show that 1
<
X
is T
is T
t
-measurable for all t [0, +]. This is clearly the case if t = + (since its
the zero function). Now suppose that t < + and pick any B B
R
. We need to show that
D := : X((), ) B : () t T
t
.
Since X is progressively measurable
A := (s, ) [0, t] : X(s, ) B B
[0,t]
T
t
.
Suppose that A is of the form A = A
1
A
2
where A
1
B
[0,t]
, A
2
T
t
. Then
C
A
:= : ((), ) A = : () A
1
A
2
.
Since is a stopping time, : () s T
t
for all s [0, t]. Since the intervales [0, s],
0 s t, generate B
[0,t]
it follows that C is the intersection of two sets in T
t
and thus C
A
is
7
in T
t
as well. Furthermore since the sets of the form A = A
1
A
2
where A
1
B
[0,t]
, A
2
T
t
generate B
[0,t]
T
t
, it is straightforward to extend the above argument to show that C
A
T
t
for any A B
[0,t]
T
t
(one needs to use the fact that the generate sigma-algebra operation
commutes with functions, see here or here). Comparing the denitions of D, A and C
A
it follows
that D = C
A
for some A B
[0,t]
T
t
. Since the t was arbitrary we have that 1
<+
X
is T
measurable.
It is easy to see that 1
=
is T
-measurable: 1
=+
1
t
X
T
t
-measurable). If t = , then we have that 1
=+
1
t
X
= X
is T
is T
-measurable.
2.4 Stopped processes
As well as simply observing the value at this time, as the name suggests, stopping times are
often used to stop the process. A process X stopped at the random time is denoted by X
,
X
(t, ) := X
(
t (), ).
It is important that stopping an adapted process at a stopping time preserves the basic
measurability properties. To show this, we rst need the following technical lemma.
Lemma 5. If X is jointly measurable and : R
+
is a measurable map (such a random
variable is called a random time) then X
is measurable.
Proof. Let 1 denote the set of all progressively measurable processes X : [0, ) such
that X
1
(B)F
is measurable as well. Thus, 1
(t,)BF
1. Applying the Functional Monotone Class Theorem
(Theorem 1) we have that 1 contains all bounded progressively measurable functions. For any
progressive processes X, it follows from Proposition 1 that Y
n
(t, ) := nX(t, ) is a sequence
of bounded progressive processes that tends pointwise to X as n . Since 1 is closed under
pointwise limits, X 1 completing the claim.
Lemma 6. Let be a stopping time. If the stochastic process X satises any of the following
properties then so does the stopped process X
.
Left-continuous and adapted.
Right-continuous and adapted.
Predictable.
Optional.
Progressively measurable.
Proof. Lemma 5 states that X is jointly measurable and is any random time then X
is
measurable (see here). It follows from the decomposition
X
t
= 1
{t}
X
t
+ 1
{t>}
X
,
that X
is also jointly measurable. Now suppose that X is progressive and T 0 is any xed
time. By denition, X
T
is B(R
+
) T
T
-measurable and, if is a stopping time, then T is
T
T
-measurable. Then, by what we have just shown above, the stopped process
8
(X
)
T
= X
T
= (X
T
)
T
is B(R
+
) T
T
-measurable. This shows that X
is progressive.
9
3 Continuous time martingales
All processes discussed in this section, with the exception of Martingales, take values in (E, B
E
),
where E is a complete separable metric space and B
E
is the collection of Borel subsets of E. We
reserve the symbol d : E E [0, ) for the metric on E. The martingales discussed here all
take values in (R, B
R
) (since we need a total order on the space for the denition of a martingale
to make sense).
3.1 Martingales and stopping times
Denition 6. Let (, T) be a measurable space. A family T
t
T : t 0 of sub-sigma-
algebras of T is called a ltration if T
s
T for any s t.
With a stochastic processes X
t
t0
we associate its natural ltration T
X
t
:= X
s
: s
t. Note we sometimes write X(t, ) instead of X
t
().
Denition 7. Given a probability space (, T, P) and a ltration T
t
T indexed by t 0, a
stochastic process M
t
t0
is called a martingale relative to (, T
t
, P) if the following hold
For almost all , M
t
() has left and right limits at every t and is continuous from the
right, that is, M(t + 0, ) := lim
h0
M(t +h, ) = M(t, ).
For every xed t 0, M(t) is T
t
measurable and integrable.
For 0 s t
E[M
t
[T
s
] = M(s) a.e.
If instead we have that E[M
t
[T
s
] M(s) (or E[M
t
[T
s
] M(s)) we say that M
t
is a super-
matingale (respectively submartingale)
2
.
Remark 3. People often do not require the rst bullet point in the denition and call mar-
tingales that satisfy it cadlag martingales (that is, continue a droite, limite a gauche). Our
assumption that a martingale is cadlag from the get go is justied by the fact that, under some
mild conditions, every martingale has a cadlag modication see Theorem 3 here.
Proposition 6. Suppose that : R R is a convex function, M
t
is a martingale and (M
t
) is
integrable. Then, (M
t
) is a submartingale. In particular, [X(t)[ is a submartingale.
Proof. Because is convex it is also continuous, thus (M
t
) is also cadlag. For any 0 s t,
by Jensens inequality we have that E[(M
t
)[T
s
] (E[M
t
[T
s
]) = (M
s
).
3.2 Doobs Optional Stopping Theorem
Theorem 4 (Optional Stopping Theorem). If
1
,
2
are two bounded stopping times such that
1
2
and M
t
is a martingale, then
E[M
2
[T
1
] = M(
1
)
holds almost surely.
To prove the above, we need the following lemma.
Lemma 7. Let X be an integrable random variable on a probability space (, T, P) and X
:=
E[X[] where T is a sigma-algebra. The collection X
[ l] = E[1
{|X
|l}
] =
E[l1
{|X
|l}
]
l
E[[X
[1
{|X
|l}
]
l
E[[X
[]
l
E[E[[X[[]]
l
=
E[[X[]
l
.
(1)
By the denition of X
, : [X
()[ l . So
_
|X
|l
[X
[dP
_
|X
|l
E[[X[[]dP =
_
|X
|l
[X[dP,
where weve used Jensens inequality to obtain inequality and the denition of E[[X[[] to
obtain the equality. It follows from (1) that ([X[1
|X
|n
)
n
converges pointwise to a function with
support of measure 0 (that is, [X[1
n
{|X
|n}
). So by the Dominated Convergence Theorem we
have that
lim
n
_
[X[1
|X
|n
dP =
_
[X[1
n
{|X
|n}
dP = 0.
The above implies that for any we can nd an N [0, ) such that
_
|X
|n
E[[X[[]dP
for all n N and all sub-sigma-algebras of T.
Proof of Theorem 4. We do this in three steps. First we prove that if C is a bounded stop-
ping time that takes nitely many values, then M
is T
] = M
. (2)
holds. Next we show repeat the above for the case that C is just a bounded stopping time.
Lastly, we show that if
1
and
2
are as in the premise, equation displayed in the conclusion
holds.
Step 1: Suppose that is a stopping time bounded by C and that () is in some nite
set t
1
, t
2
, . . . , t
l
for all . For all i = 1, 2 . . . , l let E
i
:= :
k
() = t
i
, and note that
E
1
E
2
. . . E
l
. First note that
M
= 1
E
1
M
t
1
+ 1
E
2
M
t
2
+ + 1
E
l
M
t
l
.
So by Proposition 1 M
is T
t
l
-measurable (hence, T-measurable) and by the triangle rule in
L
1
(P) it is integrable. Next, pick any Borel set B. We already know that : X((), ) B
lies in T, but we need to verify that it also lies in T
is T
.
Since is a stopping time and it is bounded by C we have that A E
i
T
t
i
and
_
AE
i
M
dP =
_
AE
i
M
t
i
dP =
_
AE
i
E[M
C
[T
t
i
]dP =
_
1
AE
i
E[M
C
[T
t
i
]dP
=
_
E[1
AE
i
M
C
[T
t
i
]dP =
_
1
AE
i
M
C
dP =
_
AE
i
M
C
dP.
11
Summing of i then gives us that
_
A
M
dP =
_
A
M
C
dP.
Since A T
was arbitrary, the above and uniqueness of conditional expectation implies that
E[M
C
[T
] = M
k
and
thus
_
A
M
C
dP =
_
A
E[M
C
[T
k
]dP =
_
A
M
k
dP.
That is, M
k
tends pointwise to M
k
[] = E[[E[M
C
[T
][] E[E[[M
C
[[T
]] = E[[M
C
[] < (the last inequality follows from
the denition of a martingale). Hence applying Vitalis Convergence Theorem (Theorem 2) we
have that M
dP = lim
k
_
A
M
k
dP = lim
k
_
A
M
C
dP =
_
A
M
C
dP.
Since the above holds for any A T
is T-measurable,
integrable and that
E[M
C
[T
] = M
. (3)
Then we would have that
E[M
2
[T
1
] = E[E[M
C
[T
2
][T
1
] = E[M
C
[T
1
] = M
1
where the second equality follows from the tower property and Proposition ??. To prove that
M
is T-measurable, integrable and that (3) holds it is sucient to prove it for some sequence
of stopping times (
k
) such that
k
C almost surely and
k
(that is, converges pointwise
from above to ). Since M
k
and
thus
_
A
M
C
dP =
_
A
E[M
C
[T
k
]dP =
_
A
M
k
dP.
That is, M
k
tends pointwise to M
k
[] = E[[E[M
C
[T
][] E[E[[M
C
[[T
]] = E[[M
C
[] < (the last inequality follows from
the denition of a martingale). Hence applying Vitalis Convergence Theorem (Theorem 2) we
have that M
dP = lim
k
_
A
M
k
dP = lim
k
_
A
M
C
dP =
_
A
M
C
dP.
Since the above holds for any A T
k
and
let E
i
:= :
k
() = t
i
. Since
k
is a stopping time we have that A E
i
T
t
i
and
_
AE
i
M
k
dP =
_
AE
i
M
t
i
dP =
_
AE
i
E[M
C
[T
t
i
]dP =
_
1
AE
i
E[M
C
[T
t
i
]dP
=
_
E[1
AE
i
M
C
[T
t
i
]dP =
_
1
AE
i
M
C
dP =
_
AE
i
M
C
dP.
Summing of i then gives us that
_
A
M
k
dP =
_
A
M
C
dP.
Since A T
k
was arbitrary, the above implies that E[M
C
[T
k
] = M
k
(almost surely, of
course).
Corollary 2. If is any stopping time and M
t
is a martingale with respect to (, T
t
, P), then
so is M
t
.
Proof. Proposition ?? establishes that for every xed t, t is a stopping time. Next Lemma
3 in here establishes that M
t
is cadlag and that for every xed t it is T
t
measurable.
References
[1] R Durrett. Probability: theory and examples. Cambridge university press, 3rd edition, 2010.
[2] AN Shiryaev. Probability. Graduate texts in mathematics. Springer, 2nd edition, 1996.
[3] Ramon van Handel. Stochastic Calculus, Filtering, and Stochastic Control. 2007.
13