MA4F7 Brownian Motion Contents March 16, 2013

advertisement
MA4F7 Brownian Motion
March 16, 2013
Contents
1 Brownian Sample Paths
1.1 Brownian Motion as a Gaussian Process . . . . . . . . . . . . . . . . . . .
1.2 Growth rate of paths . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
1.3 Regularity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
1
2
5
9
2 Brownian motion as a Markov Process
2.1 Markov transition functions . . . . . . . . . . . . . . . . . . . . . . . . . .
2.2 Strong Markov Process. . . . . . . . . . . . . . . . . . . . . . . . . . . . .
2.3 Arcsine Laws for Brownian motion . . . . . . . . . . . . . . . . . . . . . .
12
14
18
21
3 Brownian Martingales
22
4 Donsker’s theorem
32
5 Up Periscope
39
These notes are based on the 2013 MA4F7 Brownian Motion course, taught by Roger
Tribe, typeset by Matthew Egginton.
No guarantee is given that they are accurate or applicable, but hopefully they will assist
your study.
Please report any errors, factual or typographical, to m.egginton@warwick.ac.uk
i
MA4F7 Brownian motion Lecture Notes Autumn 2012
The key aim is to show that scaled random walks converge to a limit called Brownian
motion. In 1D, P{t 7→ Bt nowhere differentiable } = 1
E(Bt ) = 0, E(Bt2 ) = t and so t 7→ Bt is not differentiable at 0. By shifting gives it at
any t.
R1
We also have that P ( 0 χ(Bs > 0)ds ∈ dx) = √ 1
dx
πx(1−x)
(
1 on A
P (x + B exits D in A) = U (x) where ∆U (x) = 0 and U (x) =
. For a
0 on∂D \ A
b−log |x|
disc with inner radius a and outer radius b, U (x) = log
log b−log a and this converges to 1 as
b → ∞. Thus the probability that Brownian motion hits any ball is 1.
For random walks, P (x + r.v. exits at y) = U (x) where U (x) = 41 (U (x + e1 ) + U (x −
e1 ) + U (x + e2 ) = U (x − e2 )) which can be thought of as a discrete Laplacian. Thus we
have a nice equation for Brownian motion, but a not so nice one for random walks.
1
Brownian Sample Paths
Our standard space is a probability space (Ω, F, P).
Definition 1.1 A stochastic process (Bt , t ≥ 0) is called a Brownian Motion on R if
1. t 7→ Bt is continuous for a.s. ω
2. For 0 ≤ t1 < t2 < ... < tn we have Bt2 − Bt1 , ..., Btn − Btn−1 are independent
3. For 0 ≤ s < t we have Bt − Bs is Gaussian with distribution N (0, t − s).
But does this even exist, and if it does, do the above properties characterise B. The
answer to both is yes, and we will show these later.
We now define the terms used in the above definition, to avoid any confusion.
Definition 1.2 A random variable Z is a measurable function Z : Ω → R. In full, R
has the Borel σ-algebra B(R) and measurable means if A ∈ B(R) then Z −1 (A) ∈ F.
Definition 1.3 A stochastic process is a family of random variables (Xt , t ≥ 0) all defined
on Ω.
We do not worry what Ω is, we are only interested in the law/distribution of Z, i.e.
P(Z ∈ A) or E(f (Z)) where P(Z ∈ A) = P{ω : Z(ω) ∈ A}
If we fix ω, the function t 7→ Bt (ω) is called the sample path for ω.
The first property above means that the evaluation of Bt at ω is continuous, for almost
all ω. Sadly some books say that P{ω : t 7→ Bt (ω) is continuous } = 1 but how do we
know this set is measurable.
Definition 1.4 Z a real variable is Gaussian N (µ, σ 2 ) if it has density
P(Z ∈ dz) = √
1
2πσ 2
e−
(x−µ)2
2σ 2
dz
for σ 2 > 0, meaning integrate both sides over a set A to get the probability over A. If
σ = 0 then P(Z = µ) = 1.
1 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
1.0.1
Related Animals
The Brownian Bridge is Xt = Bt − tB1 for t ∈ [0, 1]
An Ornstein-Uhlenbeck process is one, for C > 0, of the form Xt = e−Ct Be2Ct and is
defined for t ∈ R. We will check that X here is stationary. Also (Xt+T : t ≥ 0) is still an
√
O-U process. This arises as the solution of the simplest SDE dX
= −CXt + 2C ddtB̂ , or
dt
Rt
Rt√
in other form, Xt = X0 − C 0 Xs ds + 0 2CdBt .
A Brownian motion on Rd is a process (Bt : t ≥ 0) such that B = (Bt1 , ..., Btd ) where
each t 7→ Btk is a Brownian motion on R and they are independent.
1.1
Brownian Motion as a Gaussian Process
Proposition 1.5 (Facts about Gaussians)
D
1. Z ∼ N (µ, σ 2 ) then for c ≥ 0 we
D
have cZ ∼ N (cµ, c2 σ 2 )
D
D
D
2. Z1 ∼ N (µ1 , σ12 ), Z2 ∼ N (µ2 , σ22 ) and are independent then Z1 + Z2 ∼ N (µ1 +
µ2 , σ12 + σ22 )
D
D
3. Zk ∼ N (µk , σk2 ) and if Zk → Z then limk→∞ µk = µ, limk→∞ σk2 = σ 2 and Z ∼
N (µ, σ 2 ).
The convergence above can be any one of the following.
a.s.
1. Almost surely convergence: Zk → Z means P ({ω : Zk (ω) → Z(ω)}) = 1
prob
2. In probability: Zk → Z means P (|Zk − Z| > ε) → 0 for all ε > 0
k→∞
D
3. In distribution: Zk → Z means E(f (Zk )) → E(f (Z)) for any continuous and
bounded f .
Example 1.1 I =
R1
0
Bt dt is a Gaussian variable.
1
(B
+ B2/N + ... + BN/N )
N 1/N
1
= lim
((BN/N − B(N −1)/N ) + 2(B(N −1)/N − B(N −2)/N ) + ... + N (B1 − B0 ))
k→∞ N
I = lim
k→∞
and all these are independent and so Gaussian.
1.1.1
Transforms
Definition 1.6 We define the Fourier transform, or the characteristic function to
be
φZ (θ) = E(eiθZ )
D
For example, if Z ∼ N (µ, σ 2 ) then φZ (θ) = eiθµ e−σ
2 /2
Proposition 1.7 (More facts about Gaussians)
4. φZ (θ) determines the law of
Z, i.e. if φZ (θ) = φY (θ) then P (Z ∈ A) = P (Y ∈ A).
5. Z1 , Z2 independent if and only if E((eiθ1 Z1 eiθ2 Z2 ) = E(eiθ1 Z1 )E(eiθ2 Z2 ) for all θ1 , θ2 .
2 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
D
6. φZk (θ) → φZ (θ) if and only if Zk → Z.
These all hold true for Z = (Z1 , ..., Zd ) with φZ (θ1 , ..., θd ) = E(eiθ1 Z1 +...+iθd Zd )
P
Definition 1.8 Z = (Z1 , ..., Zd ) ∈ Rd is Gaussian if di=1 λk Zk is Gaussian in R for
all λ1 , ..., λd . (Xt , t ≥ 0) is a Gaussian process if (Xt1 , ..., XtN ) is a Gaussian vector
on RN for any t1 , ..., tN and N ≥ 1.
Check that
PBrownian motion is a Gaussian process, i.e. is (Bt1 , ..., BtN ) a Gaussian
vector, or is
λk Btk Gaussian on R. We can massage this into µ1 (Bt1 − B0 ) + ... +
µN (BtN − BtN −1 ) and so is Gaussian. As an exercise, check this for Brownian bridges and
O-U processes.
Proposition 1.9 (Even more facts about Gaussians)
7. The Law of the Gaussian Z = (Z1 , ..., Zd ) is determined by E(Zk ) and E(Zj , Zk ) for j, k = 1, ..., d
8. Suppose Z = (Z1 , ..., Zd ) is Gaussian. then Z1 , ..., Zd are independent if and only if
E(Zj Zk ) = E(Zj )E(Zk )
for all j 6= k.
For 7, it is enough to calculate φZ (θ) and see that it is determined by them. For 8, noe
need only check that the transforms factor.
Example 1.2 (Bt ) a Brownian motion on R. Then E(Bt ) = 0 and, for 0 ≤ s < t,
E(Bs Bt ) = E((Bt −Bs )(Bs −B0 )+(Bs −B0 )2 ) = E(Bt −Bs )E(Bs −B0 )+E(Bs −B0 )2 = s
and similarly equals t if 0 ≤ t < s.
Do the same for Brownian bridges and O-U processes.
Theorem 1.10 (Gaussian characterisation of Brownian motion) If (Xt , t ≥ 0) is
a Gaussian process with continuous paths and E(Xt ) = 0 and E(Xs Xt ) = s ∧ t then (Xt )
is a Brownian motion on R.
Proof We simply check properties 1,2,3 in the definition of Brownian motion. 1 is
immediate. For 2, we need only check that E((Xtj+1 − Xtj )(Xtk+1 − Xtk )) splits. Suppose
tj ≤ tj+1 ≤ tk ≤ tk+1 and then E((Xtj+1 − Xtj )(Xtk+1 − Xtk )) = tj+1 − tj+1 − tj + tj = 0
as required. For 3, Xt − Xs is a linear combination of Gaussians and so is Gaussian. It
has mean zero and
E(Xt − Xs )2 = E(Xs2 − 2Xs Xt + Xt2 ) = s − 2s ∧ t + t = t − s
Q.E.D.
Suppose I =
R1
R1
Bs ds and E(I) = 0 E(Bs )ds = 0 and also
Z 1
Z 1
Z 1Z 1
Z
E(I 2 ) = E(
Bs ds
Br dr) =
E(Bs Br )dsdr =
0
0
0
0
0
0
1Z 1
s ∧ rdsdr =
0
1
3
R1R1
but we need to check that we can use Fubini, so we need to check that K = E( 0 0 |Br ||Bs |drds) <
∞. Now
Z 1Z 1
Z 1Z 1
√
K=
E(|Br ||Bs |)drds ≤
E(Br2 )E(Bs2 )drds ≤ rs < 1
0
0
0
0
as we wanted.
3 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
Lemma 1.11 (Scaling Lemma) Suppose that B is a Brownian motion on R and c > 0.
Define Xt = 1c Bc2 t for t ≥ 0. Then X is a Brownian motion on R
Proof Clearly it has continuous paths and E(Xt ) = 0. Now
1
1
E(Xs Xt ) = E( Bc2 s Bc2 t ) = s ∧ t
c
c
and also
N
X
λk Xtk =
1
N
X
λk
c
1
Bc2 tk
and this is Gaussian since Bt is Gaussian.
Q.E.D.
Lemma
( 1.12 (Inversion lemma) Suppose that B is a Brownian motion on R. Define
tB 1 t > 0
t
. Then X is a Brownian motion.
Xt =
0
t=0
Proof
N
X
λk Xtk =
1
N
X
λk tk B 1
tk
1
which is still Gaussian for tk > 0. If any of the tk = 0 then the addition to the above sum
of this term is zero, so we are fine. Clearly E(Xt ) = 0 and
1
E(Xs Xt ) = E(stB 1 B 1 ) = ts = s
s
t
t
for s < t. We also have no problem for t > 0 with the continuity of paths. However we
need to check that it is continuous at t = 0, i.e. that tB 1 → 0 as t → 0, or that 1s Bs → 0
t
√
as s → ∞. We expect that Bt ≈ ± t and so Bt /t → 0 should be clear.
However, we know that (Xt1 , ..., XtN ) = (B̂t1 , ..., B̂tN ) providing ti > 0 for a Brownian
motion B̂ and since B̂t → 0 as t → 0 surely Xt → 0 as well. We pin this down precisely:
[Xt → 0 as t → 0] = {Xq → 0, as q → 0, q ∈ Q}
= {∀ε > 0 : ∃δ > 0, q ∈ Q ∩ (0, δ] =⇒ |Xq | < ε}
∞ [
∞
\
\
1
=
{|Xq | < }
N
1
N =1 M =1 q∈Q∩(0,
M
]
and so
P[Xt → 0 as t → 0] = lim
lim lim P{|Xq | <
N →∞ M →∞ k→∞
where q1 , q2 , ... lists Q ∩ (0,
1
} = P[B̂t → 0]
N
1
M ].
Q.E.D.
We used in the above that
A1 ⊇ A2 ⊇ ... then P(∩AN ) = lim P(AN )
N →∞
A1 ⊆ A2 ⊆ ... then P(∪AN ) = lim P(AN )
N →∞
Corollary 1.13 Bt /t → 0 as t → ∞.
Bt
In fact, Bt /tα → 0 for α > 12 but lim supt→∞ √
= ∞ and lim inf t→∞
t
Bt visits every x ∈ R infinitely many times.
This brings us nicely into the next subsection.
4 of 40
Bt
√
t
= −∞ and so
MA4F7 Brownian motion Lecture Notes Autumn 2012
1.2
Growth rate of paths
Theorem 1.14 (Law of the Iterated Logarithm) Suppose that Bt is a Brownian motion on R. Then
Bt
= +1
lim sup
t→∞ ψ(t)
and
lim inf
t→∞
Bt
= −1
ψ(t)
p
where ψ(t) = 2t ln(ln t)
lim supt→∞ Xt = limt→∞ sups≥t Xs and lim sup Xt ≤ 1 means that ∀ε > 0 then sups≥t Xs ≤
1 + ε for large t, which is the same as for all ε > 0 Xt is eventually less than 1 + ε.
lim sup Xt ≥ 1 if and only if ∀ε > 0 we have sups≥t Xs ≥ 1 − ε for large t. which is
the same as for all ε > 0 there exists a sequence sN → ∞ with XsN ≥ 1 − ε.
It is on an example sheet that Xt = e−t Be2t then the Law of the Iterated logarithm
t
can be converted to get that lim sup √X
= 1.
2 ln t
We can also compare this to ZN an iid N (0, 1) and then
Proof We first show that
Bt
≤ 1) = 1
P(lim sup
ψ(t)
lim
√ sup ZN
2 ln N
=1
and this is the case if and only if
P(Bt ≤ (1 + ε)ψ(t) for large t) = 1
We first perform a calculation:
p
P(Bt > (1 + ε)ψ(t) for large t) = P(N (0, t) > (1 + ε) 2t ln(ln t))
p
= P(N (0, 1) > (1 + ε) 2 ln(ln t))
z2
Z∞
e− 2
√ dz
=
2t
√
(1+ε)
2 ln(ln t)
Lemma 1.15 (Gaussian Tails)
Z ∞
1
2
1
2
2
−a2 /2
1− 2 e
≤
e−z /2 dz ≤ e−a /2
a
a
a
a
Then we get that
Z∞
√
(1+ε)
2
z
√
e− 2
1
2
√ dz ≤
p
√ e−2((1+ε) 2 ln(ln t)) /2
2t
(1 + ε) 2 ln(ln t) 2π
2 ln(ln t)
=
1
2
p
√ e−(1+ε) ln(ln t)
(1 + ε) 2 ln(ln t) 2π
The strategy now is to control B along a grid of times tN = θN for θ > 1. Then
1
1
1
2
2
p
P(BθN > (1 + ε)ψ(θN )) ≤ √
e−(1+ε) ln(N ln θ) ≤ C(θ, ε)N −(1+ε)
2π 1 + ε 2 ln(N ln θ)
5 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
Lemma 1.16 (Borel-Cantelli part 1) If
P∞
1
P(AN ) < ∞ then
P( only finitely many AN happen ) = 1
Proof Let χAN
(
1
=
0
P
AN
and then the number of AN s that occur is given by ∞
1 χA N
c
AN
and so
E[
∞
X
χA N ] =
∞
X
1
and so
P∞
1
E[χAN ] =
1
∞
X
P(AN ) < ∞
1
χAN is finite a.s. Q.E.D. Then by BCI we have that BθN ≤ (1 + ε)ψ(θN ) for
all large N . We now need to control B over (θN , θN +1 ).
Lemma 1.17 (Reflection trick)
P(sup Bs ≥ a) = 2P(Bt ≥ a)
s≤t
for a ≥ 0
Proof Define Ω0 = {sups≤t Bs ≥ a} and then
P(Ω0 ) = P(Ω0 ∩ {Bt > a}) + P(Ω0 ∩ {Bt = a}) + P(Ω0 ∩ {Bt < a})
= 2P(Ω0 ∩ {Bt > a})
= 2P{Bt > a}
We will carefully justify this later by examining the hitting time Ta = inf{t : Bt = a}.
We consider (BTa +t − a, t ≥ 0 and check that this is still a Brownian motion.
Z ∞
a
1
Bt
2
√ ez /2 dz
P(Ta ≤ t) = P(sup Bs ≥ a) = 2P( √ ≥ √ ) = 2
t
t
2π
s≤t
a
and also
P(Ta ∈ dt) =
d
1
a 3
1
2
2
P(Ta ≤ t) = 2 √ e−a /2t t− 2 = √
e−a /2t =: φ(t)
3
dt
2
2π
2πt
and so E(Ta ) = ∞.
Q.E.D.
Thus from this we get that
2
P( sup Bs ) ≥ (1 + ε)ψ(θN )) = 2P(BθN > (1 + ε)ψ(θN )) ≤ 2C(ε, θ)N −(1+ε)
s≤θN
Borel-Cantelli part 1 still applies and so for large N we have
sup Bs ≤ (1 + ε)ψ(θN )
s≤θN
This if t ∈ [θN , θN +1 ] we have
√
Bt
(1 + ε)ψ(θN +1 )
≤
= (1 + ε) θ
N
ψ(t)
ψ(θ )
p
√
ln((N + 1) ln θ)
p
→ (1 + ε) θ
ln(N ln θ)
6 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
and thus we have that
lim sup
t→∞
√
Bt
≤ (1 + ε) θ
ψ(t)
for all ε > 0 and θ > 1 and so
lim sup
t→∞
We now show
lim sup
t→∞
If we choose tN =
θN
Bt
≤1
ψ(t)
Bt
≥ (1 − ε)
ψ(t)
for θ > 1 then
2
P(BθN > (1 − ε)ψ(θN )) ≥ C(θ, ε)N −(1−ε)
P
Lemma 1.18 (Borel-Cantelli part 2) If ∞
1 P(AN ) = ∞ and AN s are independent,
then
P( infinitely many AN occur) = 1
P
Proof Z = ∞
1 χAN is the total number of AN s that occur. From BCI we get that
E(Z) < ∞ =⇒ P(Z < ∞) = 1
or E(e−Z ) = 0 ⇐⇒ P(Z = ∞) = 1. Then
E(e
−
P
χAN
) = E(
∞
Y
e−χAN )
1
=
Y
E(e−χAN )
1∞
∞
Y
=
(1 − αP(AN ))
1
≤
∞
Y
e−αP(AN )
1
P
−α P(AN )
=e
=0
Q.E.D.
We use this on AN = {BθN > (1 − ε)ψ(θN )} but these are not independent, but
nearly √
so for large N . We finalise by correcting this. We define ÂN = {BθN − BθN −1 >
(1 − ε) 1 − θ−1 ψ(θN )} and these are independent. Then
P(ÂN ) = P(AN ) ≥ C(θ, ε)N −(1−ε)
2
BC2 tells us that infinitely many ÂN do occur a.s., i.e.
p
p
BθN ≥ (1 − ε) 1 − θ−1 ψ(θN ) − BθN −1 ≥ (1 − ε) 1 − θ−1 ψ(θN ) − (1 + ε)ψ(θN −1 )
and so
p
Bθ N
ψ(θN −1 )
−1 − (1 + ε)
≥
(1
−
ε)
1
−
θ
ψ(θN )
ψ(θN )
7 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
and
p
√
ln((N − 1) ln θ)
ψ(θN −1 )
p
√
=
→
θ−1
ψ(θN )
θ ln(N ln θ)
and so
lim sup
N →∞
p
√
BθN
−1 − (1 + ε) θ −1
≥
(1
−
ε)
1
−
θ
ψ(θN )
and taking θ large and ε small gives the result
Q.E.D.
We make some observations:
(
1
1. Can we do better? P(Bt ≤ ht for large t) =
for ht deterministic. This is
0
√
called the 0-1 law, and we see it in week 4. For ht = Ct ln ln t if C < 0 then we
get 0 and if C > 2 then we get 1. This uses an integral test for ht .
2. Random walk analogue. Suppose that X1 , X2 , ... are iid with E(Xk ) = 0, E(Xk2 ) = 1
and SN = X1 + ... + XN . then
lim sup √
N →∞
SN
=1
2N ln ln N
This was proved in 1913 but the proof was long. Was proved in a shorter manner
using the Brownian motion result in 1941.
3. Xt = tB 1 is still a Brownian motion and so lim supt→∞
t
tB 1
t
ψ(t)
or alternatively
Bs
=1
lim q
s→0
2s ln ln( 1s )
and so we have a result about small t behaviour.
4. P(B diff at 0) = 0 and if we fix T0 > 0 then define Xt = BT0 +t then this is still a
Brownian motion. Thus
Corollary 1.19 P(B diff at t0 ) = 0 for all t0 .
5. Suppose U ∼ U [0, 1] is uniform r.v. Then define Xt by the value up until U and then
monotone increasing up until 1. Then P(X diff at t0 ) = 1 but it is not differentiable
at all t. and so we cannot easily conclude that Brownian motion id differentiable
everywhere
6. Corollary 1.20
Leb{t : B is diff at t} = 0
Proof
Z
∞
Z
χ(B diff at t)dt) =
E(
0
∞
Eχ(B diff at t)dt = 0
0
Q.E.D.
The points where it is differentiable are examples of random exceptional points.
8 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
1.3
Regularity
Definition 1.21 A function f : [0, ∞) → R is α-Holder continuous, for α ∈ (0, 1], at
t if there exists M, δ > 0 such that
|ft+s − ft | ≤ M |s|α
for |s| ≤ δ.
The case of α = 1 is called Lipschitz.
The aim of the next part is to show that P(B is α Holder continuous at all t ≥ 0) = 1
provided α < 1/2 and that P(B is α Holder continuous at any t ≥ 0) = 0 provided α >
1/2.
Corollary 1.22 P(B differentiable at any t) = 0
The reasons for this are as follows. A differentiable function must lie in some cone as
f (t + s) − f (s)
→a
s
and so (f (t + s) − f (s))/s ∈ (a − ε, a + ε) for small s and thus |f (t + s) − f (s)| ≤ (|a| + ε)|s|
for small s, and so Lipschitz holds with M = |a| + |eps.
Proposition 1.23 Define ΩM,δ = {for some t ∈ [0, 1], |Bt+s −Bt | ≤ M |s| for all |s| ≤ δ}
∞
and then P(ΩM,δ ) = 0 and thus P(∪∞
M =1 ∪N =1 ΩM, 1 ) = 0
N
Proof This hasn’t been bettered since 1931. Suppose that there exists a t ∈ [K/N, (K +
1)/N ] where it is Lipschitz, i.e. |Bt+s − Bt | ≤ M |s| for |s| < δ. Then if (K + 1)/N, (K +
2)N ∈ [t, t + δ] then
|B(K+1)N − Bt | ≤
M
N
|B(K+2)N − Bt | ≤
2M
N
and so by the triangle inequality we get
|B(K+1)N − B(K+2)/N | ≤
3M
N
and then we have
P(ΩM,δ ) ≤ P
for some K = 1, ..., N − 1 : |B(K+1)N
3M
− B(K+2)/N | ≤
N
We first calculate the probability of the event on the right hand side.
3M
√
P[|N (0, 1/N )| ≤
3M
3M
] = P[|N (0, 1)| ≤ √ ] =
N
N
ZN
1
6M
2
√ e−z /2 dz ≤ √
2π
2πN
3M
−√
N
and then the last part of equation (1.1) is equal to
=
−1
P(∪N
|B(K+1)N
1
N
−1
X
6M √
3M
6M
− B(K+2)/N | ≤
)≤
(√
)= √
N
N
2π
2πN
1
9 of 40
(1.1)
MA4F7 Brownian motion Lecture Notes Autumn 2012
but this is not useful, because it does not tend to zero. We modify this by taking more
points. We already know that
|B(K+2)N − B(K+1)/N | ≤
3M
N
but we also have that
|B(K+3)N − B(K+2)/N | ≤
5M
N
|B(K+4)N − B(K+3)/N | ≤
and then we say that


P(ΩM,δ ) ≤ P  for some K = 1, ..., N − 1 :  and
and
7M
N
|B(K+1)N − B(K+2)/N | ≤
|B(K+3)N − B(K+2)/N | ≤
|B(K+4)N − B(K+3)/N | ≤
3M 
N
5M 
N
7M
N
with 4/N ≤ δ. and this is equal to
−1
{A and B and C}) ≤
P(∪N
1
N
−1
X
P({A and B and C}) ≤
1
N
−1
X
1
D
D
= √ →0
n3/2
N
and so P(ΩM,δ ) = 0.
Q.E.D.
We now turn to the other statement, that B is α-Holder continuous if α ∈ (0, 1/2).
Theorem 1.24 (Kolmogorov’s continuity) Suppose (Xt , t ∈ [0, 1]) has continuous
paths and satisfies
E[|Xt − Xs |p ] ≤ C|t − s|1+γ
for some γ, p > 0. Then X has α-Holder paths for α ∈ 0, γp ).
We use this for X a Brownian motion. In this case
√
p
E[|Bt − Bs |p ] = E[|N (0, t − s)|p ] = t − s E[|N (0, 1)|p ] = cp |t − s|p/2
We then have that Brownian motion has α-Holder continuous paths of orders
α<
and so Holder with any α <
1
2
p/2 − 1
1 1
= −
p
2 p
but not α =
1
2
Lemma 1.25 (Markov Inequality) Suppose that Z ≥ 0 is a non negative random variable. Then
1
P(Z ≥ a) ≤ E(Z)
a
Proof
E(Z) = E(Zχ{Z<a} ) + E(Zχ{Z≥a} ) ≥ E(Zχ{Z≥a} ) ≥ aP(Z ≥ a)
Q.E.D.
We use this as follows
P(|Xt − Xs | ≥ a) = P(|Xt − Xs |p ≥ ap ) ≤
10 of 40
1
C|s − t|1+γ
p
E(|X
−
X
|
)
≤
t
s
ap
ap
MA4F7 Brownian motion Lecture Notes Autumn 2012
N
Suppose we have a grid with size 2−N .
X(K−1)/2N | > 2N1 α }
We then estimate this
Then we define AN = ∪2K=1 {|Xk/2N −
N
P(AN ) ≤
2
X
P{|Xk/2N − X(K−1)/2N | >
1
1
2N α
} ≤ C2N 2−N (1+γ) 2N αp = 2−N (γ−αp)
and thus we need γ > αp or α < γ/p. This is the key idea of the proof. We now have
P
∞
1 P(AN ) < ∞ and so by Borel Cantelli I we have that only finitely many AN s occur,
i.e. there exists N0 (ω) such that
|XK/2N − X(K−1)/2N | <
1
for all K = 1, ..., 2N and for N > N0
2N α
We also know that P(N0 < ∞) = 1. We now fix ω such that N0 (ω) < ∞. We can control
X on D = { dyadic rationals i.e. of form K/2M }. Fix t, s ∈ D with
1/2M ≤ |t − s| ≤ 1/2M −1
where M ≥ N0 . We have two cases. Either we straddle two points K/2M and (K + 1)/2M
or we straddle only one point K/2M . We consider the first case, and leave the second
case to the reader. We have
t=
K + 1 1 or 0
1 or 0
+ M +1 + ... + M +L
2M
2
2
and then we have
|Xt − X(K+1)/2M | ≤
1
1
1
K +1
1
+ M +1 + ... + M +L ≤ (M +1)α
M
2
2
2
1 − 2−α
2
and similarly
|Xs − X(K)/2M | ≤
1
2(M +1)α
1
1 − 2−α
and thus
|Xs − Xt | ≤ |Xs − X(K)/2M | + |X(K)/2M − X(K+1)/2M | + |Xt − X(K+1)/2M |
1
1
2
≤ (M +1)α
+ Mα
−α
1−2
2
2
C
≤ Mα
2
and thus using the assumptions of Kolmogorov we have this less than C|t − s|α . Invoking
continuity gives the result.
11 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
2
Brownian motion as a Markov Process
Roughly, the future (Xs ; s ≥ t) is independent of the past (Xs : s ≤ t) conditional on the
present Xt .
Definition 2.1 Suppose that F, G two σ-fields on Ω are called independent if P(A ∩
B) = P(A)P(B) for all A ∈ F and B ∈ G.
Two variables X and Y are called independent if
P(X ∈ C1 , Y ∈ C2 ) = P(X ∈ C1 )P(Y ∈ C2 )
or equivalently
E(h(X)g(Y )) = E(h(X))E(g(Y ))
for all measurable bounded functions h, g.
They are equivalent as follows. We can take h = χC1 and g = χC2 and then the two
statements are the same. We then take simple functions and limits of simple functions to
get any functions, using the standard machine.
Definition 2.2 σ(X), called the σ-field generated by X is defined as
{X −1 (C) : C measurable }
Note that now X is independent of Y if and only if σ(X) is independent of σ(Y ).
Definition 2.3 σ(Xt , t ∈ I) called the σ-field generated by Xt is defined to be
σ{Xt−1 (C) : t ∈ I, C measurable }
i.e. the smallest σ field containing the above sets.
Theorem 2.4 (Markov property of Brownian motion 1) Suppose that B is a Brownian motion, and fix t0 > 0. Define Xt = Bt+t0 −Bt0 and then σ(Xt : t ≥ 0) is independent
of σ(Bt : t ≤ t0 ).
Note that this is Rvery close to independent increments.
Rt
Rt
t
For example 0 0 Bs ds is independent of 0 1 Xs ds. Need to check that 0 0 Bs ds is
measurable with respect to σ(Bt ).
Z
t0
N
X
1
B
N →∞
N K/N
Bs ds = lim
0
1
which is measurable since the limit and sum of measurable functions is measurable.
A second example is supt≤s Bt is independent of supt≤t1 Xt . We can write this as
supt≤s Bt = sup Bq = limN →∞ max{Bq1 , ..., BqN } where qi lists [0, 1] ∩ Q. Now
q∈[0,1]∩Q
w 7→ (Bq1 (w), ..., BqN (w)) 7→ max Bqi (w)
and the latter is continuous, and the former is measurable, since if C = C1 × ... × CN
then {w : (Bq1 (w), ..., BqN (w)) ∈ C} is measurable.
Definition 2.5 A collection of subsets A0 is called a π-system if it is closed under finite
intersections.
12 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
Lemma 2.6 If F0 and G0 are two different π-systems and P(A ∩ B) = P(A)P(B) for
all A ∈ F0 and B ∈ G0 then P(A ∩ B) = P(A)P(B) holds for σ(F0 ) and σ(G0 ), i.e.
independence of π-systems gives independence of generated σ-fields.
For us, σ(Xt , t ≥ 0) = σ{Xt−1 (C) : t ≥ 0, C ∈ B(R)} and this is generated by the
π-system {Xt−1
(C1 ) ∩ ... ∩ Xt−1
(CN ) : t1 , ..., tN ≥ 0, CK ∈ B(R)}
1
N
Proof (of theorem 2.4) By the above lemma, we need only check that (Xt1 , ..., XtN ) is
independent of (Bs1 , ..., BsN ). We have
i
i
h P
h P
P
P
E ei λk Btk ei µk (BT +tk −Btk ) = E e− λ̂k (Bsk −Bsk−1 ) ei µ̂k (BT +tk −BT +tk−1 )
h P
i h P
i
= E ei λk Btk E ei µk (BT +tk −Btk )
by independent increments.
Q.E.D.
Proposition 2.7 The following are true
1. a ≤ b ≤ c ≤ d then P(max Bt 6= max Bt ) = 1
[a,b]
[c,d]
2. P(∃!t? ∈ [a, b] : Bt? = max Bt ) = 1
[a,b]
3. Local maxima are dense in [0, ∞).
Proof
We can get 2 from one by using that P(max Bt 6= max Bt : ∀a ≤ b ≤ c ≤
[a,b]
[c,d]
d rational ) = 1 We can get 3 from 2 by doing this for all rational (a, b) simultaneously.
We now prove 1. Take Xt = Bc+t − Bc , and let Y = max[c,d] (Bt − Bc ) = max[0,d−c] Xt
which is independent of σ(Bs , s ≤ c). Take Z = max[a,b] (Bt − Bc ) and we aim to show
that P(Z 6= Y ) = 1. Y and Z are independent by the Markov property but Y has a
density. By the reflection principle, this implies that P(Y = Z) = 0.
Q.E.D.
At the last part of this, we have used that if E(g(Y )h(Z)) = E(g(Y ))E(h(Z)) and if
ν(dy) = P(Y ∈ dy) and µ(dz) = P(Z ∈ dz) then
Z Z
E(φ(Y, Z)) =
φ(y, z)ν(dy)µ(dz)
(
1
This can be shown by the standard machine, with first taking φ(Y, Z) =
0
Y =Z
Y 6= Z
Definition 2.8 We define the Brownian filtration by FtB := σ(Bs : s ≤ t), and we
B =∩
B
B
also define Ft+
s>t Fs and the germ field is F0+ .
B
Note that if dB
dt (t) existed then it would be in Ft+ . However, lim supt→0
since this limsup is FδB measurable for any δ > 0.
Bt
h(t)
B ,
is in F0+
Theorem 2.9 (Markov property version 2) Fix T ∈ R and define Xt = BT +t − BT .
B.
Then σ(Xt , t ≥ 0) is independent of Ft+
B so is independent of itself.
Corollary 2.10 T = 0 so σ(Bt , t ≥ 0) is independent of F0+
B is a 0/1 field.
Corollary 2.11 F0+
13 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
B measurable then {ω : Z(ω) ≤ c} has probability 0 or 1. Thus Z is a
Thus if X is F0+
constant almost surely.
Example 2.1 (Lebesgue’s Thorn) Suppose that B is a d dimensional Brownian moB .
tion and F is an open set in Rd . Take τ = inf{t : Bt ∈ F }. We claim that {τ = 0} ∈ F0+
∞
B
Indeed {τ = 0} = ∩N =N0 ∪q∈Q∩[0,1/N ] {Bq ∈ F } ∈ F1/N0 for all N0 . Lebesgue’s thorn
p
is an example of F , e.g. F = {(x, y, z) : y 2 + z 2 ≤ f (x)} a volume of rotation. Here
P(τ > 0) = 1 for thin thorns, or is 0 for thick thorns.
B and B ∈ σ(X , s ≥ 0). The π-system lemma
Proof ( of theorem 2.9) Take A ∈ Ft+
s
says we can choose B = {Xsi : i = 1, ..., m}, i.e. enough to check that
E[e−
P
λk Xsk iθχA
e
]
splits as a product of the two expectations.
B . Then
Let Xsε = Bt+ε+s − Bt+ε and note that this is independent of Ft+ε
E[e−
P
λk Xsε
k
eiθχA ] = E[e−
P
λk Xsε
k
]E[eiθχA ]
by the Markov property version 1. Letting ε → 0 and then using the DCT we get the
required result.
Q.E.D.
Example 2.2 (Shakespeare problem) Suppose we have 2 dimensional Brownian motion.
Then assume that
P(Bt , t ∈ [0, 1] traverses a tube A) = p0 > 0
and then
1
1
P( √ Bt , t ∈ [0, 1/2] traverses a tube √ A) = p0
2
2
then let
AN = {
1
2N/2
Bt : t ∈ [0, 2−N ] traverses 2−N/2 A}
∞
and then P(AN ) = p0 by scaling. Then note that AN ∈ F2B−N and let Ω0 = ∩∞
M =M0 ∪N =M
B
B
AN = [AN i.o. ] ∈ F2−M0 and so Ω0 ∈ F0+ and now
P(Ω0 ) = lim P(∪∞
N =M AN ) ≥ p0
M →∞
and so P(Ω0 ) = 1.
2.1
Markov transition functions
You may know that
P(X0 = x0 , ..., XN = xN )p(x0 , x1 )...p(xN , xN −1
and this is equivalent to
P(XN = xN |X0 = x0 , ..., XN −1 = xN −1 ) = p(xn , xN −1 )
Our aim is to come up with something similar in the continuous case. We intuitively
think of the following as the probability that we start at x and end up in dy in time t.
14 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
Definition 2.12 A Markov transition kernel is a function p : (0, ∞)×Rd ×B(Rd ) → [0, 1]
such that
1. A 7→ pt (x, A) is a probability measure on B(Rd )
2. x 7→ pt (x, A) is measurable.
Definition 2.13 A Markov process (Xt , t ≥ 0) on Rd with transition kernel stared at x
means that
Z
ptN −tN −1 (xN −1 , dxN )...pt1 (x, dx1 )
P(Xt1 ∈ A1 , ..., XtN ∈ AN ) =
A1 ,...,AN
Observe that we have that
Z
E(f (Xt1 , ..., XtN )) =
RN
Example 2.3
f (x1 , ..., xN )ptN −tN −1 (xN −1 , dxN )...pt1 (x, dx1 )
1. Brownian motion with pt (x, dy) =
2
√ 1 e−(x−y) /2t dy
2πt
2. Xt = Bt + x + ct
3. Reflected Brownian motion Xt = |x + Bt |
(
x + Bt
4. Absorbed Brownian motion Xt =
0
t<τ
where τ = inf{t : Bt = 0}
t≥τ
5. Ornstein-Uhlenbeck processes
6. RadialqBrownian motion. Suppose that Bt is Brownian motion on Rd and define
Pd
i 2
Xt =
i=1 (Bt )
7. dX = µ(X)dt + σ(X)dBt an SDE.
8. A Brownian Bridge is NOT an example. However it is a time inhomogeneous
Markov process, where p is a function of two times.
Definition 2.14 We define Btx = x + Bt i.e. Brownian motion starting at x.
We check that (Btx , t ≥ 0) is a Markov process started at x with kernel pt (x, dy) =
2
1
qt (y − x)dy where qt (z) = √2πt
e−z /2t The short proof is as follows:
P(Btx1 ∈ dx1 , ..., BtxN ∈ dxN ) = P(Btx1 ∈ dx1 , Btx2 − Btx1 ∈ d(x1 − x2 ), ..., BtxN − BtxN −1 ∈ d(xN − xN −1 ))
= qt1 (x1 − x)dx1 qt2 −t1 (x2 − x1 )dx2 ...qtN −tN −1 (xN − xN −1 )dxN
and then we should integrate both sides over A1 , ..., AN .
The longer version is as follows:
E(f (Btx1 , ..., BtxN )) = E(g(Btx1 − x, Btx2 − Btx1 , ..., BtxN − BtxN −1 ))
15 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
if
g(y1 , ..., yN ) = f (x + y1 , x + y1 + y2 , ...)
Z
=
g(y1 , ..., yN )qt1 (y1 )...qtN −tN −1 (yN )dy1 ...dyN
N
ZR
=
f (x1 , ..., xN )qt1 (x1 − x)...qtN −tN −1 (xN − xN −1 )Jdx1 ...dxN
RN
where J is the jacobian of the change of variables y1 = x1 − x,...,yN = xN − xN −1 , and
equals 1.
Definition 2.15 Xt , t ≥ 0 is a Markov process with transition kernel p means
P(Xt ∈ A|FsX ) = pt−s (Xs , A)
for all s < t.
This is shorter than the previous definition, and implies it by an induction argument,
which we will do.
2.1.1
Conditional Expectation
Suppose that we have a probability space (Ω, F, P) and a random variable X : Ω → R.
Suppose G ⊂ F is a sub σ field.
Our aim is to take E(X|G) to be the closest G measurable function to X.
The natural place for this is L2 (Ω, F, P) = {X : Ω → R|E(X 2 ) < ∞} and this is
an inner p
product space with inner product (X, Y ) = E(XY ) and a corresponding norm
||X||2 = E(X 2 ).
For X ∈ L2 (F) there exists a unique Y ∈ L2 (G) such that
(X − Y, Z) = 0
for all Z ∈ L2 (G). This means
E((X − Y )Z) = 0 ⇐⇒ E(XZ) = E(Y Z)
for all Z ∈ L2 (G). It is enough though to check for Z = χA for A ∈ G, by the standard
machine of measure theory. This expectation is then
Z
Z
XdP =
dP
(2.1)
A
A
and we write Y = E(X|G).
Proposition 2.16 For X ∈ L2 (F) there is a unique Y ∈ L2 (G) satisfying (2.1). This
can be improve to X ∈ L1 (F) or to X ≥ 0.
There are several special cases
1. G = {∅, Ω} and then E(X|G) = E(X)
(
y1 A
2. G = {∅, A, Ac , Ω} and then E(X|G) =
. If X = χB then y1 = P(B|A) and
y2 Ac
y2 = P(B|Ac ) and this is like an extension to Bayes formula
16 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
3. G = F and then E(X|G) = X.
Lemma 2.17 Suppose X is G measurable and Z is independent of G, and φ : R2 → R is
bounded and measurable. Then
E(φ(X, Z)|G) = E(φ(x, Z))|x=X
This can be proved using the measure theory machine.
Example 2.4
1. E(Bt |FsB ) = E(Bs + (Bt − Bs )|FsB ) = Bs
2. E(eBt |FsB ) = E(eBt −Bs eBs |FsB ) = eBs e(t−s)/2
3. E(Bt2 |FsB ) = E(Bs2 + 2Bs (Bt − Bs ) + (Bt − Bs )2 |FsB ) = Bs2 + 2Bs E(Bt − Bs ) + E(Bt −
Bs )2 = Bs2 + t − s
4. (X, Y ) a Gaussian vector on R2 with mean 0 and Y 6= 0. We then postulate that
E(X|σ(Y )) = αY
but we need to find α. We do this as follows. Set X = αY + (X − α)Y and we
)
need to check that E(Y (X − αY )) = 0. Thus we choose α = E(XY
, and so then
E(Y 2 )
E(X|σ(Y )) = E(αY + (X − α)Y |σ(Y )) = αY + 0.
R
5. We can do a similar thing for E( 0 1Bs ds|σ(B1 ) and set it equal to αB1 . We then
get α = 1/2.
Lemma 2.18 X ≥ 0 implies that E[X|G] ≥ 0 a.s.
Proof Y = E[X|G] means that E(Y Z) = E(XZ) for measurable Z. Let A = {ω :
Y (ω) < 0}. Then
Z
Z
XdP =
Y dP ≤ 0
0≤
A
A
and A is G measurable, so P(A) = 0.
Q.E.D.
We check that Brownian motion is Markov with this new definition. Let Xt = x + Bt .
Note that FsX = FsB and so
P(Xt ∈ dy|FsX ) = P(x + Bs + Bt − Bs ∈ dy|FsB )
= P(Z + N (0, t − s) ∈ dy)|z=x+Bs = qt−s (y − z)dy|z=x+Bs
= qt−s (y − Xs )dy
Suppose now that Xt = |Btx | and then FsX ⊂ FsB and we guess that the transition
kernel is given by
q̂t−s (y − x)dy = qt−s (y − x)dy + qt−s (−y − x)dy
since we can either reach dy or −dy at time t. We consider P(Xt ∈ dy|FsX ) = P(|Btx | ∈
dy|FsX ). Then using the tower property we get
P(|x + Bt | ∈ dy|FsB ) = q̂t−s (y − (x + Bs ))dy = q̂t−s (y − |x + Bs |)dy = q̂t−s (y − Xs )dy
and since this is already Xs measurable, conditioning on Xs does nothing.
17 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
However there is a warning. Is Xt = f (Btx ) still Markov? The answer is usually not.
In the above case,
p were were just lucky to get this. For example, consider radial Brownian
motion, Xt = B1 (t)2 + ... + Bd (t)2 and we want to show that
Z
E(f (Xt1 , ..., XtN )) = f (x1 , ..., xN )ptN −tN −1 (dxN − xN −1 )...pt1 (dx1 − x)
The measure theory machine says that it is enough to check for f (x1 , ..., xn ) =
We show this by induction:
Qn
1
φk (xk ).
n
n
Y
Y
E( φk (xk )) = E(E( φk (xk )|FtX ))
N −1
1
1
n−1
Y
= E(
φk (xk )E(φn (Xtn |FtX ))
n−1
1
n−1
Y
= E(
Z
φk (xk ))
φn (xn )ptn −tn−1 (xtn−1 , dxn )
1
and then we use induction hypothesis.
2.2
Strong Markov Process.
Define Xs (w) = BT +s − BT = BT (w)+s (w) − BT (w) (w) and is it still a Brownian motion?
This is the case if Ta = inf{t : Bt = a}, but not if T = sup{t ≤ 1 : Bt = 0}. In other
words, T must not look into the future.
Definition 2.19 (Ft , t ≥ 0) is called a filtration on (Ω, F, P) if Fs ⊆ Ft ⊆ F for s ≤ t.
The key example is (Xt , t ≥ 0) and FtX = σ(Xs : s ≤ t).
Definition 2.20 T : Ω → [0, ∞] is called a stopping time for a filtration (Ft , t ≥ 0) if
{T ≤ t} ∈ Ft for all t ≥ 0.
The key example is (FtB , t ≥ 0) and Ta = inf{t : bt = a}. We give two justifications
as to why this is a stopping time below.
{Ta ≤ t} = {sup Bs ≥ a} = { sup
s≤t
Bs ≥ a} ∈ FtB
s∈Q∩[0,t]
since sup Bs is a measurable object.
{Ta ≤ t} = ∩∞
N =1 ∪q∈Q∩[0,t] {|Bq − a| ≤
1
}
N
On the example sheet, we have TK = inf{t : Bt = K} for K ⊂ Rd closed is a stopping
time, but if K is open then it is not a stopping time, as you cannot write it in FtB .
B.
However, you can write it in Ft+
Theorem 2.21 B a Brownian motion, FtB , T is an FtB stopping time with T < ∞.
Then define Xs = BT +s − BT and it is a Brownian motion, and independent of FTB
Definition 2.22 (Information up to a stopping time) Suppose that T is a stopping
time. Then we define
FT = {A : A ∩ {T ≤ t} ∈ Ft for all t ≥ 0}
18 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
We need to check that FT is indeed a σ-field, that if S ≤ T then FS ⊆ FT for two stopping
times and that T is FT measurable.
To show the first of these note that ∅, Ω are clearly in FT . Then if A ∈ FT
Ac ∩ {T ≤ t} = {T ≤ t} \ (A ∩ {T ≤ t}) ∈ Ft
Now if A1 , A2 , ... ∈ FT then
∞
\
AN ∩ {T ≤ t} =
1
∞
\
(AN ∩ {T ≤ t})
1
To show the last note that it is enough to check that {T ≤ s} ∈ FT for all s ∈ R.
Then
{T ≤ s} ∩ {T ≤ t} = {T ≤ min(s, t)} ∈ Fmin(s,t) ⊆ Ft
We now justify the part of the Reflection principle that before was somewhat hand
wavy. This here is rigorous though.
P(Ta ≤ t) = P(Ta ≤ t, Bt ≥ a) + P(Ta ≤ t, Bt ≤ a)
and the former is P(B)t ≥ a) and the latter we consider below:
P(Ta ≤ t, Bt ≤ a) = P(Ta ≤ t, Xt−Ta ≤ 0)
= E(P(Ta ≤ t, Xt−Ta ≤ 0|FTBa ))
= E(χTa ≤t P(Xt−s ≤ 0)|s=Ta )
1
= P(Ta ≤ t)
2
Proof (of theorem 2.21) We first assume that T is discrete, i.e. Ω = ∪k {T = tk }, i.e.
T only takes the values t1 < t2 < .... Pick A ∈ FTB and look at
X
E(F (Bt+T − BT )χA ) =
E(χA χ{T =tk } F (Btk +t − Btk , t ≥ 0))
k
=
X
E(χA∩{T =tk } F (Btk +t − Btk , t ≥ 0))
k
but A ∩ {T = tk } = A ∩ {T ≤ tk } \ A ∩ {T ≤ tk−1 } ∈ FtBk and so by the first Markov
property we get the above equal to
X
E(χA χ{T =tk } )E(F (Btk +t − Btk )) = E(F (Bt , t ≥ 0))P(A)
k
We now consider general T < ∞. We approximate it by discrete stopping times.
Define
j
j−1 j
TN =
if T ∈ [
, )
N
N N
Then TN → T as N → ∞ and BTN → BT due to the continuity of B. We now must
verify that TN is a stopping time
{TN ≤ t} = {T <
j−1
B
} ∈ FB
j−1 ⊆ Ft
N
N
19 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
j
if t ∈ [ j−1
N , N ). Then the above implies that BTN +t − BTN is a Brownian motion and
is independent of FTBN ⊇ FTB If we let N → ∞ then considering characteristic functions
gives
E(ei
P
θk (BTN +sk −BTN )
χA ) = E(ei
P
θk (BTN +sk −BTN )
= E(ei
P
θk (BT +sk −BT )
where we have used the DCT in the ultimate equality.
P(A)
Q.E.D.
We consider the hitting time process (Ta : a ≥ 0). Let St = maxs≤t Bs and then
a 7→ Ta is trying to be the inverse function to St but Ta has jumps in any interval (b, c)
so this isnt a true inverse.
Proposition 2.23 Tb − Ta is independent of (Tc , c ≤ a) and Tb − Ta = Tb−a
This is an example of an independent increments process.
Proof Define Xt = BTa +t − a and this is a new Brownian motion and is independent
D
of FTBa Then Tb − Ta = inf{t : Xt = b − a} = Tb−a and Tc is measurable with respect to
FTBc ⊆ FTBa .
Q.E.D.
Proposition 2.24 Let Z = {t : Bt = 0}. Then
1. Leb(Z) = 0
2. There are no isolated points in Z.
3. Z is uncountable
4. The Hausdorff dimension of Z is 1/2.
Recall that t ∈ Z is isolated if there exists ε > 0 such that (t − ε, t + ε) ∩ Z = {t}.
Proof
Z ∞
Z ∞
χ(t ∈ Z)dt =
χ(Bt = 0)dt
leb(Z) =
0
0
R∞
and E(leb(Z)) = 0 P(Bt = 0)dt = 0
Let τs = inf{t ≥ s : Bt = 0} and then (Bτs +t − Bτs ) is a new Brownian motion and
the Law of the Iterated Logarithm gives that τs is not isolated. Let Ẑ = {τs : s ∈ Q}.
We ask whether Ẑ = Z. This is no, because we can take the last zero before a point and
this will not be of the form τs for some s.
Pick τ ∈ Z \ Ẑ and then τ is not isolated from the left. Then pick a sequence sk → τ
and then sN ≤ τsN < τ for all N and so τsk → τ .
Z is closed as it is the preimage under a continuous map of a closed set. Thus by the
deterministic lemma below we have our result.
Q.E.D.
Lemma 2.25 Suppose that A ⊂ R is closed and has no isolated points. Then A is
uncountable.
Proof Pick t0 < t1 in A and choose B0 = B(t0 , ε0 ) and B1 = B(t1 , ε1 ) disjoint. Then
choose points t00 , t01 inside B0 and A and t10 , t11 inside B1 and A and again choose disjoint
balls around these points. Continue this process.
We now have a chain B1 ⊇ B10 ⊇ ... of balls and so there exists a unique point ta
in the infinite intersection. A is closed so ta ∈ Z. If a 6= b then ta 6= tb so the set is
uncountable.
Q.E.D.
20 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
2.3
Arcsine Laws for Brownian motion
M is the unique time in [0, 1] where it is equal to its supremum, then
1
dt
P(M ∈ dt) = p
π t(1 − t)
but why is M symmetric?
The last zero T = sup{t ≤ 1 : Bt = 0} then
1
P(T ∈ dt) = p
dt
π t(1 − t)
Let L =
R1
0
χ(Bs > 0)ds, i.e. the time it is above zero. Then
1
P(L ∈ dt) = p
dt
π t(1 − t)
These are called arcsine laws due to the following:
Z
0
t
1
1
p
dt =
π
π t(1 − t)
Z
√
sin−1 ( t)
0
√
2 sin θ cos θ
2
dθ = sin−1 ( t)
sin θ cos θ
π
For the second law, note
P(T ≤ t) = P(X does not hit − Bt before time 1 − t)
and using the following
P(a + X does not hit the origin by time 1 − t) = P(B does not hit a by time 1 − t)
= 1 − P(B does hit a by time 1 − t)
= 1 − 2P(B1−t > a)
= 1 − 2P(N (0, 1 − t) > a)
a
= 1 − 2P(N (0, 1) > √
)
1−t
Z ∞
1
2
√ e−z /2 dz
=1−2
2π
√a
1−t
a
=: 1 − 2Φ √
1−t
and so
a
)
P(T ≤ t) = E(1 − 2Φ √
1 − t a=Bt
√
t|N (0, 1)|
√
= E(1 − 2Φ
)
1−t
√
Z ∞
t|x|
1
2
√ e−x /2 dx
=−
1 − 2Φ √
1−t
2π
−∞
21 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
and if you consider densities you get
4
P(t ∈ dt) = − √
2π
and this is of the form
R∞
0
Z
∞
2
tx
− 1−t
−x2 /2
e
e
0
!
√
1
x
tx
√ √
dx
+
3
2 t 1 − t (1 − t) 2
2
axe−bx dx which eventually gives the answer.
Theorem 2.26 (Lévy) Suppose St = sups≤t Bs . Then Xt = St − Bt is a reflected
Brownian motion.
Then this would give that the alst zero for X is equal to the time of maximal value
for B, and so the first two arcsine laws one would expect to be the same.
D
Corollary 2.27 T = M
Remark There is almost a random walk analogue. If we do the same thing, then we get
a reflected random walk, but it is “ sticky ” at the origin. We expect the set of sticky
times to be small though, because
Z t
χ(Bs = 0)ds = 0
0
Proof It order to prove Lévy’s theorem, it is left to the reader to check that X is a
Markov process with correct transition kernel as given before.
Then one can find P(Xt1 ∈ A1 ...XtN ∈ AN ). Then
P(Xt ∈ dy|FsX ) = P(Xt ∈ dy|FsB ) =: I + II|a=Ss −Bs
Where I and II are as below. Call Yr = Xs+r − Xs adn this is a new Brownian motion
independent of FsB . We then have two possibilities. Either I, it attains its maximum
after s, or it attains it before. We then have
Z ∞
Y
Y
I=
P(St−s
∈ dz, Bt−s ∈ dz − y)
II = P(St−s
≤ a, B ∈ a − dy)
a
Q.E.D.
3
Brownian Martingales
Definition 3.1 Fix a filtration (Ft , t ≥ 0). A process (Mt , t ≥ 0) is called a (Ft )martingale if
E(Mt |Fs ) = Ms
∀s ≤ t
Observe that this requires E|Mt | < ∞ for all t ≥ 0
Theorem 3.2 (Optional Stopping Theorem) If (Mt , t ≥ 0) is an Ft martingale and
T is a bounded stopping time then
E(MT ) = E(M0 )
22 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
There is a financial interpretation of this. Mt could be considered as a stock price.
Then the martingale property is the natural assumption, in discounted money. In other
words the best way to predict the future is to consider the present. Then the OST tells
you that the expected selling value MT equals the initial value M0 , you cannot make
money from a martingale.
Some books use the OST as the definition of a martingale.
T bounded means that there exists a K ∈ R such that P(T ≤ K) = 1
Consider Ta = inf{t : Bt = a}. This is not bounded. Then BTa = a so E(BTa ) = a
but E(B0 ) = 0 and these are not equal.
Example 3.1 (Bt ) is a martingale because E(Bt |FsB ) = E(Bs + Bt − Bs |FsB ) = Bs Take
T = Ta ∧ Tb with a < 0 and b > 0. This is the minimum of two stopping times and so is a
stopping time. It is also bounded. We now use the Optional Stopping theorem.. We have
E(BT ∧N ) = 0 and so by DCT we have E(BT ) = 0 since |BT ∧N | ≤ max(|a|, |b|). Note to
always check that this can be done. We then have
0 = E(BT ) = bP(Tb < Ta ) + aP(Ta < Tb )
and we also have that
1 = P(Tb < Ta ) + P(Ta < Tb )
Solving these gives
P(Tb < Ta ) =
−a
b−a
P(Ta < Tb ) =
b
b−a
Example 3.2 (Bt2 − t, t ≥ 0) is a martingale.
E(Bt2 |FsB ) = E(Bs2 + 2Bs (Bt − Bs ) + (Bt − Bs )2 |FsB ) = Bs2 + t − s
and so E(Bt2 − t|FsB ) = Bs2 − s as we want for a martingale. The OST says that
E(BT2 ∧N − T ∧ N ) = 0
and using DCT and MCT we get that
E(BT2 − T ) = 0
and this gives that
E(T ) = E(BT2 ) = b2 P(Tb < Ta ) + a2 P(Ta < Tb ) =
−b2 a
a2 b
+
= −ab
b−a b−a
2
Example 3.3 (eθBt −tθ /2 , t ≥ 0) is a martingale for all θ ∈ R But how is this a martingale
when it seems to be small always. The reason is the process goes to zero but the expectation
doesn’t. We now check that this is a martingale.
E(eθBt |FsB ) = E(eθBs eθ(Bt −Bs ) ) = eθBs eθ
2 /2(t−s)
and then rearranging gives it as a martingale. We then use the DCT to Ta ∧ N and
2
eθBt −tθ /2 to get, with OST,
θB
)=1
E(e Ta ∧N )−Ta ∧N θ2 /2
23 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
and Ta ∧ N → Ta and BTa ∧N → BTa = a and then as N → ∞ we get E(e−θ
Then if we let λ = θ2 /2 we get
√
E(e−λTa ) = e−
2 /2T
a
) = e−θa .
2λa
Before we showd that P(Ta ≤ t) = 2P(Bt ≥ a) and we get here
Z ∞
a 1
2
−λTa
E(e
)=
e−λt 3/2 √ e−a /2t dt = ...
t
2π
0
and try this for yourself.
The above examples give a hint that many interesting facts about Brownian motion
can be found from the Optional stopping theorem, as applied to some stopping time.
Example 3.4 Ta ∧ Tb can be done similarly and inversion can be used to give P(Ta ∧ Tb ∈
dt).
Recall that E(e−λT ) = φ(λ) = 1 + a1 λ + a2 λ2 + ... = 1 − λE(T ) + λ2 /2E(T 2 ) + ... by
comparing coefficients.
If we define Xt = Bt − ct then Xt → −∞. FindRP(Ta < ∞). We consider the
2
2
martingale eθBt −θ /2t = eθXt e(θc−θ /2)t and define Ta = {t : Xt = a}. Then OST gives
you
2
E(eθXTa ∧N e(θc−θ /2)Ta ∧N ) = 1
(
eθa {Ta < ∞}
θX
Now e Ta ∧N →
and this is dominated by eθa .
0
{Ta = ∞}, θ > 0
(
2
e(θc−θ /2)Ta {Ta < ∞}
2
(θc−θ
/2)T
∧N
a
Now e
→
. If we use θc − θ2 /2 =
0
{Ta = ∞}, θc − θ2 /2 < 0
λ then we get
E(e−λTa χ{Ta <∞} ) = e−θa
if θ > 0 and θc − θ2 /2 < 0. Now choose θ → 2c and so P(Ta < ∞) = e−2ca . This makes
sense. If a was big, then the chance is small, and if c is large, i.e. the drift is strongly
negative, then the chance is also small.
Theorem 3.3 (OST version 1) Suppose (Mt : t ≥ 0) is an Ft martingale with continuous paths. Suppose also that T is a bounded stopping time and M is bounded. Then
E(MT ) = E(M0 )
Proof Suppose first that T is discrete, and T ≤ K. Let T ∈ {t1 , .., tN } where t1 < t2 <
... < tN ≤ K. Then
N
X
E(MT ) =
E(Mtk χ{T =tk }
k=1
and also E(Mt |Fs ) = Ms and so E(Mt χA ) = E(Ms χA ) for A ∈ Fs . Also {T = tk } = {T ≤
tk } \ {T ≤ tk−1 } ∈ Ftk and so
N
X
E(Mtk χ{T =tk } = E(Mk ) = E(M0 )
k=1
24 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
j
Suppose now that we have a general stopping time T . Define TN = Nj if T ∈ [ j−1
N , N ).
TN are discrete stopping times and TN → T . Also MTN → MT by the continuity of paths
and so
E(M0 ) = E(MTN ) → E(MT )
using the DCT since |Mt (ω)| ≤ K for some K, and for all t.
(1)
(d)
Q.E.D.
(2)
Bt = (Bt , ..., Bt ) is a Brownian motion on Rd . Suppose T = inf{t : Bt = 1}.
(1)
Find P(Bt ∈ dx). Solve using T and B (1) are independent and you know P(T ∈ dt).
Proposition 3.4 (Lots of Brownian martingales) Suppose that B is a d dimensional
Brownian motion, then
Z t
∂f
1
f (Bt , t) −
(Bs , s) + ∆f (Bs , s) ds
∂s
2
0
2
∂f
∂ f
is a martingale if f ∈ C 2,1 and f, ∂f
∂s , ∂xj , ∂xi ∂xj are of at most exponential growth.
For example, Bt2 − t with f (x, t) = x2 then 12 ∆f = 1. Also Bt4 isn’t but if we subtract
2
θBt −θ2 /2t we choose f (x, t) = eθx−θ2 /2t and we subtract nothing.
0 6Bs ds it is. Also for e
If Btx = x + Bt is d dimensional Brownian motion started at x then the above is the
same with Bt replaced with Btx . We will prove it soon, and the proof uses the fact that
1
t
the Gaussian density solves ∂φ
∂t = 2 ∆φt
1
Of special importance are f (x, t) satisfying ∂f
∂s + 2 ∆f = 0 or ∆f = 0.
Consider the Dirichlet problem. Let D ⊂ Rd be open. Then find u such that
(
∆u(x) = 0
in D
(3.1)
u(x) = f (x) on ∂D
Rt
with f a given function.
Theorem 3.5 Suppose D is a bounded open set of Rd . Suppose u ∈ C 2 (Rd ) and solves
(
∆u(x) = 0
in D
u(x) = f (x) on ∂D
then
u(x) = E(f (BTx ))
where T = inf{t : Btx ∈ ∂D}.
Proof We can modify u outside of D so that it
at most exponential growth. Then
R t has
1
x
we get, by the above proposition that u(Bt ) − 0 2 ∆u(Bsx )ds is a martingale. Then the
optional stopping theorem, for T ∧ N , gives
Z T ∧N
1
x
x
u(x) = E u(BT ∧N ) −
∆u(Bs )ds = E(u(BTx ∧N ))
2
0
and by the DCT we get u(x) = E(u(BTx )) = E(f (BTx )) since u = f on ∂D.
25 of 40
Q.E.D.
MA4F7 Brownian motion Lecture Notes Autumn 2012
(
Example 3.5 Let u(x) =
1
|x|d−2
d≥3
and this satisfies ∆u = 0 except at x = 0. I
log |x| d = 2
leave this as a check for the reader, as it has been seen before many times. Let D = {x ∈
R3 : a < |x| < b} and let Ta = inf{t : |Btx | = a} and Tb = inf{t : |Btx | = b}. Then let
T = Ta ∧ Tb , and so
1
1
1
1
= E( x ) = P(Ta < Tb ) + P(Tb < Ta )
|x|
|BT |
a
b
with 1 = P(Ta < Tb ) + P(Tb < Ta ) by the law of the iterated logarithm, and so
P(Ta < Tb ) =
1
1
|x| − b
1
1
a − b
If we let b → ∞ and we get P(Ta < Tb ) → P(Ta < ∞) =
a
|x| .
Corollary 3.6 |Bt | → ∞ as t → ∞ in d ≥ 3.
Proof If Bt doesnt tend to infinity then there exists K and tN → ∞ where |BtN | ≤ K.
Let TN = inf{t : |Bt | ≥ N } then the law of the iterated logarithm says TN < ∞. Then
d−2
(N )
Xt = BTN +t − BTN is a Brownian motion and P(X N hits B(0, K)) = K
and thus
N
N
P(∩∞
hits B(0, K)}) = 0
N =K {X
and so
∞
N
P(∪∞
hits B(0, K)}) = 0
N =1 ∩N =K {X
Q.E.D.
For d = 2 let b → ∞ and then P(Ta < ∞) = 1 and then if a → 0 then P(T{0} < Tb ) = 0
and if we now let b → ∞ then P(T{0} < ∞) = 0
Corollary 3.7 In d = 2, P(B ever hits {x}) = 0 for x 6= 0
Corollary 3.8 Leb(Bt : t ≥ 0) = 0 a.s. and so is not a space filling curve.
Proof Leb(Bt : t ≥ 0) =
R
R2
χx∈(Bt ,t≥0) dx and we consider expectations.
Z
E(Leb(Bt : t ≥ 0)) = E
R2
Z
χx∈(Bt ,t≥0) dx =
R2
P(x ∈ Bt )dx = 0
Q.E.D.
Corollary 3.9 The range of (Bt , t ≥ 0) is dense in R2
The Poisson problem is the following, for D ⊂ Rd .
(
1
in D
2 ∆u = −g
u=0
on ∂D
(3.2)
and here g represents pumping heat in or out, and u is the steady state temperature.
26 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
Theorem 3.10 (Poisson version 1) Suppose u ∈ C 2 (R) solves equation (3.2) with D
bounded. Then
Z T
u(x) = E(
g(Bsx )ds)
0
with T = inf{t : Btx ∈ ∂D}
Example 3.6 u(x) =
Then u(x) = E(T ).
R2 −|x|2
d
solves the Poisson problem on a ball of radius R with g = 1.
Example 3.7 Suppose D = (0, ∞) ⊂ R and u(x) = −x2 /2. However, E(T ) 6= u(x)
Proof We can modify u away from D to have exponential growth. Then
Z t
1
x
u(Bt ) −
∆u(Bsx )ds
2
0
is a martingale. The OST then gives
Z T ∧N
Z T ∧N
1
x
x
x
∆u(Bs )ds) = E(u(BT ∧N ) +
g(Bsx )ds)
u(x) = E(u(BT ∧N ) −
2
0
0
RT
and the former tens to 0 by DCT and the latter tends to 0 g(Bsx )ds by domination with
(1)
||g||∞ T , and so we need E(T ) < ∞ but T ≤ inf{t : |Bt | = K} and the right hand side
is finite.
Q.E.D.
PDE people generally look for solutions to these problems that are in C 2 (D) ∩ C(∂D).
Sadly, one cannot in general extend these to C 2 (Rd ) or even C 2 (D̄).
Example 3.8 21 ∆u = −1 on (0, 1)2 and u = 0 on the boundary. Then u is not C 2 ([0, 1]2 )
because of the effects at the corners, as at (0, 0) we have ∆u = 0.
We can improve though, to allow solutions in C 2 (D) ∩ C(∂D) but still u(x) solves the
Dirichlet or Poisson problem. We do this as follows.
Shrink the domain in by ε by defining Dε = {x : d(x, DC ) > ε} and then Dε → D
and we can then mollify. Let Tε = inf{t : Bt ∈ ∂Dε } and then BTε → BT . Then one can
believe that you can modify u outside of Dε to be C 2 (Rd ) with exponential growth. Then
apply version 1 to Dε to get u(x) = E(u(BTε )) → E(f (BT )) as ε → 0.
By exponential growth, we mean |g(t, x)| ≤ C0 eC1 |x| for all x, t.
Proof ( of theorem 3.4) We first show that E(|Mt |) < ∞
x
x
x
2
E(|f (Btx , t)|) ≤ C0 E(eC1 |Bt | ) ≤ C0 E(eC1 Bt + e−C1 Bt ) ≤ C0 eC1 x eC1 t/2 < ∞
Rt
and similarly E(| 0 Lf (Bsx , s)ds|) < ∞ WLOG we can take x = 0 since we can shift and
we still have exponential growth.
Z t
E(Mt |FsB ) = E(f (Bt , t) −
Lf (Br , r)dr|FsB )
0
Z s
Z t
B
= E f (Bs , s) + (f (Bt , t) − f (Bs , s)) −
Lf (Br , r)dr −
Lf (Br , r)dr|Fs
0
s
Z t−s
= Ms + E(f (Bs + Xt−s , t) − f (Bs , s) −
Lf (Xr + Bs , s + r)dr|FsB )
0
Z t−s
= Ms + E(f (Z + Xt−s , t) − f (Bs , s) −
Lf (Xr + Z, s + r)dr|FsB )|Z=Bs
0
27 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
where Xr = Bs+r −Bs . The result now follows from E(g(Xt , t)−g(0, 0)−
0 where we have used g(y, t) = f (z + y, t + s). Alternatively this is
Rt
0
Lg(Xr , r)dr) =
t
Z
E(g(Xt , t) − g(0, 0) =
Lg(Xr , r)dr)
0
which is the same as
d
E(g(Xt , t)) = E(Lg(Xt , t))
dt
Then
Z
d
1
d
2
g(x, t)
e−|x| /2t dx
E(g(Xt , t)) =
d/2
dt
dt Rd
(2πt)
Z ∂g
∂φ
=
(x, t)φ(x, t) +
g dx
∂t
∂t
Rd
Z
∂g
1
=
(x, t)φ(x, t) + (∆φ)g dx
∂t
2
Rd
Z ∂g
1
(x, t)φ(x, t) + (∆g)φ dx
=
∂t
2
Rd
= E(Lg(Xt , t))
as required.
Q.E.D.
Theorem 3.11 (Heat Equation) If u ∈ C 1,2 ([0, ∞) × Rd ) is of exponential growth and
solves
(
1
∂u
t > 0, x ∈ Rd
∂t = 2 ∆u
u(0, x) = f (x) x ∈ Rd
then
u(x, t) = E(f (Btx )) =
Z
f (y)
Rd
1
2
e−|x−y| /2t dx
(2πt)d/2
Theorem 3.12 (Heat Equation on a region) Suppose D ⊂ Rd is bounded, and u ∈
C 1,2 ([0, ∞) × D) ∩ C([0, ∞) × D̄) solves

∂u
1

t > 0, x ∈ D
 ∂t = 2 ∆u
u(0, x) = f (x) x ∈ D


u(t, x) = g(x) x ∈ ∂D, t > 0
and let T = inf{t : Btx ∈ ∂D} and then
u(x, t) = E g(BTx )χ{T ≤t} + f (Btx )χ{T >t}
We prove both theorems simultaneously. Proof Fix t > 0 and consider the map
Z s
∂u 1
x
+ ∆u (Brx , r)dr
s 7→ u(Bs , t − s) −
−
∂r
2
0
in other words we run backwards in time. Then in Rd u(Bsx , t − s) is a martingale so using
the OST we get
u(x, t) = E(u(Btx , 0)) = E(f (Btx ))
28 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
as required.
On D we stop at T ∧ t which is a bounded stopping time. Then u(Bsx , s − t) is a
martingale so
u(x, t) = E(u(BTx ∧t , t − T ∧ t)) = E(χ{T ≤t} g(BTx ) + χ{T >t} f (Btx ))
as required.
Q.E.D.
Example 3.9 Brownian motion staying in a tube. Suppose we write u(x, t) = P(|Bt | <
1 for all s ≤ t). Then u(x, t) solves the equation

∂u
1 ∂2u

 ∂t = 2 ∂x2 x ∈ (−1, 1), t ∈ (0, t)
u(0, x) = 1


u(t, x) = 0 x = ±1
but note that u 6∈ C([0, t] × [−1, 1]) and the solution is
∞
X
(2k − 1)πx − (2k−1)2 π2 t
2
e
u(x, t) =
ak cos
2
k=1
2
−π t/8 and then this solves the above
Inspired by this, we consider the first term cos πx
2 e
πx
equation with initial data u(0, x) = cos 2 and then
Btx π
−π 2 t/8
χ
u(t, 0) = e
= E cos
≤ P({T > t})
2 {T >t}
and so P(T > t) ≥ e−π
2 t/8
, and this is correct asymptotically.
Example 3.10 Find P(Bsx1 , ..., Bsxd do not collide by time t). Let Btx = (Bsx1 , ..., Bsxd ) be
a d-dimensional Brownian motion. Let Vd = {x ∈ Rd : x1 < x2 < ... < xd } which is called
a cell, and then ∂Vd = {x ∈ Rd : xi = xi+1 for some i}. This solves

1
∂u

x ∈ Vd , t > 0
 ∂t = 2 ∆u
u(0, x) = f (x) x ∈ Vd


u(t, x) = 0
x = ∂Vd
and this has a famous solution
φt (x1 − y1 ) . . . φt (x1 − yd )
..
..
u(x, t) =
f (y1 , . . . , yd ) det dy1 . . . dyd
.
.
Vd
φt (xd − y1 ) . . . φt (xd − yd )
Z
where φt (z) =
2
√ 1 e−z /2t
2πt
Definition 3.13 φ : R → R is convex means φ(x) = sup{L(x)|L is linear , L ≤ φ}
Note that if φ ∈ C 2 (R) then φ is convex if and only if φ00 ≥ 0.
Lemma 3.14 (Jensen’s Inequality) If E(|X|) < ∞ and φ : R → R is convex then
E(φ(X)) ≥ φ(E(X))
E(φ(X)|F) ≥ φ(E(X|F)) a.s.
29 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
Proof
E(φ(X)|F) ≥ E(L(X)|F)
= E(aX + b|F)
= aE(X|F) + b
= L(E(X|F)) a.s.
and then taking a supremum over L ≤ φ gives the result. Note we need to ensure somehow
that we only need countably many Ls.
Q.E.D.
Corollary 3.15 Suppose (Mt ) a martingale and φ is convex, then
E(φ(Mt )|Fs ) ≥ φ(E(Mt |Fs )) ≥ φ(Ms )
Lemma 3.16 (OST for φ(Mt )) Suppose (Mt , t ≥ 0) is a martingale with continuous
paths and φ ≥ 0 is a convex function. Suppose T ≤ K is a bounded stopping time. Then
E(φ(MT )) ≤ E(φ(MK ))
Proof First assume that T is discrete, so T ∈ {t1 , ..., tM } with t1 < t2 < ... < tM . Then
E(φ(MT )) =
M
X
E(φ(Mtk )χ{T =tk } ) ≤
M
X
E(φ(MK )χ{T =tk } ) = E(φ(MK ))
k=1
k=1
Now for general T , find discrete stopping times TN ≤ K so that TN → T and then we
have E(φ(MTN )) ≤ E(φ(MK )) and using Fatou’s lemma we get
E(φ(MT )) ≤ E(φ(MK ))
Q.E.D.
We now prove a more general OST, one without boundedness of the martingale
Theorem 3.17 Suppose (Mt , t ≥ 0) is a continuous martingale, and T ≤ K is a bounded
stopping time. Then
E(MT ) = E(M0 )
Proof For discrete T this works as above and is fine.
We now consider Mt ≥ 0 and assume we have discrete stopping times TN → T and
we have E(MTN ) = E(M0 ). We then write x = x ∧ L + (x − L)+ and we have
E(M0 ) = E(MTN ) = E(MTN ∧ L) + E((MTN − L)+ )
and now E(MTN ∧ L) → E(MT ∧ L) by DCT and E(MT ∧ L) = E(MT ) − E((MT − L)+ )
Fix ε > 0 Choose L large so that
E((MTN − L)+ ) ≤ E((MK − L)+ ) < ε
and then take N large so that
|E(MTN ∧ L) − E(MT ∧ L)| ≤ ε
Finally truncate a general martingale at +L and −L. This is a bit messier though.
Q.E.D.
30 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
We now have some final remarks on the Dirichlet problem.
Can we define u(x) = E(f (BTx )) and check that it solves the Dirichlet problem. Suppose that D = {x ∈ R2 : 0 < |x| < 1} the punctured disc. If we suppose that u = 1 on
the set {x : |x| = 1} and 0 at 0 then the solution given by the Brownian motion formula
is u ≡ 1 but this doesn’t solve the boundary conditions.
However, if u is of that form, then u ∈ C ∞ (D) and ∆u = 0 on D always.
We call a point y ∈ ∂D regular if u(x) → f (y) as x → y in D. Thus if all point of
∂D are regular then u is a solution. We thus need a sufficient condition fo y ∈ ∂D to be
regular.
To show the first point, we observe that u being harmonic is the same as u satisfying
the ball averaging property, or the sphere averaging property, namely if B(x, ε) ⊂ D then
Z
1
u(x) =
u(y)dy
|B(x, ε)| B(x,ε)
or
1
u(x) =
SAε
Z
u(y)dS
∂B(x,ε)
Sphere averaging for our formula is almost obvious. If we let Sε = inf{t : Btx ∈ ∂B(x, ε)},
Xt = BSε +t − BSε and T 0 = inf{t : Xt + BSxε ∈ ∂D} then
u(x) = E(f (BTx ))
= E(f (BSxε + BTx − BSxε ))
= E(f (BSxε + XT 0 ))
= E(E(f (BSxε + XT 0 )|FSBε ))
= E(f (z + XT 0 )|z=BSx )
ε
= E(u(z)|z=BSx )
ε
=
E(u(BSxε ))
which is sphere averaging,
To do the part on regular points, the following is an equivalent definition of regular.
P(Tx > ε) → 0 as x → y where Tx = inf{t : Btx ∈ ∂D} for all ε > 0.
This is equivalent to P(σ y = 0) = 1 where σ y = inf{t > 0 : Bty ∈ DC }. The 0-1 law
says that this is either 0 or 1. Remember Lebesgue’s thorn. If y is on a thin enough spike
then P(σ y > 0) = 1 and then you cannot solve the Dirichlet problem.
A sufficient condition for y to be a regular point is the cone condition, namely if there
exists a cone in DC with vertex at y ∈ ∂D then y is regular. This is because
P(σ y ≤ ε) ≥ P(Bε ∈ cone ) = p(α) > 0
where α is the solid angle of the cone. Then letting ε → 0 gives
P(σ y = 0) ≥ p(α) > 0
and so the 0-1 law gives P(σ y = 0) = 1
Book by Richard Bass on this sort of stuff
31 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
4
Donsker’s theorem
The idea is to show random walks converge to Brownian motion. Throughout this chapter
2
we have the following.
1 , Z2 , ... are IID random variables with E(Zi ) = 0 and E(Zi ) = 1
PZ
N
and we define SN = k=1 Zk and this is the position at time N . We can interpolate the
SN to get St and this is given by
St = (N + 1 − t)SN + (t − N )SN +1
(N )
We also define Xt
=
t ∈ [N, N + 1]
(N )
S
√N t
N
but we cannot hope that Xt
→ Bt . However, we do have
(N ) D
that Xt → Bt . This is the aim of this section.
We know that
S
√N → N (0, 1)
N
by the central limit theorem, and so
p
√
SbN tc
SbN tc
bN tc
SN t
(N )
√
Xt = √ = √
+ error = p
+ error → N (0, 1) t
N
N
N
bN tc
we hope. Thus it has the same distribution as Brownian motion.
(N )
(N )
Similarly (Xt1 , ..., Xtk ) → (Bt1 , ..., Btk ) but does
(N ) D
→ max Bt
max Xt
t∈[0,1]
1
Z
Xt
D
(N )
1
Bt dt ∼ N (0, 1/3)
(4.2)
0
Z
0
(4.2) can be rewritten as
Z
dt →
0
(4.1)
t∈[0,1]
1
D
Z
χ{X (N ) >0} ds →
s
PN
1 SK D
→ and (4.3) can
N 3/2
(N )
of (Xt , t ∈ [0, 1]) =:
0
1
χ{Bs >0} ds
be rewritten almost as
(4.3)
number of times when Sk >0
N
The plan is to think
X (N ) and (Bt , t ∈ [0, 1]) = B as random
variables in C[0, 1]. We thus need to show that
D
X (N ) → B
on C[0, 1]. This is a big improvement of the Central limit theorem. We also show that
D
F (X N ) → F (B)
and we observe that
D
D
X (N ) → B =⇒ F (X N ) → F (B)
if F is continuous. In the above questions, the maximum and integral are continuous, but
the last is not, as functions C[0, 1] → R.
Definition 4.1 (E, d) a metric space. X : (Ω, F, P) → (E, d) measurable ( X −1 (B) ∈ F
for B Borel in E)is called an E-valued random variable.
32 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
We use this with E = C[0, 1] with
d(f, g) = sup |f (t) − g(t)|
t∈[0,1]
and an open ball is B(f, ε) = {g : |g(t) − f (t)| < ε} and the Borel sets are generated by
open balls.
Lemma 4.2 B(C[0, 1]) = σ(F0 ) where
F0 = {f : f (t1 ) ∈ O1 , ..., f (tN ) ∈ ON }
for t1 , ..., tN and N ≥ 0 and Oi open sets.
Proof It is easy to check that F0 is a π-system.
We check that σ(F0 ) ⊂ B(C[0, 1]). But if f ∈ {f (t1 ) ∈ O1 , ..., f (tN ) ∈ ON } then since
O1 is open there is an ε1 such that if |f − g| < ε1 then g(t1 ) ∈ O1 , and so on, then take
ε = min{ε1 , ..., εN } and we get B(f, ε) ⊂ {f (t1 ) ∈ O1 , ..., f (tN ) ∈ ON } and so it is open.
We check that B(C[0, 1]) ⊂ σ(F0 ). It is enough to check that B(f, ε) ∈ σ(F0 ). Now
B(f, ε) = {: |g(t) − f (t)| ≤ ε ∀t ∈ [0, 1]}
\
=
{: |g(t) − f (t)| ≤ ε}
t∈[0,1]∩Q
=
\
\
{: |g(t) − f (t)| < ε +
t∈[0,1]∩Q N ≥0
and also B(f, ε) =
S
N ≥1 B(f, ε
1
} ∈ F0
N
− 1/N )
Q.E.D.
An example of this is B = (Bt , t ≥ 0) is a C[0, 1] valued variable. One observes
(N )
that {ω Bt1 ∈ O1 , ..., BtN ∈ ON ∈ F0 Also Xt
are random variables in C[0, 1]. We
can show the latter using composition of measurable functions, Ω → RN → C[0, 1] given
(N )
by ω 7→ (Z1 (ω), ..., ZN (ω)) 7→ Xt (ω) and the former is measurable and the latter is
continuous.
(N ) D
Definition 4.3 X (N ) , X are (E, d) valued variables then Xt
→ X means
E(F (X (N ) )) → E(F (X))
for F : E → R bounded and continuous.
Theorem 4.4 (Donsker)
D
X (N ) → B
on C[0, 1]
D
Theorem 4.5 (Continuous Mapping) If X (N ) → X on (E, d) and G : (E, d) →
˜ which is continuous then
(Ẽ, d)
D
G(X (N ) ) → G(X)
˜
on (Ẽ, d)
33 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
Proof Take F : Ẽ → R bounded and continuous. Then
E(F (G(X (N ) ))) → E(F (G(X)))
since F ◦ G is bounded and continuous.
Q.E.D.
Corollary 4.6
(N )
max Xt
t∈[0,1]
→ max Bt
t∈[0,1]
Proof F (t) = max f (t) is a continuous map.
Q.E.D.
R1
R1
D
Consider 0 χ(X (N ) >0) dt → 0 χ(Bt >0) dt is not continuous. Also recall that if X (N ) →
t
X then this does not imply that P(XN ∈ (a, b)) → P(X ∈ (a, b)), for example consider
XN = 1/N and (a, b) = (0, 1).
Consider the simple random walk SN . Then
P(N (0, 1) ∈ Q). However,
S2
P( NN
2 D
SN
N →
N (0, 1) but P(
2
SN
N
∈ Q) = 1 6= 0 =
∈ (a, b)) → P(N (0, 1) ∈ (a, b))
D
Theorem 4.7 (Extended Continuous Mapping) If X (N ) → X on (E, d) and F :
(E, d) → R is measurable and Disc(F ) = {f : F is discontinuous atf } is such that P(X ∈
Disc(F )) = 0 then
D
F (X (N ) ) → F (X)
˜
on (Ẽ, d)
This is a big non trivial improvement.
For
set to be F is discontinuous at
R1
R 1 the indicator problem, we guess the discontinuity
f if 0 χ(f (t)=0) dt > 0, but for a Brownian path, 0 χ(Bt =0) dt = 0 and so we have the
result we want.
(N )
We now try to prove the approximation. We remind ourselves that Xt = S√NNt with
SN the simple symmetric random walk and we take (Z1 , Z2 , ...) IID with mean 0 and
D
variance 1. The aim is to show that X (N ) → B on C([0, 1]) with the supremum norm.
We embed simple random walks as follows. We define T1 = inf{t : |Bt | = 1} and then
inductively define TN +1 = inf{t ≥ TN : |Bt − BTN | = 1} and then linearly interpolate.
We have
D
(BT1 , BT2 , ...) = (S1 , S2 , ...)
We are close to a proof. We know that E(T1 ) = 1 and the strong Markov property
at TN implies that E(TN +1 − TN ) = 1 and TN +1 − TN is independent of T1 , T2 , ... and so
√
(N )
√N t then we expect X (N ) is close to this.
TN ≈ N ± O( N ) and if we take Bt = B
N
Lemma 4.8 (Skorokhod) Take Z with E(Z) = 0. Then there exists a stopping time
D
T < ∞ so that BT = Z and E(T ) = E(Z 2 ).
Financial mathematicians love this lemma. There are at least 14 different ways to
prove this.
If Bt2 − t is a martingale and so E(BT2 ∧N − T ∧ N ) = 0 or E(T ∧ N ) = E(BT2 ∧N ) and
this, by Fatou, gives E(T ) ≥ |E(BT2 ).
If we take Z to be independent of B we choose T = inf{t : Bt = Z} and so BT = Z
but E(T ) = E(E(T |σ(Z))) = E(E(Ta )|a=Z ) = ∞. This is a bit of a silly example.
34 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
Proof We first suppose Z ∈ {a, b}. Then define T by
(
b
P(BT = a) = b−a
−a
P(BT = b) = b−a
and then E(T ) = −ab.
We now take a general Z. Choose random α < 0, β > 0 independent of B, and use
T = Tα,β . We need to find the distribution of α and β. To this end we need to choose
ν(da, db) = P(α ∈ da, β ∈ db)
and we have the target distribution of
(
µ+ (dz) = P(Z ∈ dz) Z ≥ 0
µ− (dz) = P(Z ∈ dz) Z < 0
If Z ≥ 0 we need
µ+ (dz) = (BTα,β ∈ dz)
= E(P(BTα,β ∈ dz|σ(α, β)))
−α
=E
χ(β ∈ dz)
β−α
ZZ
−a
=
χ(b ∈ dz)ν(da, db)
b−a
Z
−a
=
ν(da, dz)
z−a
R
and so we choose ν(da, dz) = (z − a)µ+ (dz)π(da) where −aπ(da) = 1 and so we have
matched µ+ .
For Z < 0 we have
µ− (dz) = P(BTα,β ∈ dz)
β
=E
χ(α ∈ dz)
β−α
ZZ
b
=
χ(a ∈ dz)ν(da, db)
b−a
Z
= bµ+ (db)π(dz)
and so we choose
π(dz) = R
µ− (dz)
bµ+ (db)
and so α, β are distributed as
ν(da, db) =
(b − a)µ+ (db)µ− (da)
R
xµ+ (dx)
We thus have four things to check:
1. P(BTα,β ∈ dz) = P(Z ∈ dz)
35 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
2.
R
−aπ(da) = 1
3. E(Tα,β ) = E(Z 2 )
RR
4.
ν(da, db) = 1
These are all true, but we only check 2 and 3. Observe that 1 is by construction. For 2,
we want to show that
Z
Z
−aµ− (da) = aµ+ (da)
but we have
Z
0 = E(Z) =
Z
aµ+ (da) +
aµ− (da)
which is what we want.
For 3, observe that
E(Tα,β ) = E(−αβ)
ZZ
=
−abν(da, db)
ZZ
µ+ (db)µ− (da)
=
−ab(b − a) R
xµ+ (dx)
R
R 2
R
R
− xdµ− x dµ+ + x2 dµ− xdµ+
R
=
xdµ+
Z
= x2 dµ+R x2 dµ−
= E(Z 2 )
Q.E.D.
We use a Skorokhod trick:
(
D
BTα,β = Z
E(Tα,β ) = 1
and take IID copies (α1 , β1 ), (α2 , β2 ) independent of B. Then T1 = inf{t : Bt ∈
{α1 , β1 }},...,TN +1 = {t ≥ TN : Bt − BTN ∈ {αN +1 , βN +1 }} and define (S1 , S2 , ...) =
(BT1 , BT2 , ...) and then it has the random walk distribution that we want, i.e.
(N )
Xt
SN t
=√
N
BN t
B (N ) = √
N
The key estimate is
P ||X (N ) − B (N ) ||∞ > ε → 0
as N → ∞. Assume this, and then fix F : C[0, 1] → R that is bounded and uniformly
continuous. We get
|E(F (X (N ) )) − E(F (B (N ) ))| ≤ |E(F (X (N ) ) − F (B (N ) ))|
≤ E|F (X (N ) − F (B (N ) |
≤ 2||F ||∞ P(ΩN,ε ) + E(|F (X (N ) − F (B (N ) |χΩC )
N,ε
Fix η > 0 and use uniform continuity to choose ε so that the second term is less than
or equal to η/2.
36 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
Then choose N large to make the first term less than or equal to η/2.
We soon check that it is enough to only use uniformly continuous functions.
We now show the key estimate.
(N )
B
BT
SK
(N )
(N )
XK/N := √ := √ K = BTK /N ≈ K
N
N
N
where the approximation is the first gap we need to plug. This is the difference of the
two at the endpoints. The second gap is the difference at other parts of the paths.
We plug the first gap. Take T1 , T2 − T1 , T3 − T2 and so on IID with mean 1. Then
TN a.s.
→1
N
by the strong law of large numbers. Then
max |
i=1,...,N
Tk
k
− |→0
N
N
almost surely, from the above (analysis 1).
Let ΩN,δ = {maxi=1,...,N | TNk − Nk | ≥ δ} and then
P(ΩN,δ ) → 0
as N → ∞.
(N )
(N )
We now plug the second gap. Suppose ||Xt − Bt ||∞ > ε. Then there exists a
t ∈ [0, 1] such that
(N )
(N )
|Xt − Bt | ≥ and suppose that t ∈ [K/n, (K + 1)/N ]. Then either
(N )
|Bt
(N )
− BK/N | ≥ ε
or
(N )
|Bt
(N )
− B(K+1)/N | ≥ ε
and so
(N )
P(||Xt
(N )
− Bt
(N )
| ≥ ε for |s − t| ≤ δ + 1/N )
(N )
| ≥ ε for |s − t| ≤ 2δ)
||∞ > ε) ≤ P(ΩN,δ ) + P(|Bs(N ) − Bt
≤ P(ΩN,δ ) + P(|Bs(N ) − Bt
and then choose δ small and N large.
This concludes the proof of Donskers theorem, modulo some other minor tidy ups.
Corollary 4.9
1
Z
0
We apply F (f ) =
Z
0
1
R1
0
(N )
Xt
D
Z
dt →
1
Bt dt
0
f . The former isn’t really very useful, so we rewrite it as follows:
(N )
Xt dt
=
N
−1
X
0
1
2N
S
SK+1
√K + √
N
N
=
1
N 3/2
and so we suspect that
1
N 3/2
N
−1
X
SK → N (0, 1/3)
0
We need a tidy up lemma:
37 of 40
N
−1
X
0
SK −
SN
N 3/2
MA4F7 Brownian motion Lecture Notes Autumn 2012
D
D
D
Lemma 4.10 Suppose that XN → X and YN → 0. Then XN + YN → X.
This
is not true if YN converges to a non zero limit.
(
Z
N even
.
−Z N odd
Proof We consider characteristic functions
For example consider YN =
|E(eiθ(XN +YN ) − E(eiθX )| ≤ |E(eiθXN − eiθX )| + |E(eiθ(XN +YN ) − eiθXN )|
The former tends to zero as XN → X and the latter is as follows
|E(eiθ(XN +YN ) − eiθXN )| ≤ E(|eiθYN − 1|)
q
≤ E(|eiθYN − 1|2 )
q
≤ E(1 − eiθYN − e−iθYN + 1)
→0
Q.E.D.
R1
D
Now back to the problem in hand. We have F (f ) = 0 f and we have X (N ) → B and
then
N −1
SN
1 X
SK − 3/2
F (X (N ) = 3/2
N
N
0
Then
s SN 2
SN N
E 3/2 ≤ E 3/2 = 3 → 0
N
N
N
This uses the following
D
Lemma 4.11 If E|XN | → 0 then XN → 0
and the proof uses
|E(eiθXN − 1)| ≤ θE|XN | → 0
We have another example, with a non continuous F .
Take Ta = inf{N : SN ≥ a} and we guess that
q
D
T√N a N → τa = inf{t : Bt = a}
and we check this. We take
F (f ) = inf{t : f (t) ≥ a} ∧ 1
where the minimum with 1 is for convenience. Then
q
F (X (N ) ) = T√N a N + error
and the error is at most 1/N /. We need only show that
F (X (N ) ) → F (B)
38 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
but the problem is F is not continuous. We thus guess the discontinuity set of F :
Disc(F ) ⊂ {f : τa (f ) < 1, ∃ε > 0 such that f (t) ≤ a for t ∈ [τa , τa + ε]}
and our aim is to show that
P(B ∈ Disc(F )) = 0
If we define Xt = BTa +t − a then this is a new Brownian motion and P(Xt ≤ 0 for t ∈
[0, ε]) = 0 by the time inverted law of the iterated logarithm.
We will check the inclusion of the discontinuity set. We prove the complement statement. Choose δ > 0 and assume that supt≤F (f )−δ f (t) = a − ε for some ε > 0. Then
if
||g − f ||∞ ≤ ε/2
and this implies that F (g) ≥ F (f ) − δ. In other words, if g → f then F (g) ≥ F (f ) and
so we have
lim inf F (g) ≥ F (f )
g→f
which is semicontinuity.
We take an f where there exists SN → F (f ) and f (XN ) > a. We need to show that
F is continuous at such an f .
If f (SN ) = a + ε then take a g such that ||g − f ||∞ < ε/2 and this will have g(SN ) > a.
Thus F (g) ≤ SN and so lim supg→f F (g) ≤ F (f ) and so we are continuous here.
Theorem 4.12 The following are equivalent for XN , X over (E, d).
1. E(F (XN )) → E(F (X)) for all bounded continuous functions F : E → R.
2. E(F (XN )) → E(F (X)) for all bounded uniformly continuous functions F : E → R.
3. lim supN →∞ P(XN ∈ A) ≤ P(X ∈ A) for all closed A.
4. lim inf N →∞ P(XN ∈ A) ≥ P(X ∈ A) for all open A.
5. P(XN ∈ A) → P(X ∈ A) for all A such that P(X ∈ ∂A) = 0, where ∂A = A \ Ao .
6. E(F (XN )) → E(F (X)) for all measurable F such that P(X ∈ Disc(F )) = 0.
Proof 1 =⇒ 2 is immediate.
2 =⇒ 3 Define Fε (x) = 1 −
to χA . Then
d(x,A)
ε
+
and this is uniformly continuous and converges
P(X ∈ A) ← E(Fε (X)) = lim E(Fε (XN )) ≥ lim sup P(XN ∈ A)
N →∞
N →∞
3 =⇒ 4
P(X ∈ A) = 1 − P(X ∈ AC )
and so if A is open then AC is closed.
4 =⇒ 5
P(X ∈ Ao ) ≤ lim inf P(XN ∈ Ao ) ≤ lim inf P(XN ∈ A) ≤ lim sup P(XN ∈ A) ≤ lim sup P(XN ∈ A) ≤ P(X ∈
and so
P(X ∈ A) − P(X ∈ Ao ) = X ∈ ∂A = 0
39 of 40
MA4F7 Brownian motion Lecture Notes Autumn 2012
5 =⇒ 6 We observe that 5 is a special case with f = χA with Discf = ∂A.
Now choose α1 < α2 < ... < αN with |αi+1 − αi | < ε.
We can approximate a general f by
fε (x) =
N
X
αi χf −1 (αi ,αi+1 ]
1
and with |fε (x) − f (x)| ≤ ε. We will check E(fε (XN )) → E(fε (X)), and this is enough.
We then apply part 5 with A = f −1 (αi , αi+1 ] ∪ {x : f (x) = αi } ∪ {x : f (x) = αi+1 }.
We claim that
Disc(χA ) ⊂ Discf ∪ {x : f (x) = αi } ∪ {x : f (x) = αi+1 }
and we prove the complemented statement.
Pick x 6∈ Discf with f (x) 6= αi and f (x) 6= αi+1 . Take xN → x and then as f is
continuous we have f (xN ) → f (x), because χ(f (xN ) ∈ (αi , αi+1 ]) is discontinuous only
at αi and αi+1 .
To apply 5 we need to know that P(X ∈ {x : f (x) = αi }) = 0. To this end let
pα = P(f (X) = α) > 0. Now {α : pα ≥ 1/N } has at most N elements, and so pα > 0 for
only countably many α. Let Q = {α : pα > 0}. Then we need to choose α1 , ..., αN not
lying in Q. As Q is countable this is easy.
6 =⇒ 1 is immediate.
Q.E.D.
5
Up Periscope
Suppose that a box has N balls, half of which are black and the other half are white. We
draw at random until the box is empty. Let SK denote the number of blacks left by draw
(N )
K minus the number of whites by draw K. Let Xt = S√NNT and we expect:
D
Theorem 5.1 X (N ) → Brownian Bridge.
We consider a population model. Let SN be the population size at time N . Assume
that there is a one half probability of one individual having 2 or zero offspring each time
step.
Lemma 5.2
P(SN ) > 0|S0 = 1) ∼
2
N
Corollary 5.3
P(SN > 0|S0 = N ) ∼ 1 − (1 −
2 N
) ∼ 1 − e−2
N
(N )
If we instead choose S0 = N and linearly interpolate to get Xt
Theorem 5.4
D
X (N ) → X
where X solves Feller’s equation:
√ dB
dX
= X
dt
dt
40 of 40
=
SN t
N
we get
MA4F7 Brownian motion Lecture Notes Autumn 2012
Such theorems are called diffusion approximations. One can split the proof into two
parts:
1. Soft nonsense- Show there exists a convergent subsequence X (Nl )
2. Characterise the limit points of X (N
The first point is to do with compactness.
Theorem 5.5 (Kolmogorov) Suppose X (N ) is a continuous process and
(N )
E|Xt
− Xs(N ) | ≤ C|t − s|1+ε
for all s, t ≤ 1 and N ≥ 1. Then X (N ) has a convergent subsequence.
We can deduce this from compact sets K ⊂ C[0, 1]. For example
{f : |f (t) − f (s)| ≤ C1 |t − s|α , |f (0)| ≤ C2 }
is compact.
The second point is specific to each convergence. We look only at the population one.
How do we characterise Feller’s diffusion. We would want
p
Xt+∆t − Xt ≈ Xt B(0, ∆t)
The key estimate is the following:
(N )
(N )
X K+1 − X K
N
N
=
SK+1 SK
−
N
N
and
√ the bit on the right hand side has mean zero and variance SK , which agrees with
SK B(0, ∆t).
41 of 40
Download