2. Independent random variables. If {X The Law of large Numbers.

advertisement
2. Independent random variables.
The Law of large Numbers. If {Xi : i ≥ 1} are a sequence of independent identically
distributed random variables (on some (Ω, F , P )) with E[Xi] = m, then
1
P ω : lim [X1 (ω) + · · · + Xn (ω)] = m = 1
n→∞ n
This is the strong law of large numbers. The weak law, which is naturally weaker than the
strong law, asserts that for any ǫ > 0,
1
lim P ω : [X1 (ω) + · · · + Xn (ω)] − m ≥ ǫ = 0
n→∞
n
The strong law requires the existence of the mean, i.e {Xi } have to be integrable. On
the other hand the weak law can be valid some times even if we can only have the mean
defined as
Z a
m = lim
a→∞
x dF (x)
−a
where F is the common distribution of {Xi }.
The Central Limit Theorem. If {Xi } are independent, identically distributed and have
mean 0 and variance σ 2 , then the distribution of Zn = √1n [X1 + · · · + Xn ], approaches the
normal distribution with mean , i.e.
Z x
y2
1
1
e− 2σ2 dy
lim P [ √ [X1 + · · · + Xn ] ≤ x] = √
n→∞
n
2πσ −∞
The central limit theorem is valid also if {Xi } are not identically distributed,Pbut Xi has
n
mean 0 and variance σi2 . Then with Zn = s1n [X1 + · · · + Xn ], where s2n = i=1 σi2 , we
have
Z x
y2
1
e− 2 dy
lim P [Zn ≤ x] = √
n→∞
2π −∞
provided sn → ∞ and some additional condition known as Lindeberg’s condition is satisfied. It is that, for any ǫ > 0,
n Z
1 X
x2 dFi (x) = 0
lim
n→∞ s2
n i=1 |x|≥ǫsn
where Fi is the distribution of Xi .
Kolmogorov’s Two Series Theorem. Let {Xi } be a sequence of independent random
variables uniformly bounded by a constant C with Xi having mean 0 and variance σi2 .
Then the necessary and sufficient condition for the convergence with probability 1 of the
series
∞
X
Xi
S=
i=1
1
is
∞
X
i=1
σi2 < ∞
The sufficiency part of the proof depends on the important inequality known as Kolmogorov’s inequality.
Kolmogorov’s Inequality. Let {Xi } be indpendent random variables with mean 0 and
variance σi2 . Then
n
1 X 2
σi
P [ sup |X1 + X2 + · · · + Xj | ≥ ℓ] ≤ 2
ℓ
1≤j≤n
I=1
Tchebecheff’s inequality gives
j
1 X 2
P [|X1 + X2 + · · · + Xj | ≥ ℓ] ≤ 2
σi
ℓ
I=1
Naive calculation will yield
j
n
1 XX 2
σi
P [ sup |X1 + X2 + · · · + Xj | ≥ ℓ] ≤ 2
ℓ j=1
1≤j≤n
I=1
which is not good enough. Let us define
Ej = {ω : |X1 + X2 + · · · + Xk | < ℓ for k ≤ (j − 1),
|X1 + X2 + · · · + Xj | ≥ ℓ}
We are interested in estimating
P (E) =
n
X
P (Ej )
j=1
Denoting by Sj =
Pj
i=1
Z
Xi ,
Sn2 dP
=
Z
=
Z
≥
Z
Ej
Ej
(Sj + (Sn − Sj ))2 dP
Ej
[Sj2 + 2Sj (Sn − Sj ) + (Sn − Sj )2 ] dP
Ej
Sj2 dP ≥ ℓ2 P (Ej )
The cross term is 0 because Ej , Sj are independent of Sn − Sj . Because Ej are disjoint
and their union is E, summing over j = 1, 2, . . . , n, we get
Z
2
sn ≥
s2n dP ≥ ℓ2 P (E)
E
2
This allows the tail of the series
Tn = sup |Sj − Sn |
j≥n
to be estimated by
∞
1 X 2
σ
P [Tn ≥ ǫ] ≤ 2
ǫ j=n+1 j
proving Tn → 0 with probability 1. The converse depends on an inequality as well. Assume
lim sup |Sn | < ∞ with positive probbaility. Let
Fn = {ω : sup |Sj | ≤ ℓ}
1≤j≤n
so that En+1 = Fn − Fn+1 are disjoint. Then there is an ℓ such that P (Fn ) ≥ δ > 0 for all
Z
Z
Z
Z
Z
2
2
2
2
Sn+1 dP −
Sn2 dP
Sn dP =
Sn+1 dP +
Sn+1 dP −
Fn
F
Fn+1
En+1
Fn
Z n
2
]dP
=
[2Xn+1 Sn + Xn+1
Fn
Z
2
≥
Xn+1
dP
Fn
2
= P (Fn )σn+1
2
≥ δ σn+1
On the other hand |Sn+1 | ≤ C + ℓ on En+1 . Therefore
Z
Z
1
2
2
2
σn+1 ≤ [(C + ℓ) P (En+1 ) +
Sn+1 dP −
Sn2 dP ]
δ
En+1
Fn
And the telescoping sum
XZ
[
2
Sn+1
dP
En+1
−
Z
Sn2 dP ]
Fn
is bounded by ℓ2 . Providing the estimate
X
j
σj2 ≤
1
[(C + ℓ)2 + ℓ2 ]
δ
Actually we have shown some thing stronger. If
X
σj2 = +∞
j
3
then
P [lim sup |Sn | = ∞] = 1
n→∞
An easy corollary is that if {Xi } are independent random variables with E[Xi] = 0 and
E[Xi2 ] = σi2 then for any sequence of constants {aj } satisfying
X
j
σj2 a2j < ∞
the series
S=
X
j
will converge with probability 1,
4
aj X j
Download