Proof of the Halmos-Savage Factorization Theorem (The-

advertisement
Proof of the Halmos-Savage Factorization Theorem (Theorem 3 of the Stat 643 Summary).
We’ll take Theorem 1 as given, prove Theorem 3 given Theorems 1 and 2,
and then prove Theorem 2.
So first assume that Theorems 1 and 2 have been established.
Suppose that T is sufficient. Pθ ¿ λ ¿ µ implies that
dPθ dλ
dPθ
=
·
a.e. µ
dµ
dλ dµ
(and both functions on the right of this equality are non-negative). By Theorem
2, ∃ a nonnegative function gθ such that
dPθ
= gθ ◦ T a.s. λ
dλ
I claim that as measures on X ∗ , λ and µ are
Let X ∗ = {x| dλ
dµ (x) > 0}.
equivalent. Clearly, λ ¿ µ on X ∗ . So suppose that B is a B-measurable subset
of X ∗ that has λ (B) = 0. But
Z
dλ
dµ
λ (B) =
B dµ
while since B ⊂ X ∗ the integrand here is positive. So it must be that µ (B) = 0.
Thus λ and µ are equivalent on X ∗ .
So then consider
¸
·
dPθ
dλ dPθ
dλ
∆=
− (gθ ◦ T ) ·
=
− gθ ◦ T
dµ
dµ
dµ dλ
Letting
dPθ
(x) 6= gθ (T (x))}
dλ
since the second term in this intersection has λ probability 0, µ (D) = 0 and
thus
dλ
dPθ
= (gθ ◦ T ) ·
a.e. µ
dµ
dµ
D = {x| ∆ (x) 6= 0} = X ∗ ∩ {x|
and taking h =
dλ
the forward implication is proved.
dµ
Conversely, suppose that the factorization holds, that is that fθ =
dPθ
dµ = (gθ ◦ T ) · h a.e. µ.
1
Now
dλ
dµ
=
∞
X
i=1
=
∞
X
i=1
= h·
for k =
ci
dPθi
dµ
ci h · (gθi ◦ T )
∞
X
i=1
ci (gθi ◦ T )
= h · (k ◦ T )
P∞
i=1 ci gθi .
Then the fact that Pθ ¿ λ ¿ µ again implies that
dPθ dλ
dPθ
=
·
a.e. µ
dµ
dλ dµ
so that
(gθ ◦ T ) · h =
Then define
dPθ
· h · (k ◦ T ) a.e. µ.
dλ

 gθ ◦ T
gθ∗ ◦ T =
k◦T
 0
On X ∗
dλ
>0
dµ
otherwise.
if
dPθ
= gθ∗ ◦ T
dλ
and this set has λ probability 1, i.e.
dPθ
= gθ∗ ◦ T a.e. λ
dλ
so that Theorem 2 then implies that T is sufficient.
So Theorem 3 has been proved (conditional on Theorems 1 and 2).
Now consider the proof of Theorem 2 (assuming the truth of Theorem
1).
Suppose that T is sufficient. Let B ∈ B and note that by sufficiency ∃ a
B0 -measurable function P (B|B0 ) such that for all θ
Eθ (IB |B0 ) = Pθ (B|B0 ) = P (B|B0 ) a.s. Pθ
Then for B0 ∈ B0
Z
B0
P (B|B0 ) dPθ = Pθ (B ∩ B0 ) for all θ ∈ Θ
2
(1)
by the definition of conditional expectation. Then since λ =
Z
P (B|B0 ) dλ = λ (B ∩ B0 )
P
ci Pθi ,
B0
and so (again by the definition of conditional expectation)
Eλ (IB |B0 ) = P (B|B0 ) a.s. λ
and thus
Eλ (IB |B0 ) = P (B|B0 ) a.s. Pθ for all θ
(2)
(that is, this object serves as the conditional expectation wrt each Pθ and wrt
λ).
Now consider Pθ0 and λ0 , respectively the restrictions of Pθ and λ to B0 and
dP 0
the R-N derivative dλθ0 on (X , B0 ). By Lehmann’s Theorem, ∃ non-negative
F-measurable functions gθ such that
dPθ0
= gθ ◦ T
dλ0
(3)
We need to show that gθ ◦ T also serves as an R-N derivative of Pθ wrt λ. So
let B ∈ B.
Z
Pθ (B) =
P (B|B0 ) dPθ
(by fact (1))
ZX
Eλ (IB |B0 ) dPθ
(by fact (2))
=
X
Z
Eλ (IB |B0 ) dPθ0
(since Eλ (IB |B0 ) is B0 -measurable)
=
X
Z
Eλ (IB |B0 ) gθ ◦ T dλ0 (by relationship (3))
=
X
Z
Eλ (IB |B0 ) gθ ◦ T dλ
=
X
Z
Eλ (IB · gθ ◦ T |B0 ) dλ (since gθ ◦ T is B0 -measurable)
=
ZX
IB · gθ ◦ T dλ (X ∈ B0 and the definition of conditional expectation)
=
ZX
gθ ◦ T dλ
=
B
That is, gθ ◦ T works as the R-N derivative of Pθ wrt λ and the first half of the
theorem is proved.
So conversely, suppose that ∃ non-negative F -measurable functions gθ
such that
dPθ
= gθ ◦ T a.s. λ
dλ
3
Since gθ ◦ T is B0 -measurable,
dPθ0
= gθ ◦ T a.s. λ0
dλ0
(4)
To show that T is sufficient, it will be enough to show that for any B ∈ B
Eθ (IB |B0 ) = Eλ (IB |B0 ) a.s. Pθ for all θ
So for fixed θ and B ∈ B, define a measure ν on (X , B) by
Z
IB dPθ = Pθ (C ∩ B)
ν (C) =
C
Now ν ¿ Pθ and
dν
= IB a.s. Pθ
dPθ
For all B0 ∈ B0 ,
Z
IB dPθ
ν (B0 ) =
B0
Z
=
Eθ (IB |B0 ) dPθ
(5)
(by the defintion of conditional expectation)
B0
So for ν 0 the restriction of ν to B0
dν 0
= Eθ (IB |B0 ) a.s. Pθ0
dPθ0
(since Eθ (IB |B0 ) is B0 -measurable and has the right integrals).
ν 0 ¿ Pθ0 ¿ λ0
dν 0 dPθ0
dν 0
=
·
a.s. λ0
dPθ0 dλ0
dλ0
(6)
Then since
So since Pθ0 ¿ λ0 and facts (4) and (6) hold
dν 0
= Eθ (IB |B0 ) · gθ ◦ T a.s. Pθ0
dλ0
On the other hand, from (5), we still have
dν
dPθ
= IB a.s. Pθ , so that
dPθ
dν
= IB ·
= IB · gθ ◦ T a.s. Pθ
dλ
dλ
This then implies by the definition of conditional expectation that
dν 0
= Eλ (IB · gθ ◦ T |B0 ) a.s. Pθ0
dλ0
4
(7)
(The object on the right isR obviously B0 -measurable, and if we multiply by IB0
and integrate dλ0 , we get IB∩B0 gθ ◦ T dλ = ν (B0 ).) So
dν 0
= Eλ (IB |B0 ) · gθ ◦ T a.s. Pθ0
dλ0
(8)
So from (7) and (8)
Eθ (IB |B0 ) · gθ ◦ T = Eλ (IB |B0 ) · gθ ◦ T a.s. Pθ0
This is enough to imply that
Eθ (IB |B0 ) = Eλ (IB |B0 ) a.s. Pθ0
(9)
Why? Let A = {x| gθ (T (x)) = 0} and note that
Z
gθ ◦ T dλ = 0
Pθ (A) =
A
So X − A has Pθ probability 1 and so does {x| Eθ (IB |B0 ) · gθ ◦ T =Eλ (IB |B0 ) ·
gθ ◦ T } ∩ (X − A). And on this set, Eθ (IB |B0 ) =Eλ (IB |B0 ).
So (9) holds and Theorem 2 has been proved (conditional on Theorem 1).
5
Download