Technical Complements to the paper

advertisement
Technical Complements to the paper
Shared Uncertainty in Measurement Error Problems, With Application to
Nevada Test Site Fallout Data
by Yehua Li, Annamaria Guolo, F. Owen Hoffman and Raymond J. Carroll
Web Appendix A: Sketch of Technical Arguments
A.1
Computation of Σ−1
s (ρs , γC )
As is easily seen,
Σs (ρs , γC ) = (1 − γC ){(1 − ρs )Ds + ρs vs vsT };
2
Ds = diag(σ1s,total
, ..., σn2 s s,total );
vs = (σ1s,total , ..., σns s,total )T ;
Σ−1
s (ρs , γC )
(
Ds−1 vs vsT Ds−1
ρs
(1 − ρs )
−
= (1 − γC )
(1 − ρs )2 1 + ρs (1 − ρs )−1 vsT Ds−1 vs
)
(
ρs
−1
T −1
−1
−1
−1
D v s v s Ds ;
Ds −
= (1 − γC ) (1 − ρs )
1 + (ns − 1)ρs s
−1
−1
Ds−1
)
|Σs (ρs , γC )| = (1 − γC )ns |(1 − ρs )Ds | × |1 + ρs (1 − ρs )−1 vsT Ds−1 vs |
= (1 − γC )ns (1 − ρs )ns −1 {1 + (ns − 1)ρs }
ns
Y
2
σis,total
.
i=1
A.2
Complete Conditionals for Gibbs Sampler
We will use Gibbs sampler to draw samples from the posterior distributions of Xis , Lis , ρs ,
γC , β, µs , σs2 , θ. First we need to derive the complete conditional distributions for these
variables and decide the sampling schemes from these complete conditionals. For those
variables whose complete conditional distribution can not be easily generated, we will use a
Metropolis step to update them. This algorithm is also referred as Metropolis-within-Gibbs
algorithm.
• If we impose prior distribution Normal(0, σµ2 s ) to µs , then
ns
X
σs2 σµ2 s
σµ2 s
Lis , 2
π(µs |rest) = Normal
σs2 + ns σµ2 s i=1
σs + ns σµ2 s
!
.
• Let π(σs2 ) be an Inverse Gamma distribution, i.e., π(σs2 ) ∝ (σs2 )−(as +1) exp{−1/(bs σs2 )},
then

(
π(σs2 |rest) = IG as + ns /2, (1/2)
ns
X
(Lis − µs )2 + b−1
s
i=1
)−1 
.
• Let 1ns be a vector of ns 1’s, then the complete conditional for Ls = (L1s , · · · , Lns s )T
is
(
1
1
(Ls − µs 1ns )T (Ls − µs 1ns ) −
(Ls − Ws )T Ds−1 (Ls − Ws )
2
2σs
2γC
1
T −1
− (Ls − Xs ) Σs (ρs , γC )(Ls − Xs )
2
1 T −2
∝ exp − Ls {σs I + γC−1 Ds−1 + Σ−1
s (ρs , γC )}Ls
2
π(Ls |rest) ∝ exp −
−2
−1 −1
−1
+LT
s {µs σs 1ns + γC Ds Ws + Σs (ρs , γC )Xs }
−1
= Normal(Ω−1
s hs , Ωs ),
(A.1)
−1 −1
−2
−1
where Ωs = σs−2 I + γC−1 Ds−1 + Σ−1
s (ρs , γC ), hs = µs σs 1ns + γC Ds Ws + Σs (ρs , γC )Xs .
• We let the prior for β to be Normal(0, σβ2 I), then the complete conditional for β is
π(β|rest) ∝
T
Y (exp[Zis
β + log{1 + θ exp(Xis )}])Yis
i,s
(
)
1
exp − 2 β T β .
T
1 + exp[Zis β + log{1 + θ exp(Xis )}]
2σβ
We use a Metropolis step to update β. Denote the current value of β as βcurr , then
generate a candidate value of β by kernel q(βcurr , βnew ) = Normal(βcurr , 0.01I). The
new value is accepted with probability
(
)
π(β = βnew |rest)
α = min 1,
.
π(β = βcurr |rest)
• Complete conditional for γC :
−N/2
π(γC |rest) ∝ I(ac ≤ γC ≤ bc )γC


exp −
1 X
2γC
i,s


2
(Wis − Lis )2 /σis,total

×
Y
−1/2
|Σs (ρs , γC )|
s
(
)
1X
(Xs − Ls )T Σ−1
× exp −
s (ρs , γC )(Xs − Ls )
2 s




1 X
−N/2
2
(Wis − Lis )2 /σis,total
∝ I(ac ≤ γC ≤ bc )γC (1 − γC )−N/2 exp −

2γC i,s
"
#
X
1
(Xs − Ls )T {(1 − ρs )Ds + ρs vs vsT }−1 (Xs − Ls ) .
× exp −
2(1 − γC ) s
Again, we use a Metropolis step to update γC . A new value γC,new is generated from
Uniform[ac , bc ]. We accept the new value with probability
(
)
π(γC = γC,new |rest)
α = min 1,
.
π(γC = γC,curr |rest)
• Complete conditional for ρs :
1
π(ρs |rest) ∝ I(cs ≤ ρs ≤ ds )|Σs (ρs , γC )|−1/2 exp − (Xs − Ls )T Σ−1
s (ρs , γC )(Xs − Ls )
2
"
1
−(ns −1)/2
−1/2
∝ I(cs ≤ ρs ≤ ds )(1 − ρs )
{1 + (ns − 1)ρs }
exp − (1 − γC )−1
2
)
#
(
ρs
−1
T −1
−1
−1
T
D vs vs Ds (Xs − Ls ) .
×(1 − ρs ) (Xs − Ls ) Ds −
1 + (ns − 1)ρs s
We use a Metropolis step similar to that of γC to update ρs . The candidate value is
generated from Uniform[cs , ds ].
• Complete conditional for θ:
π(θ|rest) ∝ π(θ)
Y
(H[ZisT β + log{1 + θ exp(Xis )}])Yis
i,s
×(1 − H[ZisT β + log{1 + θ exp(Xis )}])1−Yis .
Again, θ is updated by a Metropolis step similar to above. We consider the following
two schemes.
(θ.1) For the truncated normal prior, we use the prior π(θ) as the proposal density. A
new value θnew is generated from π(·), and then it is determine to be accepted with
probability
(
)
Pθ (θnew )
α = min 1,
,
Pθ (θcurr )
where Pθ (t) =
Q
T
Yis
(1−H[ZisT β+log{1+t exp(Xis )}])1−Yis .
i,s (H[Zis β+log{1+t exp(Xis )}])
In the analysis, we used proper priors for θ, such as the Truncated Normal priors.
(θ.2) For the approximate Jeffreys prior, we used a transition kernel q(θcurr , θnew ) =
T N (θcurr , σq2 ), where T N (µ, σ 2 ) is the truncated normal distribution with density function I(x ≥ 0)φ{(x − µ)/σ}/Φ(µ/σ), φ and Φ are the density and distribution function
of a standard normal random variable. Then the new value θnew is accepted with
probability
(
)
π(θ = θnew |rest)/Φ(θnew /σq )
α = min 1,
.
π(θ = θcurr |rest)/Φ(θcurr /σq )
A.3
Method of Scoring for EM Algorithm
The method of scoring in Section 4.2 is
(k+1)
b
Θ
1
(k)
(k)
(k)
−1
b
=Θ
1 − A (Θ1 )B(Θ1 ),
T
b
b
where A(Θ1 ) = ∂ 2 Q/∂Θ
1 ∂Θ1 and B(Θ1 ) = ∂ Q/∂Θ1 , these terms given as
(b)
(b)
Pis (β, δ, Xis ) = H[ZisT β + log{1 + exp(δ + Xis )}],
b
XX
∂Q
(b)
= B −1
{Yis − Pis (β, δ, Xis )}Zis ,
∂β
b s,i
(b)
b
XX
∂Q
exp(δ + Xis )
(b)
,
= B −1
{Yis − Pis (β, δ, Xis )}
(b)
∂δ
1 + exp(δ + Xis )
b s,i
b
XX
∂2Q
(b)
(b)
−1
=
−B
Pis (β, δ, Xis ){1 − Pis (β, δ, Xis )}Zis ZisT ,
T
∂β∂β
b s,i
(b)
b
XX
exp(δ + Xis )
∂2Q
(b)
(b)
= −B −1
Pis (β, δ, Xis ){1 − Pis (β, δ, Xis )}
Zis ,
(b)
∂β∂δ
1 + exp(δ + Xis )
b s,i

(b)
2
b
 exp(δ + X ) 
XX
∂ 2Q
(b)
(b)
is
−1
=
−B
P
(β,
δ,
X
){1
−
P
(β,
δ,
X
)}
.
is
is
is
is
(b) 
2

∂δ
1
+
exp(δ
+
X
)
s,i
b
is
A.4
Formulae For Computing (8)
Let Pis (β, δ, x) = H[ZisT β + log{1 + exp(δ + x)}]. Then
X
∂ log f1
=
{Yis − Pis (β, δ, Xis )}Zis ,
∂β
s,i
X
∂ log f1
exp(δ + Xis )
=
,
{Yis − Pis (β, δ, Xis )}
∂δ
1 + exp(δ + Xis )
s,i
X
∂ log f1
= (σs2 )−1 (Lis − µs ),
∂µs
i
X
∂ log f1
= (2σs4 )−1 (Lis − µs )2 − ns /(2σs2 ),
2
∂σs
i
−
−
−
X
∂ 2 log f1
=
P (β, δ, Xis ){1 − P (β, δ, Xis )}Zis ZisT ,
T
∂β∂β
s,i
X
∂ 2 log f1
exp(δ + Xis )
=
Zis ,
P (β, δ, Xis ){1 − P (β, δ, Xis )}
∂β∂δ
1 + exp(δ + Xis )
s,i
X
∂ 2 log f1
exp(δ + Xis )
=
−
{Yis − Pis (β, δ, Xis )}
2
∂δ
{1 + exp(δ + Xis )}2
s,i
+
X
s,i
exp(δ + Xis )
P (β, δ, Xis ){1 − P (β, δ, Xis )}
1 + exp(δ + Xis )
∂ 2 log f1
= I(s = s′ )ns /σs2 ,
∂µs ∂µs′
(
)
X
∂ 2 log f1
′
−6
2
4
(Lis − µs ) − ns /(2σs ) ,
− 2 2 = I(s = s ) σs
∂σs ∂σs′
i
−
(
(
)
X
∂ 2 log f1
−
= I(s = s′ ) σs−4 (Lis − µs ) .
2
∂µs ∂σs′
i
)2
,
Web Figures W.1 and W.2
30
0
10
θ
20
0
e+00
2
e+04
4
e+04
6
e+04
8
e+04
1
e+05
e+04
8
e+04
1
e+05
1.0
1.5
β1
2.0
2.5
Iteration
0
e+00
2
e+04
4
e+04
6
Interation
Figure W.1: History for the excess relative risk parameter θ per gray and the gender effect
β1 for thyroiditis, in the case of shared Berkson uncertainties and using the approximate
Jeffreys prior.
300
250
200
0
50
100
θ
150
0e+00
2e+04
4e+04
6e+04
8e+04
1e+05
6e+04
8e+04
1e+05
2
0
1
β1
3
4
Iteration
0e+00
2e+04
4e+04
Iteration
Figure W.2: History for the excess relative risk parameter θ per gray for neoplasm, in the
case of shared Berkson uncertainties and using the approximate Jeffreys prior.
Download