An alternative hypothesis ( ) hypothesis

advertisement
5. Tests of hypotheses and
condence intervals
An alternative hypothesis (Ha)
gives a set of alternaties to the null
hypothesis
Consider the linear model with
E (Y) = X and V ar(Y) = We may test
H0 : C = d vs Ha : C 6= d
where
C is an m k matrix of constants
d is an m 1 vector of constants
This can also be expressed as
Y = X + where E () = 0 and V ar() = .
Typical null hypothesis (H0)
species the values for one or more elements of species the values for some linear functions of the elements of The null hypothesis is rejected if it is shown to
be suciently incompatible with the observed
data.
288
287
Failing to reject H0 is not the same as proving
H0 is true.
too little data to accurately estimate
C
relatively large variation in (or Y)
if H0 : C = d is false, C ; d may be
\small"
You can never be completely sure that you
made the correct decision
Type I error (signicance level)
Type II error
289
Basic considerations in specifying a null
hypothesis H0 : C = d
(i) C should be estimable
(ii) Inconsistencies should be avoided,
i.e., C = d should be a consistent
set of equations
(iii) Redundancies should be eliminated,
i.e., in C = d we should have
rank(C ) = number of rows in C
290
Example 5.1 Eects model from Example 3.2
Yij = + i + ij i = 1; 2; 3
j = 1; : : : ; ni
In this case
2Y
66 Y
66 Y
64 Y
Y
11
12
21
31
32
Y33
3 21
77 66 1
77 = 66 1
75 64 1
1
1
1
0
0
0
1 0
0
0
1
0
0
0
0
0
0
1
1
1
3
77
77
75
2
3 6 75 + 666 64 2 64 11
12
1
21
2
31
3
33
32
3
77
77
75
291
+ 1 = cT H0 : + 1 = 60 seconds
against
HA : + 1 6= 60 seconds
(two-sided alternative)
Or we can test
H0 : + 1 = 60 seconds
against
HA : = 1 < 60 seconds
(one-sided alternative)
By denition
E (Yij ) = + i is estimable.
In this case
We can test
213
6 7
where c = 664 10 775
0
Note that this quantity is estimable, e.g.,
0 0 0 0)Y):
cT = + 1 = E (( 12 21
Then, any solution b = (X T X );X T Y to the
normal equations X T X b = X T Y yields the
same value cT b and it is the unique blue for
cT .
We will reject H0 : cT = 60 if
cT b = cT (X T X );X T Y
is too far away from 60.
293
292
(ii) Dierence between the mean response for
two treatments is estimable
1 ; 3 = ( + 1) ; ( + 3)
= 12 21 0 ;31 ;31 ;31 E (Y)
ans we can test
H0 : 1 ; 3 = 0 vs. HA : 1 ; 3 6= 0
The unique blue for
1 ; 3 = (0 1 0 ; 1) = cT is
cT b for any b = (X T X );X T Y
Reject H0 : 1 ; 3 = cT = 0 if cT b is too
far from 0.
294
In Example 3.2 (on Page 3.7) we found several
solutions to the normal equations:
(iii) It would not make much sense to attempt to test
H0 : 1 = 3 vs. HA : 1 6= 3
because 1 = [0 1 0 0] = cT is not
estimable
Although E (Y1j ) = + 1 neither nor 1
has a clear interpretation.
Dierent solutions to the normal equations
produce dierent values for
^1 = CT b = CT (X T X );X T Y
The data alone do not provide information
about whether or not 1 = 3.
2
66 00
b = 66
40
0 0
1 0
2
0 11
0 0 0
0 37 2 Y:: 3 2 0 3
0 77 666 Y1: 777 = 666 61 777
0 75 4 Y2: 5 4 71 5
1
Y3:
60
3
2 1 ;1 ;1 0 3 2 Y 3 2 69 3
::
1 6 ;1 2:5 1 0 777 666 Y1: 777 = 666 ;8 777
b = 664
3 ;1 1 4 0 5 4 Y2: 5 4 2 5
0 0 0 0 Y3:
0
2 2 ;1 ;1
66 61 61 6
0
6;
b = 66 16 2 1
64 ; 6 0 1
; 16 0
;1 3 2
6 7 Y::
0 777 66 Y1:
6
0 775 4 Y2:
0 31 Y3:
295
2
3
0 1 0 ;1 7
6
(iv) For C = 4 1 1 0 0 5
1 0 0 1
consider testing
2 3
2 3
;3 7
;3
6
H0 : C = 4 60 5 vs. HA : C 6= 64 60 75
70
70
In this case C is estimable, but
there is an inconsistancy. If the null
hypothesis is true,
2
3 2 3
;
;3
1
3
C = 64 + 1 75 = 64 60 75
70
+ 3
Then + 1 = 60 and + 3 = 70
implies
(1 ; 3) = ( + 1) ; ( + 3)
= 60 ; 70 = ;10
Such inconsistancies should be avoided.
297
3 2 66:66 3
77 666 ;5:66 777
75 = 66
7
4 4:33 75
2:33
296
2
3
0 1 0 ;1 7
6
(v) For C = 4 1 1 0 0 5
1 0 0 1
consider testing
2
3
2
3
;10 7
;10 7
6
6
H0 : C = 4 60 5 vs. HA : C 6= 4 60 5
70
70
In this case C is estimable and the equations
specied by the null hypothesis are consistent.
There is a redundancy
[1 1 0 0] = + 1 = 60
[1 0 0 1] = + 3 = 70
imply that
[0 1 0 ; 1] = 1 ; 3
= ( + 1) ; ( + 3)
= 60 ; 70 = ;10
298
The rows of C are not linearly independent,
i.e., rank(C ) < number of rows in C .
There are many equivalent ways to remove a
redundancy:
"
#
" #
H0 : 11 10 00 01 = 60
70
"
#
"
H0 : 01 11 00 ;01 = ;10
60
"
#
"
H0 : 01 10 00 ;11 = ;10
70
#
In each case:
The two rows of C are linearly independent
and
rank(C ) = 2
= number of rows in C
The two rows of C are a basis for the
same 2-dimensional subspace of R4.
This is the 2-dimensional space spanned by
the rows of
2
3
0
1
0
;
1
C = 64 1 1 0 0 75
1 0 0 1
#
"
#
"
#
50
H0 : 12 21 00 ;11 = 130
are all equivalent.
We will only consider null hypotheses of the
form H0 : C = d where rank(C ) = number of
rows in C .
299
This leads to the following concept of a
\testable" hypothesis.
Defn 5.1: Consider a linear model E (Y) = X where V ar(Y) = and X is an n k matrix.
For an m k matrix of constants C and an
m 1 vector of constants d, we will say that
H0 : C = d
is testable if
(i) C is estimable
(ii) rank(C ) = m = number of rows in C
301
300
To test H0 : C = d
(i) Use the data to estimate C .
(ii) Reject H0 : C = d if the estimate of C is to far away from d.
How much of the deviation of the
estimate of C from d can be attributed to random errors?
{ measurement error
{ sampling variation
Need a probability distribution for
the estimate of C Need a probability distribution for a
test statistic
302
Normal theory Gauss-Markov model
2 3
Y1
Y = 64 .. 75 N (X ; 2I )
Yn
Result 5.1.
A least squares estimate b for minimizes
(Y ; X b)T (Y ; X b)
For a testable null hypothesis
H0 : C = d
the OLS estimator for C is
C b = C ( X T X ); X T Y
(i) Since C is estimable, C b is invariant to
the choice of (X T X );. (Result 3.10).
(ii) C b is the unique b.l.u.e. for C . (Result
3.11).
Any solution to the normal equations
(X T X )b = X T Y
has the form
b = (X T X );X T Y
303
(iii)
E (C b) ; d) = C ; d
V ar(C b ; d) = V ar(C b)
= 2 C ( X T X ); C T
This follows from V ar(Y) = 2I , because
V ar(C b) = V ar(C (X T X );X T Y)
= C (X T X );X T V ar(Y) X [(X T X );]T C T
= C (X T X );X T (2I )X [(C T X );]T C T
= 2C (X T X );X T X [(X T X );]T C T
Since C is estimable, C = AX for some A and
V ar(C b) = 2AX (X T X );X T [X (X T X );X T ]T AT
= 2AX (X T X );X T AT
= 2C (X T X );C T
305
304
(iv) C b ; d N (C ; d, 2C (X T X );C T )
This follows from the normal theory
Gauss-Markov assumption
Y N (X ; 2I );
property (iii) and Result 4.1.
(v) When H0 : C = d is true,
C b ; d N (0; 2C (X T X );C T )
(vi) Dene
SSH0 = (C b ; d)T [C (X T X );C T ];1(C b ; d)
then
1
2 2
2 SSH0 m( )
where
2 = 12 (C ; d)T [C (X T X );C T ];1(C ; d)
306
This follows from Result 4.7 using
C b ; d N (C ; d; 2C (X T X );C T )
A = 12 [C (X T X );C T ];1
= V ar(C b ; d) = 2C (X T X );C T
Clearly, A = I is idempodent.
We also need the estimability of C and
rank(C ) = m = number of rows in C
to ensure that C (X T X );C T is positive denite
and (C (X T X );C T );1 exists. Then,
d.f. = rank(A) = rank(C (X T X );C T )
=m
Since C (X T X );C T is positive denite, we have
2 = 212 (C ; d)T [C (X T X );C T ];1(C ; d)
> 0
unless C ; d = 0. Hence 2 = 0 if and only if
H0 : C = d is true.
Consequently, from Result 4.7, we have
(vii) 12 SSH0 2m
if and only if
H0 : C ; d is true.
308
307
Then,
SSresiduals = eT e
= YT (I ; PX )Y
is invariant to the choice of (X T X ); used to
obtain PX = X (X T X );X T and b.
To obtain an estimate of
V ar(C b ; d) = 2C (X T X );C T
we need an estimate of 2.
(viii)
Since E (Y) = X is estimable,
^ = X b = X (X T X );X T Y = PX Y
Y
is the unique b.l.u.e. for X .
where
Consequently, the residual vector
^ = (I ; PX )Y
e=Y;Y
is invariant to the choice of (X T X ); used to
obtain PX = X (X T X );X:T
and
E (SSresiduals) = (n ; k)2
k = rank(X) = rank(PX)
n ; k = rank(I ; PX)
and it follows that
MSresiduals = SSnresiduals
;k
is an unbiased estimator of 2.
309
310
Result (viii) is obtained by applying Result 4.7
to
SSresidual = YT(I ; PX)Y
(ix) 12 SSresiduals 2n;k
To show this use the assumption that
Y N (X ; 2I ) and apply Result 4.7 to
1
T 1
2 SSresiduals = Y 2 (I ; PX ) Y
E (YT AY) =
=
T A + tr(A)
T X T (I ; PX )X + tr((I ; PX )2I )
"
this is a zero matrix
= 2tr(I ; PX )
= 2(n ; k)
where k = rank(X).
This used the assumption of a Gauss-Markov
model, but does not use the normality assumption.
%
E (Y) = X = V ar(Y) = 2I = Clearly A = 12 (I ; PX )2I = I ; PX is idempodent and the noncentrality parameter is
2 = T A
= 12 T X T (I ; PX )X = 0
%
this is a matrix
of zeros
312
311
(x) SSH0 and SSresiduals are independently
distributed.
To show this note that SSH0 is a
function of
Then, by Result 4.4, C b and e are independent
because
Cov(C b; e) =
=
=
=
C b = C (X T X );X T Y
and SSresiduals is a function of
e = (I ; X )Y
By Result 4.1,
" # "
#
C b = C (X T X ); X T Y
e
I ; PX
has a multivariate normal distribution
because Y N (X ; 2I ).
313
this is A
COV (C (X T X );X T Y; (I ; PX )Y)
C (X T X );X T (V ar(Y))(I ; PX )T
C (X T X );X T (2I )(I ; PX )
2C (X T X ); X T (I ; PX ) = 0
%
This is a matrix of
zeros since it is
the transpose of
(I ; PX )X = X ; X = 0
Consequently, SSH0 is independent of
SSresiduals, and it follows that
314
(xi)
SSH 0
2
F = SSm residuals
2
=
(n;k)
SSH0
m
SSresiduals
n;k
Perform the test by rejecting H0 : C = if
Fm;n;k(2)
F > F(m;n;k);
with noncentrality parameter
2 = 12 (C ; d)T [C (X T X );C T ];1(C ; d)
0
where is a specied signicance level (Type
I error level) for the test.
and 2 = 0 if and only if H0 : C = d is true.
315
316
Type I error level:
n
o
= Pr F > Fm;n;k; j H0 is true
%
when H0 is true,
H0
F = MSMS
residuals
has a central F distribution
with (m; n ; k) d.f.
This is the probability of incorrectly rejecting
a null hypothesis that is true.
317
318
Type II error level:
0.8
Central F Distribution with (4,20) df
= Prffail to reject H0 H0 is falseg
0.4
0.6
= PrfF < Fm;n;k; H0 is falseg
%
when H0 is false;
F = MSMSH0
0.2
f(x;n)
= PrfType II errorg
0.0
residuals
0
1
2
3
4
5
has a noncentral F distribution
with (m; n ; k) d:f : and
noncentrality parameter > 0:
x
319
Power of a test:
power = 1 ; = PrfF > Fm;n;k; H0 is falseg
%
this determines the value
of the noncentrality
parameter.
320
321
0.8
Central and Noncentral
F Distributions with (4,20) df
Noncentral
Noncentral
F(1)
F(1)
0.4
Noncentral
Noncentral
F(3)
F(3)
Noncentral
Noncentral
F(5)
F(5)
0.0
0.2
f(x;v,w)
0.6
Central
Central
FF
0
1
2
3
4
5
x
322
0.6
0.8
Central and Noncentral F Densities
with (5,20) df and noncentrality parameter = 1.5
Central F
0.4
0.2
0.0
f(x;v,w)
Noncentral F
0
1
2
3
4
5
x
323
0.6
0.8
Central and Noncentral F Densities
with (5,20) df and noncentrality parameter = 3
Central F
0.4
0.0
0.2
f(x;v,w)
Noncentral F
0
1
2
3
4
5
x
324
0.6
0.8
Central and Noncentral F Densities
with (5,20) df and noncentrality parameter = 10
Central F
0.4
0.2
0.0
f(x;v,w)
Noncentral F
0
1
2
3
4
5
x
325
0.6
0.8
Central and Noncentral F Densities
with (5,20) df and noncentrality parameter = 20
For a xed type I error level (signicance level)
, the power of the test increases as the
noncentrality parameter increases.
Central F
2 = 12 (C ; d)T [C (X T X );C T ];1(C ; d)
"
"
0.4
size of
how much
the error the actual
variance value of
C
deviates
from the
hpothesized
value d
"
the design
of the experiment
(Note: the number
of observations
also aects
degrees of
freedom.)
0.0
0.2
f(x;v,w)
Noncentral F
0
1
2
3
4
5
x
326
Example 5.2 Eects of three diets on blood
coagulation times in rats.
Diet factor: Diet 1, Diet 2, Diet 3
Response: blood coagulation time
Model for a completely randomized experiment
with ni rats assigned to the i-th diet.
Yij = + i + ij
where
ij NID(0; 2)
for i = 1; 2; 3 and j = 1; 2; :::; ni.
Here, E (Yij ) = + i is the mean coagulation
time for rats fed the i-th diet.
327
Test the null hypothesis that the mean blood
coagulation time is the same for all three diets
H0 : + 1 = + 2 = + 3
against the general alternative that at least two
diets have dierent mean coagulation times
HA : ( + i) 6= ( + j ) for some i 6= j :
Equivalent ways to express the null hypothesis
are:
H0 : 1 = 2 = 3
"
#
0
1
;
1
0
H0 : C = 0 0 1 ;1
2 3
66 1 77 " 0 #
64 75 = 0
2
3
328
2 3
# 6 7 " #
H0 : C = 00 10 01 ;11 664 1 775 = 00
2
3
"
"
this is d
Obtain the OLS estimator for C C b where b = (X T X );X T Y
and evaluate
SSH0 = (C b ; 0)T [C (X T X );C T ];1(C b ; 0)
3
X
=
ni(Yi: ; Y::)2
i=1
SSH0
2 on 3 ; 1 = 2 d:f :
ni
3 X
X
SSresiduals = YT (I ; PX )Y =
(Yij ; Yi:)2
MSH0 =
SSresiduals on
MSresiduals = X
3
(ni ; 1)
i=1
i=1 j =1
3
X
(ni ; 1) d:f :
i=1
Reject H0 in favor of Ha if
F = MSMSH0 > F(2;P3i=1(ni;1));
residuals
How many observations (in this case rats) are
needed? Suppose we are willing to specify:
(i)
(ii)
(iii)
(iv)
= type I error level = .05
n1 = n2 = n3 = n
power .90 to detect
a specic alternative
( + 1) ; ( + 3) = 0:5
( + 2) ; ( + 3) = 330
329
For this particular alternative
2 3
"
#6 7 "
#
;1 66 1 77 = :5
C = 00 01 10 ;
1 4 2 5
3
and the power of the F-test is
n
o
power = Pr F(2;3n;3)(2) > F(2;3n;3);:05
where
"
# " #!T
2 = 12 :5 ; 00
h T ; T i;1 " :5 # " 0 #!
C (X X ) C
" ;# 0
h
i
= [:5 1] C (X T X );C T ;1 :15
%
X has n rows of (1 1 0 0)
n rows of (1 0 1 0)
n rows of (1 0 0 1)
331
In this case,
2 3n
6
X T X = 664 nn
n
n n n 37
n 0 0 77
0 n 05
0 0 n
and a generalized inverse is
2
3
66 00 n0;1 00 00 77
(X T X ); = 66 0 0 n;1 0 77
4
5
0 0 0 n;1
Then,
"
#
C (X T X );C T = n1 21 12
and
" #
2
T
;
T
;
1
= [:5 1][C (X X ) C ] :51
= n2
Choose n to achieve
:90 = power
= Pr F(2;3n;3) n2 > F(2;3n;3):05
332
0.6
0.4
0.0
0.2
Power
0.8
1.0
Power versus Sample Size
F-test for equal means
0
10
20
30
40
Sample Size
333
For testing
H0 : ( + 1) = ( + 2) = = ( + k )
against
HA : ( + 1) 6= ( + j )for somei 6= j
# This file is stored in power.spl
#===========================================
use
n <- 2:50
power <- 1 - pf(qf(.95,2,3*n-3),2,3*n-3,n/2)
#
#
F = MSMSH0
residuals
Unix users show insert the motif( )
function here to open a graphics window
where
par(fin=c(7.5,8.5),cex=1.2,mex=1.5,lwd=4,
mar=c(5.1,4.1,4.1,2.1))
plot(c(0,50), c(0,1), type="n",
xlab="Sample Size",ylab="Power",
main="Power versus Sample Size\n
F-test for equal means")
lines(n, power, type="l",lty=1)
F(k;1;Pki=1(ni;1))(2)
k
X
2 = 12 ni(i ; :)2
i=1
with
#============================================
k
X
nii
: = i=1k
X
i=1
334
ni
335
50
To obtain the formula for the noncentrality parameter, write the null hypothesis as
31 2 3
0H20 : 0 = C = 3 2
0
0 1 0 0
B
6 0 0 1 0 7 6 0
B
@64 .. 0 . . . . 75 ; 64 ..
0 0 0 1
Use
0
2
66 11
66 1
66
61
X = 666 1
66 ..
66 ..
64 .
1
1
1
0
0.
...
..
1 0
n1
n
..
..
n
:
1
n
:
0 0
0
1
1.
...
..
0 n
n
k
:
77CC 66 5A 64 ..
1
2
n
n
k
:
k
77
75
3
07
0 77
0 77
0 777
0. 77
. 77
1 777
15
1
2
66 nn: nn1
1 1
X T X = 666 n2
64 ..
nk
20 0
66 0 n;1
1
T
;
(X X ) = 6666 0 0
4 ..
3
n2 nk 7
77
7
n2
. . . 775
nk
0 0 n;2 1 ...
0 0 0
0
0
n;k 1
3
77
77
77 :
5
Then
2 = 12 (C ; 0)T [C (X T X );C T ];1(C ; 0)
k
X
= 12 ni(1 ; :)2
i=1
336
Consider minimum power of the test that rejects H0 when
F = MSMSH0
residuals
For n1 = n2 = = nk = n we have
2 n w2
Then, the power of the test is at least
> F(k;1;Pki=1(ni;1));
against alternatives where
ji ; :j w for at lease one i and a specied w.
Then,
337
power Pr F(k;1;k(n;1)) (n w2) > F(k;1;k(n;1);
%
%
noncentral F
percentile
random
of a
variable
central F
distribution
F = MSMS
H0
residuals
k
X
2 = 12 ni(i ; :)2
i=1
w2 minfn1; : : : ; nkg
The following graph shows minimum power for
= :01; k = 4; w = :50 as a function of the
common sample size n.
338
339
0.6
0.4
0.0
0.2
Power
0.8
1.0
Minimum Power vs. Sample Size
At least one difference exceeds sigma*0.
0
20
40
60
80
100
Sample Size
340
#======================================================#
# min.power.vs.sample.size()
#
# -------------------------#
# Input : k = # of groups
#
#
n = maximum sample size in the
#
#
group(n1=n2=...=nk)
#
#
alpha = type I error level
_
#
#
w =spesifies alternative |ai-a.| >= w*sigma #
# Output: line pot (Minimum Power vs. Sample Size)
#
# Note : Unix users must activate a motif() window
#
#
before using this function.
#
#======================================================#
min.power.vs.sample.size _ function(k,n,alpha,w)
{
n _ 2:n
power _ 1 - pf(qf(1-alpha,k-1,k*(n-1)),k-1,k*(n-1),n*w^2)
}
#
#
#
#
plot(c(0,max(n)), c(0,1),type="n",xlab="Sample Size",
ylab="Power", main=paste("Minimum Power vs. Sample
Size\nAt least one difference exceeds sigma*",w,sep=""))
lines(n, power, type="l",lty=1)
Condence intervals for estimable functions of //[.2in]
Defn 5.2: Suppose Z N (0; 1) is distributed
independently of W 2v , and then the distribution of
t = qZW
v
is called the student t-distribution with v degrees of freedom.
We will use the notation
t tv
The following statement can be used to
execute this function
min.power.vs.sample.size(4,100,.01,.5)
341
342
0.4
Central t Densities
0.2
0.0
0.1
f(x;df)
0.3
2 df
5 df
30 df
-4
-2
0
2
4
x
343
# This code is stored in:
tden.spl
#=================================================#
# t.density.plot()
#
# -----------------------#
# Input : degrees of freedom; it can be a vector. #
#
(e.g.) t.density.plot(c(2,5,7,30))
#
#
creates curves for df = 2,5,7,and 30 #
# Output: density plot of Student t distribution. #
# Note: Unix users must first use
motif()
#
#
to open a grpahics window before
#
#
using this function.
#
#================================================ #
# The following code creates a legend;
}
t.density.plot <- function(df)
{
x <- seq(-4,4,,100)
#
#
#
#
#
#
#
#
# draw x,y axis and title
plot(c(-4,4), c(0,.4), type="n",
xlab="x", ylab="f(x;df)",
main="Central t Densities")
legend( x = rep(0.85,length(df)) ,
y = rep(.4,length(df)) ,
legend = paste(as.character(df),"df") ,
lty = seq(1,by=2,length=length(df)) ,
bty = "n")
This function can be executed as follows.
Windows users should delete the motif( )
command.
motif( )
source("tden.spl")
par(fin=c(7,8),cex=1.2,mex=1.5,lwd=4)
t.density.plot(c(2,5,30))
for(i in 1:length(df)) {
lty.num <- 2*i-1
# specify the line types.
f.x <- dt(x,df[i])
# calculate density.
lines(x, f.x, type="l",lty=lty.num)
# draw.
}
344
345
For the normal-theory Gauss-Markov model
Y N (X ; 2I )
and from Result 5.1.(iv) we have for an estimable function, cT , that the OLS estimator
cT b = c T ( X T X ) ; X T Y
follows a normal distribution, i.e.,
cT b N (cT ; 2cT (X T X );c) :
It follows that
T
T
Z = q c2 Tb ; cT ; N (0; 1)
c (X X ) c
From Result 5.1.(ix), we have
1
1 T
2 = 2 Y (I ; PX )Y
2(n;k)
where k = rank(X ).
Using the same argument that we used to derive Result 5.1.(x), we can show that cT b is
distributed independently of 12 SSE.
First note that
"
#
cT b
(I ; PX )Y
" T T ; T#
= c ((XI ;XP) )X Y
X
has a joint normal distribution under the
normal-theory Gauss-Markov model. (From
Result 4.1)
346
Note that
Cov(cT b; (I ; P ; X )Y)
= (cT (X T X );X T )(V ar(Y))(I ; PX )T
= (CT (X T X );X T )(2)(I ; PX )
= 2CT (X T X );X T (I ; PX )
=0
"
this is a matrix of zeros
347
Then,
t = q ZSSE
2(n;k)
=
pc2TcTb;(XcTTX );c
q
SSE
2(n;k)
= r SSEcT bT;cT T
Consequently, (by Result 4.4)
cT b is distributed independently
of e = (I ; PX )Y
which imples that
cT b is distributed independently
of SSE = eT e.
;
;k) c (X X ) c
(n
t(n;k)
%
call SSE
n;k the MSE
348
349
0.2
0.0
0.1
f(x;df)
0.3
0.4
Central t Density
( 5 degrees of freedom )
-4
-2
0
2
4
x
350
It follows that
1; =
(
T
cT t
Pr ;t(n;k);=2 p c b ;
MSE cT(XTX);c (n;k);=2
q
= Pr cT b ; t(n;k);=2 MSE cT(XTX);c cT q
cT b + t(n;k);=2 MSE cT(XTX);c
and a (1 ; ) 100% condence interval for
cT is
q
cT b ; t)(n ; k); =2
cT b + t(n;k);=2
q
MSE cT(XTX);c ;
MSE cT(XTX);c
351
)
For brevity we will also write
cT b t(n;k);=2 ScT b
where
q
ScT b = MSE cT(XTX);c :
For the normal-theory Gauss-Markov model
with Y N (X ; 2I ); the interval
cT b t(n;k);=2ScT b
is the shortest random interval with probability
(1 ; ) of containing CT .
352
Condence interval for 2:
For the normal-theory Gauss-Markov model
with Y N (X ; 2I ) we have shown that
The resulting (1;)100% condence interval
for 2 is
0
1
SSE
SSE
@ 2
A
(n;k);=2 ; 2(n;k);1;=2
SSE = YT (I ; PX )Y 2
(n;k)
2
2
Then,
1 ; = Pr 2(n;k);1;=2 SSE
2 (n;k);=2
8
9
< SSE
=
SSE
2
= Pr : 2
(n;k);1;=2 ;
(n;k);=2
353
0.0
0.04
f(x;df)
0.08
0.12
Central Chi-Square Density
0
5
10
15
x
20
25
30
354
Download