6. Reparameterization: Matrix notation: sight, may actually be equivalent.

advertisement
6. Reparameterization:
Models that may appear to be dierent at rst
sight, may actually be equivalent.
Example 6.1
Two-way classication in Example 3.3. Consider the \cell mean" model.
i = 1; 2
j = 1; 2
k = 1; 2
Yijk = ij + ijk
Matrix notation:
2
6
6
6
6
6
6
6
6
6
6
6
6
6
4
Y111
Y112
Y121
Y122
Y211
Y212
Y221
Y222
3 2
77 6
77 6
6
77 6
6
77 = 6
6
77 6
6
77 6
6
75 6
6
4
1
1
0
0
0
0
0
0
or
0
0
1
1
0
0
0
0
3
0
0
0
0
1
1
0
0
07
0 77
0 777
0 77
0 77
0 777
15
1
2
66
64
11
12
21
22
2
6
3 666
77 666
75 + 66
66
66
4
358
357
66
66
66
66
64
Y111
Y112
Y121
Y122
Y211
Y212
Y221
Y222
or
1
77
66 1
77
61
77 = 666 1
77
66 1
75
64 1
1
1
1
1
1
1
0
0
0
0
0
0
0
0
1
1
1
1
1
1
0
0
1
1
0
0
0
0
1
1
0
0
1
1
1
1
0
0
0
0
0
0
0
0
1
1
0
0
0
0
0
0
0
0
1
1
0
0
0
0
0
0
0
0
1
1
77 6
6
77 6
6
77 6
6
77 6
6
75 6
6
6
4
3
77
77
77
77
77
77
75
Y = W + where ijk NID(0; 2)
The \eects" model:
Yijk ; + i + j + ( )ij + ijk
where
i = 1; 2
ijk NID(0; 2) j = 1; 2
k = 1; 2
Matrix
notation:
2
3
2
32
111
112
121
122
211
212
221
222
The models are \equivalent" in the sense that
the space spanned by the columns of W is the
same as the space spanned by columns of X .
Consequently, you can nd matrices F and G
such that
1
2
1
2
11
12
21
22
Y = X + 3
77
77
77
77
77
75
2
6
6
6
6
6
6
6
W = X 666
6
6
6
6
6
4
0
0
0
0
0
0
1
0
0
0
0
0
0
0
0
0
1
0
3
07
0 77
0 77
0 777
0 77 = XF
0 77
0 777
05
1
and
2
6
X = W 664
359
0
0
0
0
0
1
0
0
0
1
1
1
1
1
1
0
0
0
0
1
1
1
0
1
0
0
1
0
1
1
0
0
0
0
1
0
0
0
0
1
0
0 37
0 77 = W G
05
1
360
Then,
(i) rank(X ) = rank (W )
Example 6.2 Regression model for the yield of
a chemical process in Example 3.1.
Yi = 0 + 1X1i + 2X2i + i
%
%
-
(ii) Estimated mean responses are the same:
Y^ = X (X T X );X T Y
= W (W T W );1W T Y
or
Y^ = PX Y = PW Y
yield
temperature
time
An \equivalent" model:
Yi = 0 + 1(X1i ; X1:) + 2(X2i ; X2:) + 1
(iii) Residual vectors are the same
e = Y ; Y^ = (I ; PX )Y
= (I ; PW )Y
361
For the rst model:
2
3
2
Y
1 X11
1
66
7
66
66 Y2 777
66 1 X12
66 Y3 77 = 66 1 X13
4 Y4 5
4 1 X14
Y5
1 X15
= X + 362
The space spanned by the columns of X is the
same as the space spanned by the columns of
W.
X21
X22
X23
X24
X25
3
77 2
77 6
77 4
5
0
1
2
2
3 6
6
75 + 6
6
6
6
4
1
2
3
4
5
3
7
7
7
7
7
7
5
2
X = W 64
Y1
Y2
Y3
Y4
Y5
3
21
77
61
75 = 664 1
1
X11 ; X1
X12 ; X1
X13 ; X1
X14 ; X1
1 X15 ; X1
= W + X21 ; X2
X22 ; X2
X23 ; X2
X24 ; X2
X25 ; X2
3
2
2
3
77 0
6
75 4 1 5 + 6
6
4
2
3
X1 X2
7
1 0
0 1
5 = WG
and
For the second model:
2
66
64
1
0
0
1
2
3
4
5
3
77
75
2
W = X 64
3
1 ;X1 ;X2 7
0 1 0 5 = XF
0 0 1
and
Y^ = PX Y = PW Y
e = Y ; Y^ = (I ; PX )Y = (I ; PW )Y
363
364
Defn 6.1: Consider two Gauss-Markov models:
Model 1: E (Y) = X and V ar(Y) = 2I
Equivalently, if the linear space spanned by the
columns of X is the same as the linear space
spanned by the columns of W .
The previous examples illustrate that if one
model is a reparameterization of the other,
then
(i) rank(X ) = rank(W )
Model 2: E (Y) = W and V ar(Y) = 2I
where X is an n k model matrix and W is an
n q model matrix.
(ii) Least squares estimates of response means
are the same, i.e., Y^ = PX Y = PW Y
We say that one model is a reparameterization
of the other if there is a k q matrix F and a
q k matrix G such that
W = XF and X = W G :
(iii) Residuals are the same, i.e.,
e = Y ; Y^ = (I ; PX )Y = (I ; Pw)Y
(iv) An unbiased estimator for 2 is provided by
MSE = SSE=(n-rank(X)) where,
SSE = eT e =
YT (I ; PX )Y
= YT (I ; PW )Y
365
366
Result 6.1. Suppose two Gauss-Markov
models:
Model 1: E (Y) = X V ar(Y) = 2I
Model 2: E (Y) = W V ar(Y) = 2I
Reasons fro reparameterizing models:
are reparameterizations of each other, and let
F be a matrix such that W = XF . Then
(i) Reduce the number of parameters
Obtain a full rank model
Avoid use of generalized inverse
(i) If CT is estimable under model 1, then
= F and CT F is estimable under
Model 2.
(ii) Make computations easier
In Examples 6.1 and 6.2,
W T W is a diagonal matrix and
(W T W );1 is easy to compute.
(iii) More meaningfull interpretation of parameters.
(ii) Let ^ = (X T X );X T Y and ^ =
(W T W );W T Y. If CT is estimable, then
CT ^ = CT F ^
(iii) if H0 : CT = d is testable under Model
1, then H0 : CT F = d is testable under
Model 2.
367
368
Proof:
(i) If CT is estimable for Model 1, then (by
Result 3.9 (i))
CT = aT X for some a :
Hence,
CT F = aT X F = aT W
which implies that CT F is estimable for
Model 2.
(ii) Since CT is estimable, the unique b.l.u.e.
is
CT ^ = CT (X T X );X T Y
= aT X T (X T X );X T Y
= aT PX Y for some a
369
By a similar argument
Then,
PW PX = PX
PW = P + W T
= (PX PX )T
= PWT PXT
= P W PX
= PX
371
Since CT F is also estimable, the unique
b.l.u.e. for CT F is
CT F (W T W );W T Y = aT XF (W T W );W T Y
= aT W (W T W );W T Y
= aT PW Y
for the same a.
Hence, the estimators are the same if PX =
PW . To show this, note that
PX W = PX XF = XF = W
which implies
PX PW = PX W (W T W );W T
= W (W T W );W T
= PW
370
Download