Simultaneous Inferences and Other Regression Topics

advertisement
Simultaneous Inferences and
Other Regression Topics
KNNL – Chapter 4
Bonferroni Inequality
Suppose we have 2 events: A1 and A2 with: P  A1   P  A2   
Prob that A1 and/or A2 occur :
P  A1
A2   P  A1   P  A2   P  A1
A2 
Prob that neither A1 nor A2 occur (complementary event of A1

P A1


A2  1  P  A1
 P A1

A2   1  P  A1   P  A2   P  A1
A2 ) :
A2 
A2  1  P  A1   P  A2 

For these events: P A1

A2  1      1  2
Application: We want simultaneous Confidence Intervals for b0 and b1 such that we can
be (1-)100% confident that both intervals contain true parameter:
A1 ≡ Event that CI for b0 does not cover b0 A2 ≡ Event that CI for b1 does not cover b1
Then: The probability that both intervals are correct is ≥ 1-2
Thus, if we construct (1-(/2))100% CIs individually, Pr{Both Correct} ≥ 1-2(/2) = 1-
Joint Confidence Intervals for b0 and b1
Goal: Want Confidence Intervals for b 0 , b1 so that we can be
(1- )100% Confident that BOTH intervals contain the true parameter.
"Trick:" Make each confidence interval at 1   / 2 100% Confidence
B  t 1   4  ; n  2 
1   / 2 100% CI for b0 :
1   / 2 100% CI for b1 :
b0  Bs b0 
b1  Bs b1
Note: If we want to be 95% confident that both intervals are
correct, we set-up 97.5% confidence intervals for each parameter
df
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
t(.975,df) t(.9875,df)
12.706
25.452
4.303
6.205
3.182
4.177
2.776
3.495
2.571
3.163
2.447
2.969
2.365
2.841
2.306
2.752
2.262
2.685
2.228
2.634
2.201
2.593
2.179
2.560
2.160
2.533
2.145
2.510
2.131
2.490
2.120
2.473
2.110
2.458
2.101
2.445
2.093
2.433
2.086
2.423
2.080
2.414
2.074
2.405
2.069
2.398
2.064
2.391
2.060
2.385
2.056
2.379
2.052
2.373
2.048
2.368
2.045
2.364
2.042
2.360
Simultaneous Estimation of Mean Responses
• Working-Hotelling Method: Confidence Band for
Entire Regression Line. Can be used for any number
of Confidence Intervals for means, simultaneously
• Bonferroni Method: Can be used for any g
Confidence Intervals for means by creating
(1-/g)100% CIs at each of g specified X levels
^

^
W  2 F 1   ; 2, n  2 
Working-Hotelling: Y h  Ws Y h
^

^
Bonferroni: Y h  Bs Y h

B  t 1    2 g   ; n  2

Bonferroni t-table ( = 0.05, 2-sided)
g
1
2
3
4
5
6
7
8
9
10
1-.05/(2g)
0.9750
0.9875
0.9917
0.9938
0.9950
0.9958
0.9964
0.9969
0.9972
0.9975
df
t(1-.05/2g,df) t(1-.05/2g,df) t(1-.05/2g,df) t(1-.05/2g,df) t(1-.05/2g,df) t(1-.05/2g,df) t(1-.05/2g,df) t(1-.05/2g,df) t(1-.05/2g,df) t(1-.05/2g,df)
1
12.706
25.452
38.188
50.923
63.657
76.390
89.123
101.856
114.589
127.321
2
4.303
6.205
7.649
8.860
9.925
10.886
11.769
12.590
13.360
14.089
3
3.182
4.177
4.857
5.392
5.841
6.232
6.580
6.895
7.185
7.453
4
2.776
3.495
3.961
4.315
4.604
4.851
5.068
5.261
5.437
5.598
5
2.571
3.163
3.534
3.810
4.032
4.219
4.382
4.526
4.655
4.773
6
2.447
2.969
3.287
3.521
3.707
3.863
3.997
4.115
4.221
4.317
7
2.365
2.841
3.128
3.335
3.499
3.636
3.753
3.855
3.947
4.029
8
2.306
2.752
3.016
3.206
3.355
3.479
3.584
3.677
3.759
3.833
9
2.262
2.685
2.933
3.111
3.250
3.364
3.462
3.547
3.622
3.690
10
2.228
2.634
2.870
3.038
3.169
3.277
3.368
3.448
3.518
3.581
11
2.201
2.593
2.820
2.981
3.106
3.208
3.295
3.370
3.437
3.497
12
2.179
2.560
2.779
2.934
3.055
3.153
3.236
3.308
3.371
3.428
13
2.160
2.533
2.746
2.896
3.012
3.107
3.187
3.256
3.318
3.372
14
2.145
2.510
2.718
2.864
2.977
3.069
3.146
3.214
3.273
3.326
15
2.131
2.490
2.694
2.837
2.947
3.036
3.112
3.177
3.235
3.286
16
2.120
2.473
2.673
2.813
2.921
3.008
3.082
3.146
3.202
3.252
17
2.110
2.458
2.655
2.793
2.898
2.984
3.056
3.119
3.173
3.222
18
2.101
2.445
2.639
2.775
2.878
2.963
3.034
3.095
3.149
3.197
19
2.093
2.433
2.625
2.759
2.861
2.944
3.014
3.074
3.127
3.174
20
2.086
2.423
2.613
2.744
2.845
2.927
2.996
3.055
3.107
3.153
21
2.080
2.414
2.601
2.732
2.831
2.912
2.980
3.038
3.090
3.135
22
2.074
2.405
2.591
2.720
2.819
2.899
2.965
3.023
3.074
3.119
23
2.069
2.398
2.582
2.710
2.807
2.886
2.952
3.009
3.059
3.104
24
2.064
2.391
2.574
2.700
2.797
2.875
2.941
2.997
3.046
3.091
25
2.060
2.385
2.566
2.692
2.787
2.865
2.930
2.986
3.035
3.078
26
2.056
2.379
2.559
2.684
2.779
2.856
2.920
2.975
3.024
3.067
27
2.052
2.373
2.552
2.676
2.771
2.847
2.911
2.966
3.014
3.057
28
2.048
2.368
2.546
2.669
2.763
2.839
2.902
2.957
3.004
3.047
29
2.045
2.364
2.541
2.663
2.756
2.832
2.894
2.949
2.996
3.038
30
2.042
2.360
2.536
2.657
2.750
2.825
2.887
2.941
2.988
3.030
40
2.021
2.329
2.499
2.616
2.704
2.776
2.836
2.887
2.931
2.971
50
2.009
2.311
2.477
2.591
2.678
2.747
2.805
2.855
2.898
2.937
60
2.000
2.299
2.463
2.575
2.660
2.729
2.785
2.834
2.877
2.915
70
1.994
2.291
2.453
2.564
2.648
2.715
2.771
2.820
2.862
2.899
80
1.990
2.284
2.445
2.555
2.639
2.705
2.761
2.809
2.850
2.887
90
1.987
2.280
2.440
2.549
2.632
2.698
2.753
2.800
2.841
2.878
100
1.984
2.276
2.435
2.544
2.626
2.692
2.747
2.793
2.834
2.871
∞
1.960
2.241
2.394
2.498
2.576
2.638
2.690
2.734
2.773
2.807
Simultaneous Predictions of New Responses
• Scheffe’s Method: Widely used method for making
simultaneous tests and confidence intervals. Like WH, based on F-distribution, but does increase with g,
the number of simultaneous predictions
• Bonferroni Method: Can be used for any g
Confidence Intervals for means by creating
(1-/g)100% CIs at each of g specified X levels
^
Scheffe: Y h  Ss pred
^
Bonferroni: Y h  Bs pred
S  gF 1   ; g , n  2 

B  t 1    2 g   ; n  2

Regression Through the Origin
• In some applications, it is believed that the
regression line goes through the origin
• This implies that E{Y|X} = b1X (proportional relation)
• Note, that if we imply that all Y=0 when X=0, then
the variance of Y is 0 when X=0 (not consistent with
the regression models we have fit so far)
• Should only be used if there is a strong theoretical
reason
• Analysis of Variance and R2 interpretation are
changed. Should only use t-test for slope
Regression Through the Origin
 i ~ N  0,  2  independent
Model: Yi  b1 X i   i
Least Squares Estimation:
n
Q   Yi  b1 X i 
2

i 1
n
Q
 2 Yi  b1 X i   X i  Setting derivative to 0, and solving for b1 
b1
i 1
n
n
n
XY bX
i 1
i i
1
i 1
2
i
 b1 
XY
i 1
n
X
i 1
^
Y i  b1 X i
i i
2
i


 X 
   n i Yi
2 
i 1 
  Xi 
 i 1

n
n
^
SSE   ei2
ei  Yi  Y i
s 2  MSE 
i 1
SSE
n 1



^
MSE  X 
X h2 
MSE
2
2
2

s b1  n
s Yh 
s pred  MSE 1  n
n
2
2
2 

Xi
Xi


  Xi 
i 1
i 1
i 1


1   100% CI for b1 : b1  t 1   / 2  , n  1 s b1

2
h
1   100% CI for E Yh   b1 X h : Y h  t 1   / 2  , n  1 s
^
1   100% CI for Yh ( new) : Y h  t 1   / 2  , n  1 s pred
^

^
Yh
Measurement Errors
• Measurement Error in the Dependent Variable (Y): As
long as there is not a bias (consistently recording too
high or low), no problem (Measurement Error is
absorbed into ).
• Measurement Error in the Independent Variable (X):
Causes problems in estimating b1 (biases downward)
when the observed (recorded) value is random. See
next slide for description.
• Measurement Error in the Independent Variable (X):
Not a problem when the observed (recorded) value
is fixed and actual value is random (e.g. temperature
on oven is set at 400⁰ but actual temperature is not)
Measurement Error in X (Random)
Observed (recorded) Value: X i*
True (unobserved) Value: X i
 i  X i*  X i
True Model: Yi  b 0  b1 X i   i  b 0  b1  X i*   i    i  b 0  b1 X i*    i  b1 i 
Assumptions: No Bias in Measurement Error, and uncorrelated with Random Error
E  i   E  i   E  i i   0


   X i* ,  i  b1 i   E  X i*  E  X i*   i  b1 i   E  i  b1 i  


 E  X i*  X i    i  b1 i  
 E  i   i  b1 i   E  i i   b1 E  i2    b1 2  i 
 The recorded value X i* is not independent of the "error" term:   i  b1 i 
 E Yi | X
*
i
b
*
0
b X
*
1
*
i
where b  b1
*
1
 X2
 X2   Y2
 b1
Inverse Prediction/Calibration
Goal: Predict a new X value based on an observed new Y value, based on existing Regression:
Model: Yi  b 0  b1 X i   i

^
Y  b0  b1 X
Observe a new (typically easy to measure) Yh (new )
and want to predict X h (new ) corresponding to it (difficult to measure)
^
Point Estimate: X h (new ) 
Yh (new )  b0
b1
Approximate 1   100% Prediction Interval: X h (new )  t 1   / 2  , n  2  s pred X 
^
2

 ^
 
X h (new )  X  
MSE  1 
 
2
where: s pred X   2 1   n
2 
b1  n
Xi  X 


i 1




t 1   / 2  , n  2   MSE
Approximate Interval is appropriate if
is small (say < 0.1)
n
2
b12  X i  X
2
i 1


Bonferroni or Scheffe adjustments should be made for multiple simultaneous predictions
Choice of X Levels
• Note that all variances and standard errors depend on
SSXX which depends on the spacing of the X levels, and
the sample size.
• Depending on the goal of research, when planning a
controlled experiments, and selecting X levels, choose:
 2 levels if only interested in whether there is an effect and its
direction
 3 levels if goal is describing relation and any possible curvature
 4 or more levels for further description of response curve and
any potential non-linearity such as an asymptote
Download