# Chapter 5 ```Chapter 5
Regression with a Single Regressor:
Hypothesis Tests and Confidence Intervals
 Solutions to Exercises
1
(a) The 95% confidence interval for β1 is {−5.82 &plusmn; 1.96 &times; 2.21}, that is −10.152 ≤ β1 ≤ −1.4884.
(b) Calculate the t-statistic:
t act =
βˆ 1 − 0 −5.82
=
= −2.6335.
SE( βˆ 1) 2.21
The p-value for the test H0 : β1 = 0 vs. H1 : β1 ≠ 0 is
p-value = 2Φ( −|t act |) = 2Φ (−2.6335) = 2 &times; 0.0042 = 0.0084.
The p-value is less than 0.01, so we can reject the null hypothesis at the 5% significance level,
and also at the 1% significance level.
(c) The t-statistic is
t act =
βˆ 1 − (−5.6) 0.22
=
= 0.10
SE( βˆ 1)
2.21
The p-value for the test H0 : β1 = −5.6 vs. H1 : β1 ≠ −5.6 is
p-value = 2Φ (−|t act |) = 2Φ (−0.10) = 0.92
The p-value is larger than 0.10, so we cannot reject the null hypothesis at the 10%, 5% or 1%
significance level. Because β1 = −5.6 is not rejected at the 5% level, this value is contained in
the 95% confidence interval.
(d) The 99% confidence interval for β0 is {520.4 &plusmn; 2.58 &times; 20.4}, that is, 467.7 ≤ β 0 ≤ 573.0.
2.
(a) The estimated gender gap equals \$2.12/hour.
(b) The hypothesis testing for the gender gap is H0 : β1 = 0 vs. H1 : β1 ≠ 0. With a t-statistic
t act =
βˆ 1 − 0 2.12
=
= 5.89,
SE ( βˆ 1) 0.36
the p-value for the test is
p-value = 2Φ( −|t act |) = 2Φ (−5.89) = 2 &times; 0.0000 = 0.000 (to four decimal places)
The p-value is less than 0.01, so we can reject the null hypothesis that there is no gender gap at a
1% significance level.
Solutions to Exercises in Chapter 5
(c) The 95% confidence interval for the gender gap β1 is {2.12 &plusmn; 1.96 &times; 0.36}, that is,
1.41 ≤ β1 ≤ 2.83.
(d) The sample average wage of women is βˆ0 = \$12.52/hour. The sample average wage of men is
βˆ 0 + βˆ 1 = \$12.52 + \$2.12 = \$14.64/hour.
(e) The binary variable regression model relating wages to gender can be written as either
Wage = β 0 + β1 Male + ui ,
or
Wage = γ 0 + γ 1Female + vi .
In the first regression equation, Male equals 1 for men and 0 for women; β 0 is the population
mean of wages for women and β 0 + β1 is the population mean of wages for men. In the second
regression equation, Female equals 1 for women and 0 for men; γ 0 is the population mean of
wages for men and γ 0 + γ 1 is the population mean of wages for women. We have the following
relationship for the coefficients in the two regression equations:
γ 0 = β 0 + β1 ,
γ 0 + γ 1 = β0 .
Given the coefficient estimates β̂ 0 and β̂ 1 , we have
γˆ 0 = βˆ 0 + βˆ 1 = 14.64,
γˆ 1 = βˆ 0 − γˆ 0 = − βˆ 1 = −2.12.
Due to the relationship among coefficient estimates, for each individual observation, the OLS
residual is the same under the two regression equations: uˆ i = vˆ i. Thus the sum of squared
residuals, SSR = ∑ i =1 uˆ i2, is the same under the two regressions. This implies that both
n
2
SER = ( SSR
and R 2 = 1 − SSR
are unchanged.
TSS
n −1 )
1
&middot;
In summary, in regressing Wages on Female, we will get
Wages = 14.64 − 2.12 Female,
R 2 = 0.06, SER = 4.2.
3.
The 99% confidence interval is 1.5 &times; {3.94 &plusmn; 2.58 &times; 0.31) or 4.71 lbs ≤ WeightGain ≤ 7.11 lbs.
4.
(a) −3.13 + 1.47 &times; 16 = \$20.39 per hour
(b) The wage is expected to increase from \$14.51 to \$17.45 or by \$2.94 per hour.
(c) The increase in wages for college education is β1 &times; 4. Thus, the counselor’s assertion is that
− 2.50
β1 = 10/4 = 2.50. The t-statistic for this null hypothesis is t = 1.470.07
− 14.71, which has a
p-value of 0.00. Thus, the counselor’s assertion can be rejected at the 1% significance level. A
95% confidence for β1 &times; 4 is 4 &times; (1.47 &plusmn; 1.97 &times; 0.07) or \$5.33 ≤ Gain ≤ \$6.43.
29
30
Stock/Watson - Introduction to Econometrics - Second Edition
5
(a) The estimated gain from being in a small class is 13.9 points. This is equal to approximately 1/5
of the standard deviation in test scores, a moderate increase.
(b) The t-statistic is t act = 13.9
= 5.56, which has a p-value of 0.00. Thus the null hypothesis is
2.5
rejected at the 5% (and 1%) level.
(c) 13.9 &plusmn; 2.58 &times; 2.5 = 13.9 &plusmn; 6.45.
6.
(a) The question asks whether the variability in test scores in large classes is the same as the
variability in small classes. It is hard to say. On the one hand, teachers in small classes might able
so spend more time bringing all of the students along, reducing the poor performance of
particularly unprepared students. On the other hand, most of the variability in test scores might be
beyond the control of the teacher.
(b) The formula in 5.3 is valid for heteroskesdasticity or homoskedasticity; thus inferences are valid
in either case.
7.
3.2
= 2.13 with a p-value of 0.03; since the p-value is less than 0.05, the null
(a) The t-statistic is 1.5
hypothesis is rejected at the 5% level.
(b) 3.2 &plusmn; 1.96 &times; 1.5 = 3.2 &plusmn; 2.94
(c) Yes. If Y and X are independent, then β1 = 0; but this null hypothesis was rejected at the 5% level
in part (a).
(d) β1 would be rejected at the 5% level in 5% of the samples; 95% of the confidence intervals would
contain the value β1 = 0.
8.
(a) 43.2 &plusmn; 2.05 &times; 10.2 or 43.2 &plusmn; 20.91, where 2.05 is the 5% two-sided critical value from the t28
distribution.
(b) The t-statistic is t act = 61.57.4− 55 = 0.88, which is less (in absolute value) than the critical value of
20.5. Thus, the null hypothesis is not rejected at the 5% level.
(c) The one sided 5% critical value is 1.70; tact is less than this critical value, so that the null
hypothesis is not rejected at the 5% level.
9.
(a) β =
1 1
X n
(Y1 + Y2 +
L + Y ) so that it is linear function of Y , Y , …, Y .
n
1
2
n
(b) E(Yi|X1, …, Xn) = β1Xi, thus
E ( β |X1 ,
K, X ) = E X1 1n (Y + Y + L + Y )|X , K, X )
11
=
β (X + L + X ) = β
Xn
n
1
1
2
n
1
n
1
n
1
10. Let n0 denote the number of observation with X = 0 and n1 denote the number of observations with
X = 1; note that
∑
n
i =1
( X i − X )2
Y = n1 Y1 +
n
n0
n
Y0
∑
=∑
n
i =1
X i = n1 ; X = n1| n;
n
i =1
X i2 − nX 2 = n1 −
1
n1
n12
n
∑
n
i =1
(
X iYi = Y1 ;
= n1 1 −
n1
n
)=
n1n0
n
; n1Y1 + n0Y0 = ∑ i =1Yi , so that
n
Solutions to Exercises in Chapter 5
From the least squares formula
∑ ( X − X )(Yi − Y ) ∑i=1 Xi (Yi − Y ) ∑i=1 XiYi − Yn1
= n
=
βˆ1 = i=1 n i
∑i=1 ( Xi − X )2
∑i=1 ( Xi − X )2
n1n0 |n
n
=
n
n
n
n
n
n

(Y1 − Y ) =  Y − 1 Y1 − 0 Y0  = Y1 − Y0
n0
n0 
n
n 
n 
n n + n0
n
and βˆ0 = Y − βˆ1 X =  0 Y0 + 1 Y1  − (Y1 − Y0 ) 1 = 1
Y0 = Y0
n
n

n
n
11. Using the results from 5.10, βˆ0 = Ym and βˆ1 = Yw − Ym . From Chapter 3, SE(Ym ) =
SE(Yw − Ym ) =
sm2
nm
Sm
nm
and
s2
+ nww . Plugging in the numbers βˆ0 = 523.1 and SE ( βˆ0 ) = 6.22; βˆ1 = −38.0 and
SE( βˆ1 ) = 7.65.
12. Equation (4.22) gives
σ β2ˆ =
0
var (H i ui )
( )
n  E H 
2
i
2
, where H i = 1 −
&micro;x
( )
E X i2
Xi .
Using the facts that E (ui | X i ) = 0 and var (ui | X i ) = σ u2 (homoskedasticity), we have



 i



E (Hi ui ) = E u −
=0−
&micro;x
( )
2
i
E X



i i



X u = E (ui ) −
&micro;x
( )
E Xi2
E [ Xi E (ui |Xi )]
&micro;x
&times; 0 = 0,
E ( Xi2 )
and


 

 i

 
E [(Hi ui ) ] = E u −
2
&micro;x
E  Xi2 
 2 
 
 
i i  

 
 
Xu


 2
 i



2


 &micro; 
&micro;x
2
2 2 
x


= E u − 2  2  Xiui +  2  Xi ui 

E  Xi 
 E  Xi  


2
 &micro;   2  2 
&micro;


x
= E u − 2 x E  Xi E  ui2|Xi   + 
 E  Xi E  ui |Xi  
 2  
 2



E  Xi 
 E  Xi   
 2
 i 


2
 &micro;   2
&micro;
x
= σ u2 − 2 x 2 &micro;xσ u2 + 
 E Xi σ u2
E  Xi 
 E  Xi2    

&micro;2 
=  1 −  x 2  σ u2 .
 E  Xi  



Because E ( H i ui ) = 0, var ( H i ui ) = E[( H i ui )2 ], so
31
32
Stock/Watson - Introduction to Econometrics - Second Edition

&micro; x2
var (H i ui ) = E[( H i ui )2 ] =  1 −
 E X i2

( )

 σ u2 .


We can also get
( )= E
E H
2
i


 



 
1−
&micro;x
( )
E Xi2
 2 
 
 
i  

 
 

X







= E 1− 2
&micro;x
( )
E Xi2

2
 &micro;
x
Xi + 
 E Xi2
( )
 
 Xi2 
 

2
 &micro; 
&micro;2
&micro;2
x
=1− 2 x 2 + 
E Xi2 = 1− x 2 .

E ( Xi )  E ( Xi2 ) 
E ( Xi )


( )
Thus
σ β2ˆ =
0
=
13. (a)
(b)
(c)
(d)
14. (a)
var (Hiui )
 nE H 2 2 
i


( )
E ( Xi2 )σ u2
(

&micro;x2  2
1 −
σ
 E Xi2  u
σ u2


=
=
2


&micro;x2
&micro;x2 
n 1 −

n 1 −
 E Xi2
 E Xi2 



( )
( )
( )
E ( X )σ
=
.
2
i
)




2
u
nσ X2
n[E Xi2 − &micro; x2 ]
Yes
Yes
They would be unchanged
(a) is unchanged; (b) is no longer true as the errors are not conditionally homosckesdastic.
X
From Exercise (4.11), βˆ = ∑ aiYi where ai = n i 2 . Since the weights depend only on Xi but
∑ j=1 X j
not on Yi , β̂ is a linear function of Y.
E( βˆ | X1,
(b)
K, X ) = β + ∑
n
Var ( βˆ |X1,
(c)
n
i=1
K, X ) = ∑
Xi E(ui | X1,
∑in=1 X 2j
n
i=1
K, X ) = β since E(u |X , K, X ) = 0
n
i
Xi 2Var (ui | X1,
n
 ∑in=1 X 2j 
2
K, X ) =
n
1
n
σ2
∑in=1 X 2j
(d) This follows the proof in the appendix.
15. Because the samples are independent, βˆm,1 and βˆw,1 are independent. Thus
var ( βˆm,1 − βˆw,1 ) = var ( βˆm,1 ) + var( βˆw,1 ). Var ( βˆm ,1 ) is consistently estimated as [SE ( βˆm ,1 )]2 and
Var (βˆ ) is consistently estimated as [SE ( βˆ )]2 , so that var( βˆ − βˆ ) is consistently estimated
w ,1
w ,1
m ,1
w ,1
by [SE ( βˆm,1 )] + [SE ( βˆw,1 )] , and the result follows by noting the SE is the square root of the
estimated variance.
2
2
```