Linear Model Under General Variance Structure: Heteroscedasticity •

advertisement
Linear Model Under General Variance Structure:
Heteroscedasticity
• A Definition of Heteroscedasticity
In this section, we consider a special case of the model
y = X β + e, or yt = xt β + et, t = 1, .., T.
We relax assumption A4, while maintaining assumption A3.
§ That is, we assume that
E(et et*) = V(et) = σ2 wt2 if t = t*,
= Cov(et, et*) = 0 if t ≠ t*
where wt > 0, t = 1, …, T. While maintaining a zero covariance
across et’s (as implied by assumption A3), this allows for the variance
of et, V(et) = σ2 wt2, to vary across observations (which relaxes
assumption A4). A changing variance of et across observations is
called heteroscedasticity.
§ Given V(e) = σ2 ψ, this corresponds to the following specification for
the (T×T) matrix ψ
w12 0

2
 0 w2
ψ=  .
.

 0 0
. 0

. 0
. . .

. wT2 .
This implies that the variance of et, V(et), is proportional to wt2 (or
equivalently, the standard deviation of et is proportional to wt). In the
special case where σ2 = 1, then V(et) = wt2, and wt is simply the
standard deviation of et, t = 1, 2, …, T.
§ The matrix ψ being non-singular, we have
1/(w12 )
0

2
 0 1/(w2 )
ψ-1 =  .
.

0
 0
0 

.
0 
.
. , → P =

. .1/(wT2 )
.
1/ w1 0
 0 1/ w
2

 .
.

0
 0
0 
. 0 
. . 

. .1/ wT 
.
where ψ-1 = P' P.
§ Then, model M* becomes
y* = X* β + e*, P y = P X β + P e, or
 y1 / w1   x1 / w1 
 y / w  x / w 
 2 2  2 2
 .  =  .  β+

 

yT / wT  xT / wT 
 e1 / w1 
e / w 
 2 2
 . .


eT / wT 
This implies that the t-th observation (yt, xt) is weighted by the inverse
of wt.
§ The generalized least squares (GLS) estimator of β is
βg = (X' ψ-1 X)-1 X' ψ-1 Y
= [ Σ Tt=1 xt' xt/wt2]-1 [ Σ Tt=1 xt' yt/wt2]
= [ Σ Tt=1 (xt/wt)' (xt/wt)]-1 [ Σ Tt=1 (xt/wt)' (yt/wt)].
It is also called a weighted least squares estimator, where the weights
are the inverse of wt.
§ This is as a nice intuitive interpretation. Observations that have a
larger variance (i.e. a larger wt) are less reliable and are weighted less.
Alternatively, observations that have a smaller variance (i.e. a smaller
wt) are more reliable and are weighted more.
• Heteroscedasticity Case 1: Two Unknown Variances
Assume that the sample is partitioned into two sub-samples: T = Ta + Tb.
§ The first sub-sample has Ta observations:
2
ya = Xa β + ea,
where Ya is (Ta×1), Xa is a (Ta×K) matrix of explanatory variables, β
is a (K×1) vector of parameters, and ea a (Ta×1) error term vector
where ea ~ (0, σa2 I T ).
a
§ The second sub-sample has Tb observations: Yb = Xb β + eb,
where yb is (Tb×1), Xb is a (Tb×K) matrix of explanatory variables, β
is a (K×1) vector of parameters, and eb is a (Tb×1) error term vector
where eb ~ (0, σb2 I T ).
b
§ In the case where σa2 ≠ σb2, we have
Y  X 
Y = X β + e, or  a  =  a  β +
 Yb   X b 
 ea 
e  , where
 b
2
 ea 
 ea  s a ITa
E(e) = E   = 0, and V(e) = V   = 
eb 
eb   0
0 
.
2
s b ITb 
Using our earlier notation (where V(e) = σ2 ψ), let σ2 = 1. It follows
that V(et) = wt2 = σa2 if t belongs to first sub-sample
= σb2 if t belongs to the second sub-sample,
and
σ a2 IT
a
ψ= 
 0


.
2
σ b ITb 
0
§ Then, the GLS estimator of β is βg = (X'ψ-1 X)-1 X' ψ-1 y
X 
= [(Xa' Xb') ψ-1  a  ]-1 (Xa' Xb') ψ-1
 Xb 
 Ya 
Y ,
 b
= [(Xa' Xa)/σa2 + (Xb' Xb)/σb2]-1 [(Xa' Ya)/σa2 + (Xb' Yb)/σb2]
= [(Xa/σa)'(Xa/σa)+(Xb/σb)'(Xb/σb)]-1 [(Xa/σa)'(Ya/σa)+(Xb/σb)'(Yb)/σb].
3
• Parameter Estimation Under Case 1
§ Obtain the least squares estimator of β, βs = (X'X)-1 X'Y which is a
consistent estimator of β. Estimate the associated error term
es = Y – X βs which is a consistent estimator of e.
§ Evaluate σiu2 = (yi – Xi βs)' (yi – Xi βs)/(Ti – K), which is an unbiased
and consistent estimator of σi2, i = a, b.
§ Evaluate feasible generalized least squares (FGLS) estimator of β,
βfg = [(Xa' Xa)/σua2 + (Xb' Xb)/σub2]-1 [(Xa' Ya)/σua2 + (Xb' Yb)/σub2).
§ The above FGLS estimator βfg is a consistent, and asymptotically
efficient estimator of β, satisfying
βfg ≈ N[β, (Xa' Xa/σau2 + Xb' Xb/σbu2)-1] as T → ∞.
• Hypothesis Testing Under Case 1
Consider the hypothesis
H0: σa2 = σb2
Null hypothesis:
Alternative hypothesis: H1: σa2 ≠ σb2.
Recall that the sum of squared standard normal variables is distributed as
?2. It follows that, under the normality of e,
[(Ti – K)/σi2] σiu2 = (Yi – Xi βs)' (Yi – Xi βs)/σi2
2
, or σiu2/σi2 ∼
= ΣTt=1 (yi – Xit βs)2/σi2 ∼ χ (T
-K)
i
χ (2Ti − K ) /(Ti
– K), i = a, b.
§ Since σau2 and σbu2 are independently distributed, this implies
2
σ au
/ σ a2
2
σ bu
/ σ b2
∼
χ (2T
/(Ta − K )
χ (2T
/(Tb − K )
a −K )
b −K )
= F(Ta -K,Tb -K) .
§ With null hypothesis H0: σa2 = σb2, we have σau2/σbu2 ∼ F(Ta -K,Tb -K) ,
implying that, under H0, the test statistic (σau2/σbu2) has an F
distribution with (Ta – K) and (Tb – K) degrees of freedom.
4
§ This suggests the following test procedure:
q
Choose the level of significance α = P(Type-1 error) =
P(reject H0| H0 is true)
q
Find c1 and c2 satisfying:
α/2 = P( F( T −K ,T − K ) < c1], and
a
b
α/2 = P( F( T −K ,T − K ) > c2].
a
b
q
Do not reject H0 if c1 ≤ (σau2/σbu2) ≤ c2,
q
Reject H0 if (σau2/σbu2) < c1, or (σau2/σbu2) > c2.
• Heteroscedasticity Case 2: Multiplicative Heteroscedasticity
Consider the model Y = X β + e,
where E(e) = 0, and
E(et et*) = Cov(et, et*) = 0 if t ≠ t*,
= V(et) = σt2 = exp(zt α) if t = t*, t = 1, …, T,
zt α = [1 z t1
α1 
α 
n
. z tn ]  2  = α1 + Σi=2
αi zti,
 . 
 
α n 
α is a (n×1) paramater vector and the zti’s are explanatory variables.
§ It follows that
n
n
V(et) = σt2 = exp(α1 + Σi=2
αi zti) = exp(α1)⋅ Π i=2
exp(αi zti).
This expression indicates that the variance of et depends in a
multiplicative fashion on the explanatory variables zt. This is thus a
case of multiplicative heteroscedasticity.
5
• Parameter Estimation Under Case 2
§ Obtain the least squares estimator of β, βs = (X'X)-1 X'Y which is a
consistent estimator of β. Also evaluate the error term vector:
es = y – X βs which is a consistent estimator of e.
 es1 
§ Given es =  M  , regress ln(est2) on zt = [1
esT 
z t1 . z tn ] ,
t = 1, …, T, to
obtain the least squares estimator
αs = ( ΣTt=1 zt' zt)-1[ ΣTt=1 zt' ln(est2)].
The estimator exp[zt αs] is asymptotically proportional to σt2.
§ Evaluate the FGLS estimator of β,
βfg = [ ΣTt=1 xt'xt/exp(zt αs)]-1 [ ΣTt=1 xt' yt/exp(zt αs)].
§ The above FGLS estimator βfg is a consistent, and asymptotically
efficient estimator of β, satisfying βfg ≈ N(β, V(βfg)] as T → ∞.
§ The variance of βfg, V(βfg), can be consistently estimated by
G ⋅ [ ΣTt=1 xt' xt/exp(zt αs)]-1
where the coefficient of proportionality G is
G = [ ΣTt=1 (yt - xt βfg)2/exp(zt αs)]/(T-K).
6
Download