Document 11163280

advertisement
Digitized by the Internet Archive
in
2011 with funding from
Boston Library Consortium IVIember Libraries
http://www.archive.org/details/efficientivestimOOkuer
DEWEY
HB31
I'
.M415
working paper
department
of
economics
massachusetts
institute of
technology
50 memorial drive
Cambridge, mass. 02139
WORKING PAPER
DEPARTMENT
OF ECONOMICS
EFFICIENT IV ESTIMATION FOR
AUTOREGRESSIVE MODELS
WITH CONDITIONAL HETEROGENEITY
Guido Kuersteiner
No.
99-08
March, 1999
MASSACHUSEHS
INSTITUTE OF
TECHNOLOGY
50 MEMORIAL DRIVE
CAMBRIDGE, MASS. 02142
EFFICIENT IV ESTIMATION FOR AUTOREGRESSIVE
MODELS WITH CONDITIONAL HETEROGENEITY
Guido M. Kuersteiner*
First Draft: December 1996
Current Version: December 1998
Massachusetts Institute of Techonolgy
Department of Economics
50 Memorial Drive, E52-251B
Cambridge, MA 02215
Tel (617) 253 2118
'This paper
my
Ph.D. dissertation at Yale University. I wish to thank Peter C.B. Philhps for
I have also benefited from comments by Donald Andrews, Oliver
Linton, Chris Sims and participants of the econometrics workshop at Yale and seminars at Barcelona,
Brussels, Chicago, CREST, Harvard, LSE, Minnesota, MIT, Northwestern, Oxford, Pennsylvania, Princeton and Stanford .All remaining errors are my own. Financial support from an Alfred P. Sloan Doctoral
Dissertation Fellowship is gratefully acknowledged.
is
part of
continued encouragement and support.
Abstract
This paper analyzes autoregressive time series models where the errors are assumed
weak additional restrictions on their distribuUnder these conditions Quasi Maximum Likelihood estimators of the autoregressive
parameters are no longer efficient in the GMM sense. The main result of the paper is
to be martingale difference sequences with
tion.
the construction of efficient semiparametric instrumental variables estimators for the autoregressive parameters.
The optimal instruments
are a linear function of the innovation
sequence.
shown that a frequency domain approximation of the optimal instruments leads to
an estimator which only depends on the data periodogram and an unknown linear filter.
Semiparametric methods to estimate the optimal ffiter are proposed.
The procedure is equi\/alent to GMM estimators where lagged observations are used
as instruments. The number of instruments is allowed to grow at the same rate as the
sample. No lag truncation parameters are needed to implement the estimator which makes
It is
it
particularly appealing from an applied point of view.
1.
Introduction
This paper develops new instrumental variables (IV) estimators for autoregressive time
series models when the errors are uncorrelated but not independent. The specification
includes error processes which are conditionally heteroskedastic of
unknown
functional
form. Efficiency gains are obtained without having to specify a model for the dependence in
the errors.
The setup
is
general enough to account for stylized facts in
series displaying features
many economic
time
such as thick tailed distributions and time dependent conditional
variances.
Classical efficiency results for the quasi
autoregressive parameters such as
Hannan
maximum
likelihood estimator
(QMLE)
of the
(1973) depend on independence of the errors.
In the more general case of conditional heteroskedasticity considered here the
QMLE does
GMM
lowerbound for the asymptotic covariance matrix which now
no longer attain a
depends on fourth moments of the innovation process. In Kuersteiner (1997) it is shown
how a decomposition of the higher moment terms leads to a lower bound for the covariance
matrix. An instrumental variables estimator based on this decomposition is shown to
achieve the lower bound for the covariance matrix in the class of IV estimators with
instruments that are linear in the innovations.
The
feasible
of Hayashi
GMM
estimators developed in this paper are similar to the estimators
and Sims (1983), Stoica, Soderstrom and Friedlander (1985) and Hansen and
Singleton (1991, 1996). In this literature lagged observations are used as instruments to
account for unmodelled
MA(q)
innovations which lead to inconsistent
OLS
estimators.
Here lagged observations are used as instruments to account for unmodelled conditional
heteroskedasticity of the error terms which renders
OLS
inefficient.
Apart form the different motivation for the use of instruments, inefficiency versus inconsistency, this paper extends the previous literature as it explicitly treats feasible estimation.
The number of instruments in our case is allowed to grow at the same rate as the sample
size. This is made possible at the cost of an additional restriction on the fourth order cumulants compared to the treatment in Kuersteiner (1997). The advantage of making this
assumption is that the estimator can be implemented without the need for a truncation
or bandwidth parameter for the number of instruments used.
Since the optimal instruments are unobservable they need to be estimated nonparametrically. Assumptions about the generating mechanism of the volatility process or more
generally the dependence in higher moments are replaced by smoothness assumptions for
higher order cumulant spectra of the errors. This setup allows for the treatment of dependence in higher moments as a nuisance parameter. Nonparametric estimators of this
nuisance parameter are used to construct the optimal instruments.
Other semiparametric procedures proposed to handle conditional heteroskedasticity
include Robinson (1987) and Newey (1991). No parametric assumptions about the form
of conditional heteroskedasticity are made in these treatments. However, in order to estimate the conditional variance these authors have to assume independent errors. This
assumption has precluded direct application of their techniques to the stochastic conditional variance case. Hidalgo (1992) relaxes the iid assumption for the errors but has
to assume instead that the conditional variance is a smooth function of an independent
stationary process.
GLS
metric
Hansen (1995)
He assumes
framework.
Brownian motion
in a semipara-
that the conditional variance process converges to a
Sample path continuity of the
in the limit.
model
treats the stochastic volatiHty
limit process then allows for
consistent kernel estimation of the conditional variance.
Hansen (1985) and Hansen, Heaton and Ogaki (1988) prove existance
lowerbound in the presence of condimensional
and
nonlinear
heteroskedasticity. The high
character of these instru-
More
generally
of instrumental variable estimators achieving a
ditional
GMM
ments has so far precluded implementation of such an estimator.
Here we limit ourselves to the implementation of optimal procedures in the much
smaller class of IV estimators with instruments that are linear functions of the observations. Ordinary least squares is a particular member of this class. In the general case of
conditional heteroskedasticity it is inefficient. This paper achieves the construction of a
feasible version of the most efficient IV estimator with hnear instruments.
The remainder of the paper is organized as follows. Section 2 describes the model assumptions. Section 3 develops the
domain approximation
for the
IV estimator
efficient
IV estimator
is
for the
AR{p) model.
derived in Section
4.
A frequency
Section 5 shows that a
semiparametric estimator with the same optimality properties can be constructed. Some
Monte Carlo simulations are reported in Secition 6 and concluding remarks are made
Section 7. Some important lemmas are quoted in Appendix A while the main results
the paper are proved in Appendix B.
in
of
Model
2.
We
by defining the stochastic environment of the model. Let {Q ,J^ ,P) he a. general
probability space and define a filtration J^t to be an increasing sequence of cr-fields such
that J^t ^ -^t+i C ^ V i. There is a doubly infinite sequence of random variables generating
the filtration Tf We assume that we observe a sample of size n of a univariate time series
yt where i = {1, ...,n}. More specifically, we assume that yt is generated by the following
autoregressive model
start
<l^{L)yt
where
(p'
is
=
et is
(01,
=
St
(1)
a martingale difference sequence generating Tt- Here
is
..., (f)p)
assumed that
the vector of parameters describing the
has
(l){L)
the parameter vector
(p.
all
roots outside the unit circle.
The martingale
correlation between the errors. However,
Rather we allow
for
dependence
(f){L)
= l — (^iL—... — (f)pU'.
mean equation
We
of the model.
It
are interested in estimating
difference assumption for et implies absence of
it is
in higher
not assumed that the errors are independent.
than second moments to account
for thick tails
and conditional heteroskedasticity.
Assumption A-1.
(T^,
(J
E {et) = cr^ <
(s)
<
DO for
E [e'te't^s) >^
Remark
s
>
(i)
oo,
et is strictly
(ii) (f>{L)
0, (iv)
some
a>
has
stationary and ergodic,
all
roots outside the unit
E [etet-s^t-r) =
for all
E {et
for s
^
r,
s,r
>
\
J^t-i)
= 0,E [e^
circle, (Hi) E{e'te't_s
0, (v)
^
\s\
\a {s)\
|
"
J-'t-i)
^'^)
=B<
=
oo,
s.
Assumption (iv) is added to the assumptions in Kuersteiner (1997) in order
to simplify the form of the optimal instruments. It is somewhat restrictive as it rules out
1.
-
some nonsymmetric parametric examples such as EGARCH. The IV estimators proposed
in Section 3 are still consistent and asymptotically normal if (iv) fails. However, in this
case they lose their optimality properties.
Remark
2.
It
should he emphasized that no assumptions about third moments are made.
In particular this allows for skewness in the error process.
Remark
GARCH
Asssumption (A-1) under certain conditions.
Nelson (1990) obtains sufficient conditions for stationarity and ergodicity of the GARCH(1,1)
model. The martingale difference property follows immediately from the definition of
a GARCH process. Assumption (A-liv) is shown to hold for the ARCH{p) case in
Milhoj (1985) for symmetric innovation densities. The same argument extends to the
GARCH (p.q) case. If the innovation distribution is normal then fourth moments are
known
3.
to exist for the
(p,q) processes satisfy
GARCH (1, 1)
+ 2'^if3i + 0\ <
case if 87!
1.
This condition
is
valid
and thus covers the ARCH case. In Milhoj (1 985) and BoUerslev (1 986) the autocorrelation structure a (s) is shown to be identical to the AR{p) and ARMA{maoc(p, q),q)
case for ARCH{p) and GARCH {p,q) respectively. This implies that the summability
condition holds if fourth moments exist. Similar arguments can be made to show that
stochastic volatility models satisfy the assumptions.
for
P=
,
Based on the results in Kuersteiner (1997), we will now introduce the optimal instrumental variables estimator for the AR{p) model. The estimator is constructed by reweighting the innovation sequence by the unconditional fourth moments a (k) + o"^ of the error
process.
these
3.
Without parametric assumptions about the form of conditional heterogeneity
moments
typically have to be estimated.
Instrumental Variables Estimator
The parameter
in this
paper
vector
can be consistently estimated by OLS. Under the assumptions
OLS amounts
to an inefficient
IV
estimator in the class of
IV
estimators
with linear instruments. Kuersteiner (1997) derives the form of the optimal instrument as a
function of the fourth moments and the impulse response function of the underlying process
yt in
a
more general
context. In this section these
more general
results are specialized to
the autoregressive model.
Let zt G M.P be J^t-i measurable and square integrable, strictly stationary and ergodic.
Then, the instrument satisfies the moment condition E [(</> (L) yt) zt] = where (L) yt =
An instrumental variables estimator then
Et- Let y't = [yt,yt-i, ...,yt-p] and </> = [l, —<p']
(f>
.
is
defined as
=arg min
f=P+i
with an explicit solution in the case of an autoregressive model as
<t>=[ZY.i]'^ Z'Y
(1)
=
where Y'
Y=
that
[yi+p,
.
.
.
,yn]
+ e. By
Y-icf)
and
y!_-^
=
[yp,-.,yn-i] where
the ergodic theorem
As argued before instruments
follows
it
which are linear
zt
y'j
is 4> is
=
[yt,yt-i, ...,yt-p+i]
Note
consistent.
in past observations are considered.
Such instruments are equivalent to instruments that are linear functions of past innovations. This set up is formalized here. Using (f)~ (z) = X^jIq^^J-^''' *^^ time series yt can
be expressed as a linear filter of past shocks yt = Yl%.o''P(i>,j^t-j- Also, let
=E
ak
and
b'f^
=
(V'</.,fe-i,
with
,'4'4,,k-p)
In Kuersteiner (1997)
the z-th element of
it is
ip^^^
= ^ (f^) + (^^
(e^tetk)
=
afc^ of a^ Yl'kLi Wk,i\
k
for
shown that
for Zf
<
<
=
Then
0.
(2)
define
Yll^=i '^k^t-k
P^ =
[6i, ...,6^].
with a^ G
W such
ooVi the asymptotic distribution of
is
that for
normal
with mean zero and covariance matrix
—KX)
771
where
Am =
[o-i,
...,0^].
The
lower
bound
S=
where
It
O'k
=
is
\im{^' Pln^;n'Pm)-'
(3)
should be noted that under Assumption (A-1), Qrn = Dio-gioLi, ...,0^)then follows immediately that the optimal instrument achieving (3) has weights
it
A
^k/oi-k-
different
almost surely. This
is
way
of writing the optimal instrument
a special case of the
The optimal instrument matrix
Z
more general formulation
now
=
is zt
defined by setting z[
lim^^oo Pm^m^T'
in Kuersteiner (1997).
=
[zt,i, ...zt-p+i^p] with
t
It should be emphasized that the
Zt+i,k = Y^%zO'sr^^t-i and letting Z' = [zp, ...,Zn-i]
k -th instrument, i.e. the instrument for parameter (pj^ is lagged by k periods and has a
convolution filter which is also shifted by k elements.
is
.
The estimator introduced
be a function of the seqeunce
Define
(f)Q.
zt
=
not feasible and needs to be approximated. Let ho
{fcfc/ctfc}fc^i of optimal weights and the true parameter value
so far
is
Yfk'Ji'^ h/otk^t-ki^o) with
approximate version of
4> is
et{(f)o)
=
(p'oYt
and
let
Z=
[zp,...,Zn-i]
The
now
^(/io)=(zV_i)"'zV.
(4)
A feasible estimator is then obtained by replacing Hq with a consistent estimate h. The proof
of feasibility proceeds in two steps. In Section 4
and
in Section 5
we
establish that ^yn{^{h)
—
it is
4>{^o))
In the next two sections the Approximation (4)
domain
shown that
y/n{(f){ho)
—
</>)
=
Op(l)
=
Op{l).
is
represented in terms of frequency
This imposes no practical limitations since the freqeuncy domain version
by the exclusion of only a few observations at the beginning of the sample.
The difference is asymptotically of order Op{n~^) and therefore does not affect the first
differs
integrals.
from
(4)
order asymptotic properties of the estimator.
Formulating the estimator in the frequency domain however
offers the potential for im-
provements in terms of computational efficiency since the frequency domain formulation
6
can be used to implement FFT-algorithms. Using FFT algorithms reduces the computational complexity of ^{h) from 0{'n?) to 0(n log n). This improvement is substantial
in applications where the dataset is extremly large and the estimator is computed many
times.
Leading examples are model selection procedures
in forecasting applications
and
simulation studies.
4.
IV Estimation
in the
In this section a frequency
Frequency Domciin
domain approximation
to the optimal
IV estimator
Consider the inverse of the spectral density for the AR{p) model f~y (A)
and
let
[(/((e*"^)]
useful later
and
=
9yy{^>
'^)-
The
definition of the
=
is
^
derived.
|(?!)(e*^)|
spectrum of the squared errors
will
be
is
oo
/,2,2 (A)
=
(27r)-i
^{k)e--i\k
J2
(1)
fe=— oo
Define the lag operator a(A)
=
[e'^, ....,6*^^]
by a(A)*. Also introduce the matrix
A (A) =
and denote the complex conjugate transpose
a(A)*a (A)
.
Then
gyy{(t>, A)
can be represented
as
'
9yy(.(t>A)
1
a(A)
a{Xy
A{X)
Define
=
77(0, A)
d\ngyy{^,\)
(2)
and introduce the discrete Fourier transform of the data
the periodogram as In,yy (A)
With
these definitions
=
1^^;^
we turn
(A)
The main
Re
[In,zy (A)
-y= Y17=i Vt^
a
-itX
and
.
|
domain implementation of the
the beginning. It is easy to show that
1
=
=
to the frequency
mental variables estimator introduced at
4>
as cOy (A)
(A)]
d\
instru-
-1
Re[In,zy{\)\dX
+ Op{n-^)
(3)
contribution of the paper consists in deriving an approximation to this frequency
domain formulation that does only depend on observable quantities. In a first step we
decompose the instrument into observable data and an unobservable optimal filter. In the
next section it is shown how the unknown optimal filter can be consistently estimated.
For the purpose of this and the next section, we introduce the spaces L'^ [— 7r,7r] of
functions / [— 7r,7r] —> C^ such that J" |/| d\ < oo. Also, define the spaces C^ [— 7r,7r] of
functions / [— 7r,7r] —>
such that f is k times continuously differentiable. Throughout,
the function Rn (A) will denote a generic remainder term whose definition can change. We
start by approximating the discrete Fourier transform of the instrument variables ztj
:
:
W
^., (A)
=
l^,k (A)
e^^V(e-^^)a;, (A)
+ i?^,^ (A)
(4)
_-i
= ^-„ ^e-^>0+^) and^i?^^^) ^ „-i/2 ^oo^ _^^-,A;^^^^.
^^^
and ^{k) = {i/jQ/akjipi/ai+k,
Then let the
define the sequences ^ = {ipQj'ipi,
where
'A,
/^., (A)
_
}
}
j^^^
inner
product
= -^ Yl'^o ^J^
Rl^^s (A)
Unj (A) = Ylt=i-j ^te"^^
a compact form
—
^^^^n,j (A)
Y17=i ^te^^^-
We
be defined
for a
=
••'
Kmp)^^^^''
{sq, si,
...}
R^
f.
where
(A) in
=
[i?^
^
^^ ^^ analogous way u^^ (-A)
(A)
is
,
and
...i?^p (A)]
the complex conju-
(A) as
l.^
.
=
+ RriMk)^>')-
denote the vectors of stacked error terms by -R^ (A)
[^n,^(i)^^)'
K,i,^^^
gate of uJzi^ (A) Define
s
can now write the remainder term
-iX
<fc(A) = -W(A)0(e- )K,i,W
We
sequence
^^,1 (A)
=
l^ (A)
(5)
't/i.p
The
properties of
/,/,
determine the asymptotic distribution of the instrumental
(A)
The next lemma
ables estimator.
(A)
gives a representation of
operators. This shows that the smoothness of l^ (A)
vari-
terms of convolution
inherited from the smoothness of
AR(p) spectrum.
the
Lemma
4.1. Let l^^k (A)
sumption (A-1) and
assume that
Irj
is
(A) in
/,/,
=
(A)
-t-
1^
1
^^'^^ otk
—^
a{z)* has
= ZT=-oo ^j^''^'
(A)
all
{fa * v) (A)
and
-I-
-^r)
{(f),
A)
.
8.10, that
Ir,
(A)
G
.
Also
Then
^
= (^ - ^)
^^^h a,
The properties of
For fa & L} [— 7r,7r] and
fj.
satisfying As-
et
characteristic roots outside the unit circle.
Using the convolution operator
4.
,
.
B
Proof. See Appendix
Remark
(e?e?_fc)
(—A) can be represented as
J —IT
where fa
= E
being the coefficients of the power series expansion of 4>{z)~
ipj
=
(f){z)
l^ (A)
= E^o ^1^7^"'^^^'^''^
C'' [-7r,7r]
rj
E C^
Irj
*
a compact notation for
(A)
[—-n,-n]
can
it
l^i
(A) is
now be determined from
FoUand
follows from
Ij^
(A)
=
those of fa
(1984),
Theorem
oo
implying that J] j=0
\j\ < OO. While fa e L^ [-7r,7r] is
«fc+j
it is not necessary. Alternatively if a^ > a >
for some
sufRcient to obtain this result
a and
all
k then sup
a^.
|
|
< a
^
<
oo and Yl'^o
These arguments show that
central limit theorem in Appendix A.
if
f]
E
C"^ [— 7r,7r]
The
DFT
.
Irj
oik+j
\j\
< a-'^T=okj\\j\ <oo
(A) is sufficiently
smooth
to apply the
representation of the discrete Fourier transforms of the instruments in terms of the
for the
data allows to obtain a frequency domain version of
4>
without the need to go
through an explicit calculation of the instruments in the time domain. The approximation
relies on the fact that convolutions in the time domain are transformed into multiplications
8
=
domain and the
by (f){e'^^).
in the frequency
multiphcation of
The
fact that the residuals
W
lOz,,
by k periods leading to
transforms in the following
u,
^z, (A)
cl^{e-'^)uy (A) l^ (A)
(A)
,-tAp
where Rn(X)
=
e~*^'^w^^ (A)
way
-iX
a (A)* ©i?^ (A) and
©
is
the element by element product.
=
=
e'^PcOz,
where the symmetry of
Rn{X),
(6)
The correspond-
is
(-A)
e'^cvz,
u>z (A)*
+
'^z, (A)
ing expression for the conjugate transpose of Uz (A)
+ R^{-X)
4,{e'^)u;y{X)*l^{-X)
(7)
(-A)
Z^ (A) is used.
discrete Fourier transform approximation for Zt introduced at the beginning of
produces an asymptotically equivalent estimator based on the periodogram of
this section
yt
is then obtained from (4)
and stacking the resulting
discrete Fourier transform of the instrument matrix
by lagging each
The
can be computed by a simple
ujy (A)
and an unknown
It is
filter.
convenient to define
/i^((^,A)
= Re
l^{-X)(t>{e'^)a{-\)
and
h{(l),X)
By
=Re
l^{-X)<P{e'^)
.
substituting for equations (6) and (7), (3) can be approximated by
-1
kho) =
^n,yy (A) /l^ (^, A)
dX
In,yy (A)
/
h{(f),
X)dX.
(8)
J— IT
It is
shown
in the proof of Proposition (4.2) that 0(/io)
1
(t){ho)
-
<^o
=
—
4^
=
Op[n
-^Z^)
and
-1
•n,yy (A)/i^((^,A)dA
J
In,ee (A)
Ir,
(A)
dX
+
Op {n'^'^
such that consistency follows from ergodicity and the fact that
E
f
In,ee (A)
Ir^
(A)
dX
=
a^ f
J —TT
transparent from equation
/^
(A)
dX
=
0.
J —IT
depends
on knowledge of the true parameter values and the correlation structure of the squared
errors. Feasible versions of 4>{ho) will be discussed in Section 5 below.
Under the assumption that the weight matrix Re [Z^ (—A) (e*'^)] is known, the asIt is
(8)
that ^(/iq)
is
infeasible as
it
stands, since
it
</>
ymptotic distribution of
(A. 4).
The next
^(/lo) is
now a
straight forward consequence of
proposition summarizes this result.
Lemmas
(A. 2)
and
Proposition 4.2. Let (f>{L)yt = et where all roots of (f)(L) are outside the unit circle. If
et satisfies Assumption (A-1) then for ^ defined in (1) and 4>{ho) defined in (8) we have
4){ho) — 4> = Op (n"-^/^) and
^(4>iho)-<Po)^N{0,a-^E)
where E
is
defined in
(3).
Proof. See Appendix B
The remainder of the paper
will
now be concerned with
parametric estimator with the same distribution as
5.
the construction of a semi-
(f).
Adaptive Estimation
To develop a feasible efficient IV procedure,
Re [l^ (A) (p (e~*^)] a^^d h^ {(f>, A) = Re \l^ (A)
cf)
has to be established that h
it
{&~''^)
is
called adaptive.
No
apart from efficiency issues,
is
A
semiparametric
confusion should arise between this
use of the terminology and the literature on feasible local
Bickel (1982), Kreiss (1987), Linton (1993)
=
a (A)] can be replaced by consistent
estimates without affecting the limiting properties of the estimator.
estimator having this property
A)
(cf),
minimax estimators such
and Steigerwald (1994). The main
as
difference,
the fact that here a nonparametric correction to the crite-
is made while the local minimax
Newton Raphson improvement to a consistent
rion function
literature
first
makes a nonparametric one step
stage estimator.
Different approaches to prove adaptiveness are used in the semiparametric literature.
is used in Robinson (1987,1988) in the context oiiid models and partially
models and by Hidalgo (1992) in the context of time series regression models. Newey
(1991) applies similar techniques as Robinson (1987) to the instrumental variables case for
iid data. Andrews (1994) develops a general methodology based on stochastic equicontinuity arguments and applies it to the partially linear framework. Andrews' approach will
Direct calculation
linear
be used here to break the proof into two parts. First, it is shown that a nonparametric
estimate h{^,X) converges to h{(f)Q,X) uniformly with probability one. The second step is
to established that uniformly in a shrinking neighborhood of the true filter h{(pQ,X) the
distribution of an estimator is arbitrarily close to the distribution of the estimator based
on the true filter.
This argument will now be formalized. Let
where C is the complex plane. Then introduce a
n = |/i:
[-7r,7r]
/i
= Re
L (-A)
(^(e'^)
l^
:
[— 7r,7r]
—
>
C^ and
set of functions
4>
'
H defined
;Re[Z^ (-A)] ,Re ^ (e'^)
["TTj^] ~^
C
as
e C^ [-n
,
Tr]\
.
(1)
where
C'^ [— tt,
tt]
the L°° Sobolev
denotes the space of k times continuously differentiable functions. Define
norm
of order one as
1=
sup
11/
(A) 11+
sup
A£[-7r,7r]
AG[-7r,7r]
10
I^W
where ||.||
on 7^ as
is
the Euclidean matrix
norm
=
p{hi,h2)
{HjP)
is
a complete metric space.
Lemma
We proceed
(4.1) that
l^
A)
((/>,
G C"
defined by
-
||i,/,,i
If
.
+
li,,2\\l
—
1101
(trAA*)
-
'
.
Introduce the metric
?f'2lll
given by (1) and
is
(f)
[-tt,it]
\\A\\
/,/,
by
(5)
then
it
follows from
Therefore h{(t>,X) G H.
by defining the estimator for h (</>, A) We have established that we can
= {Y_iY-i)~^Y-iY or from its frequency
obtain a consistent estimate 4> for example from
function of some fixed parameter value
before.
Residuals
as
a
domain analog introduced
(p are obtained as in Kreiss (1987) from
.
cf)
e«
{<f^)
=
such that the estimated error
measurable part (0
—
et
£t
((/'o)
{(f))
n
(</>- 0o)
-^yt-p)
{yt-i,
We
form the following
t=p+k+l
else
dr,
>
where the sequence d„
for all
n with dn =
O {n
^1'^+^) for
is
<
some
z^
<
1/2.
The
dn are used to avoid "too large" values for a^^(^). Truncation was
More
introduced by Bickel (1982) in the context of score estimation.
our context
J-'t-i
statistics.
=
oikQ))
truncation numbers
(2)
can be decomposed into the true error and the
,yt-p)-
{yt-i,
(po)
+
closely related to
Hidalgo's (1992) semiparametric frequency domain estimator.
Simulation
experiments indicate that the truncation playes no role in practice and can therefore be
ignored in applications.
Next, an estimate for b^
=
(27r)~^
J^^
A) e^^^dX
77 {4>,
is
The
needed.
vector hk contains
we
the impulse response function of the AR{p) model evaluated at different points. Here
want to express
definition of
6fc
77
b^ directly as a function of the underlying yli?-parameters.
(0, A) in (2)
4)~'^
and the expansion
{z)
=
with ip^j
Yl''p4>,j'^^
=
From
for j
the
<
0,
can be written as
bk
'^4>,k-p
where the
all s
>
coefficients ip^^j satisfy the recursion ip^^
and
tp^^Q
coefficients of the
=
1
Let
(see Kreiss, 1987).
polynomial expansion of
(f)~^
ifJ^
(z)
.
—
(pi^p^^s-i
This vector
"
0<^.o
<^1
1
<?^2
-01
—
(t>p4'4,,s-p
is
=
first
^o^
p
+
1
the solution to
"
"
"
1
—
denote the vector of the
1
^<t>A
(3)
1
-
11
.
^*,P
.
_
_
which
denoted by
is
defined in
pxp+1
(3).
t/)?
Then,
=
^~^ei where
ei is the first unit vector
bp+i denote the vector of coefficients
let
selector matrix ^i picking the last
p elements from a p
and
^o
^p^^2
+1
x
1
$
is
the matrix
'^4>,p+\-
vector
Using a
we have
1
hp+i
=
Si^l
=
r^5i$-^ei
1
<t>V
where we define T^ = (/>! if p = 1. In a similar way we obtain bp+s = T^Si^ ^ei. The
vectors h\
.b^ can now be expressed as functions of the underlying parameters by
.
.
bk
where the convention T9
=
with the indicator function
coefficient b^
is
Ip
=
is
{.}
=
5p,fcr("^"'t°''^-^1^5i$-iei
assumed and the
1 if
selector matrix Sp^k
the expression inside the bracket
continuous in the underlying parameters for
all finite
is
is
defined by
true.
The
Fourier
k and can therefore
be consistently estimated from a consistent estimate ^.
A nonparametric estimate of h (cj), A) is now defined as
n—p—l
and
hr (^,.A
No
=Re
(A)0(e
-i\
(4)
is needed. The reason is, that h (</>, A) is already a conbounded sequence and a twice continuously differentiable function. In
implicitly contain a bandwidth since for every
inside the stationary region
additional kernel smoothing
volution between a
fact,
the bk
(f)
they will decay to zero quickly.
We
need the following matrix h^{4',X), whose elements are continuous func^) and which is defined by
will also
tions of
hn{4'-,
h^^
The
(^, a)
= Re
[l^ (A)
^ (e-^^) a (A)
success of a semiparametric estimator depends on the ability to uniformly estimate
the weights aj^. Additional assumptions about the
moments
of the driving error process
on fourth moments such conditions necessarily
involve higher than fourth moments. Here we prove uniform convergence by a mean square
argument which necessitates summability assumptions on eighth moments. The following
assumption is sufficient to prove the main result.
are needed to assure this. Since aj depends
12
Assumption B-1. Let
Cs.,.e (ii,
.
.
.
,
ifc-i)
he the k-th order cumulant of the error process
Then
et.
^(1 + \tj\) \c,,„, {ti,..., tk-i)\ <
•
J]]
tl
=
oo, for all j
1,
...,
k
-
1
and k
=
2, 3, ..,8
t-k-l
Assumption (B-1) implies that higher order cumulant spectra
assumption enables us to state the following
Proposition 5.1. Let hn{4>n^ -^) he as deRned
assume that ^„ -^ (pQ in probability. Then
sup
of order eight exist. This
result.
Assumptions (A-1, B-1) hold and
in (4), let
hn{(t>n^>^)-h{(t)Q,X)
=Op(l)
AG[-7r,7r]
^
n
—
Proof. See Appendix B
We proceed to define the semiparametric estimator ^„{hn) by replacing Hq
with a nonparametric estimate (4) We will establish that
=
n
as
P
Also
oo.
[hn{(l>n->
P
(p
^'H] -^
^)
1
f
/i„((/)„,
as
A),
/i
((/>o,
A)
>^) —
any
for
>
)
(5
>
as
>
oo and
^ oo.
n
/i((/)q,
A)
.
Vn
By
applying
Lemma
(A. 2)
In,yy (A)
h
it
((/),
(^^„(^n)
-
(ho))
4>n
=
can be shown that ior h
A)
=
In,yy (A)
Rc
Op (1)
(5)
.
eH
[l^ (A) (/.(e-'^)a (A)]
+In,ee (A) h^^ (0, \)
+
h
((/.,
(6)
A) i?„ (A)
where the remainder term Rn (A) is such that ^Jn f Rn{X)'^ (A) dX = Op
uous function (A) with absolutely summable Fourier coefficients. Let
(1) for
(7)
any contin-
<;
h^^
such that h^^
{(pQ,
A)
then follows
= Re
= Re [l^ (—A)] =
V
(5)
icf>,X)
(^, a)
Ir/
= Re
[l^
(-A)
(A)
and
[i^
(-A)
4>{e'^)<t>^\e'^)
4>{e'^)<Po'i<^'^)
if
£^ In,yy (A)
(A^(^„, A)
-
h^
(</>o,
A)) dX
=
Op{l)
(8)
and
j^In,e.{X){h^^Q>^,X)-lr,{X))dX
=
Op{l)
/l„(A)(/i„(^„,A)-/i(<^o,A))c!A
=
Op{l).
13
(9)
(10)
(8)
can be established easily with the help of Proposition
^"
-
/„,,, (A) (/i^(,^„, A)
<
sup
r In,yy{X)dX
/l^(0„,A)-^-((^O,A)
J —n
,^A^
sup
2
Ae[-7r,7r]
last
h
=
=
{a a){b b)
expression goes to zero by (5.1)
(9)
we work with
hn{4>n, A),
is a positive scalar and the second
where a and b are two conformable vectors. The
and the fact that sup;^£t_ ^.^i ||a(A)|| is bounded. To
inequality uses the fact, that In,yy (A)
first
inequality uses tr{ab ba)
prove
||a(A)||%j,(0)^0
sup
/V(-A)^(e^^)-Z^,o(-A).^o(e
Ae[—tt.tt]
where the
by the following argument
h^ (00, A)) d\
Ae(-7r,7r]
<
(5.1)
0^
Vn
=
cP{e'^),
=
(h)
the metric space {H,p) defined in
= Me'^)
ct>^
y/n
(1).
Also
let
Hq
-
Hq)
dX
=
h{(j)Q,X)
,
and
In,ee (A) {h^^
{(f),
A)
-
/,,,o)
+ Rn (A)
(h
J —TT
for
(5
h G H. Following Andrews (1994),
>
any given
??,
>
e
limsupPf
Vn (hn)
-
Vn (ho)
>
< limsupPf
Vn (hn)
-
Vn (ho)
> &,hn€H,p{hn,ho) <6
^
n—>(x>
+ lim sup P
<
there exists a
1?)
'
(hn ^ 7i oi p{hn, ho)
sup
\\vn {h)
>
6
— f„
(/io)||
>
>
=0.
i?
|
<
£-
\h€n,p{hr,M)<i
we have established
in Proposition (5.1) that
lim sup
n—too
Therefore
^
hmsup P
^-"^
Since
(9) follows if for
such that
P (Jin^H
V
or p (hn,hQ]
^
6]
^
if
limsupP
"-°°
sup
ll^n (/i)
—
'"n (^o)ll
>
<^
'i^
(11)
I
yft.e-H,p(/i„,/io)«5
then the following theorem can be established.
^
Theorem
5.2. Let hn{^ny^)
defined in (4). Let assumption (A-1) hold and let ^„
be a previous estimator for which ^„ —> (pQ in probability or almost surely. Then, the
semiparametric estimator 4>(hn) defined by
^ [hn
=
In,yy{X)hlQ)^,X)dX
)
•TT
/
J
14
./— TT
/„,yy (A) /i„(^„,
A)dA
.
has a limiting distribution characterized by
^
(4,{hn)
-
4>o)
^N{0,a-^E)
.
Proof. See Appendix B
This result establishes the
feasibility of
a semiparametric estimator that improves on
the efficiency of the conventional Gaussian estimator in the presence of higher order depen-
The frequency domain
dence.
representation allows to avoid estimating the instruments
each observation in the sample. Instead an optimal
for
filter
applied to the periodogram
of the data leads to an asymptotically equivalent procedure. Moreover, the fact that the
optimal
a convolution integral in the frequency
filter itself is
domain
solves the
problem
of truncating the approximation of the optimal instrument at a given lag in a natural
and elegant way and eliminates the need
for lag
truncation parameters for the number of
instruments used.
Monte Carlo Simulations
6.
Monte Carlo experiment is reported. To keep the exposition as simple
on an AR{1) model. We consider what the efficiency gains/losses of
In this section a small
as possible
the
IV
we
focus
estimator are relative to a correctly specified likelihood procedure and relative to
OLS.
The
following questions are of interest:
estimator achieve efficiency gains,
how much
how
Under what circumstances does the optimal IV
QMLE and
big are they relative to the Gaussian
by not specifying the true likelihood. These questions are analyzed for
mechanism is an ARCH{1) process.
generate samples of size n = 256, n = 512 and n = 1024 from the following model
is lost
the case where the true generating
We
yt
where
Ut
~
et is
generated by the
ARCH{1)
A''(0,1). Starting values are yo
=
=
+ et
(fyyt-i
process
and
ej
eq
(i)
= Ufh/ where ht = 'Yq + 7i£t_i with
= 0. Small sample properties of three
be defined below are evaluated for different values of 0, 7;^ G [0, 1)
from Milhoj (1985) that asymptotic normality established in previous chapters
different estimators to
It is clear
only obtains for values of 7^ € [0, -^1/3). Nevertheless, simulation results are reported for
parametrizations outside this interval in order to analyze the robustness of the proposed
IV
procedure to departures from the assumptions. The parameter 70
is
fixed at
.1
for all
experiments.
The parameter
is
(f)
^
mator
is
is
The
least squares esti-
=
Yl^=2ytyt-i/Y17=2yt-i- ^^^ optimal instrumental variables
obtained from the consistent first stage estimator 4>oLS ^^
denoted by 0„
estimator
estimated by three different estimators.
OLS
IT
In,yy (X) h"" {4)QLS, >^)dX
15
/
J — -K
In,yyWH^OLS^^)d>^
where h^
ated by
and
A)
{(f),
h{(p, A)
computed
are
the Ukelihood estimator
(1),
obtained from maximizing
is
(j)^
7i;n =
70,
/(</.,
as explained in Section 5. If the data are gener-
-^E (in
and /it = 7o + li^l-iEngle (1982) to maximize the likelihood.
with
St
=
yt
—
We
<t>yt-i
Figure
Figure
QMLE
1
shows the potential
+
^t
use the
(2)
^)
BHHH
algorithm described in
1
efficiency gains of the
as a function of the autoregressive parameter
IV estimator relative to the Gaussian
</>.
The
efficiency gains are
from the asymptotic covariance matrix when the generating mechanism
plicitly,
the asymptotic covariance matrix oi
c^Ils (</',7o,7i)
=
=
is (1).
computed
More ex-
can be expressed as
(p^
^^-^ E'^'^^.+i
(3)
7i)^ and a,+i = 27§7'i+V[(l - 7i)^(l - 37?)] + o"^. The asymptotic
covariance matrix for the optimal IV estimator can be obtained from (3). It is given by
where a^
(70/I
-
fv-i
^/y
(<?!',
7o,7i)
=
"I
-1
-^E^'V\
(4)
i=0
Figure
plots cr^y
1
(</>,
.1,7;^)
/cr^^^
(c;!),
.l,7i) for
cf)
6
[0,1)
and
different values of 71.
' OLS
These theoretical gains are contrasted to the empirical efficiency of the estimators 4>n
and 4>n based on 3000 replications for sample sizes 256, 512 and 1024. The results
(j)^
are summarized in Table 1.
Table 1
,
As expected,
parameter
sive
is
gains for the
above
.5.
IV
estimator are achieved for models where the autoregres-
This conforms with the theoretical analysis based on asymptotic
approximations. For the sample sizes considered here, the theoretical efficiency gains are
not achieved completely.
The
improves with the sample
to 512. It
is
The most
significant increase takes place
also interesting to note that the
for values of 7^
dependence
table shows that the relative efficiency of the
size.
>
IV
procedure maintains
\/l/3. In fact the gains are strongest
when both
ARCH
estimator
size
256
properties even
autocorrelation and
in the conditional variance are strong.
Figure 2 shows the empirical densities of the three estimators
no
its
IV
from
4)^
,
4>n
^^d ^„
when
effects are present.
Figure 2
The graph confirms the information summarized
identical
and
(^„
in the tables:
The
three estimators are
under iid conditions. Figure 3 shows the empirical distributions of
= .9 and 7^ = .9.
for a sample size of 1024 when
</>
Figure 3
16
4>n
>
4'n
Here 0„ clearly dominates the two other estimators in terms of efficiency and mean
and median unbiasedness. The IV estimator has surprisingly good properties even though
the asymptotic theory used for
construction does not hold for this set of parameter
its
values.
Table 2 contains the means and medians
on 3000
is
bias tends to be largest for the
(f).
For a fixed
other hand,
7.
4'n
^'^'^
^n
when n
=
512 based
is little
largest
(p, it is
affected
2
IV estimator, but the
smaller than the difference of the former with
with
^^^
,
replications.
Table
The
^^^^
^„
for
when
=
7^
4>^
.5.
difference
between
(/>„
and ^„
The bias for 0„
and ^„ increases
The bias of the ML estimatoi, on the
.
by the parametrization of the model.
Conclusions
This paper develops
efficient
IV estimators
for autoregressive
models with martingale
novations. Popular parametric examples of such innovation processes are
A
and stochastic
volatility models.
these effects in
many macroeconomic and
The paper shows
vast empirical literature
that
GMM
OLS.
ARCH, GARCH
documents the presence
of
financial time series.
that estimation of the autoregressive parameters by standard Gaussian
ML techniques leads to inefficient estimators. An important result
inate
in-
in Kuersteiner (1997)
is
estimators based on lagged instruments improve efficiency and therefore domIt is
shown how
to construct the best
GMM estimator
based on instruments
that are linear in past observations.
A common
GMM
problem of
procedures based on a large number of instruments is
bad small sample performance. This is mainly due to estimates of a high dimensional
weight matrix. We introduce a novel decomposition of the weight matrix which leads to an
their
orthogonahzation of the instrument space. This decomposition has the advantage that
it
implemented by using FFT algorithms. Moreover, the
estimator does not require a bandwidth choice for the number of instruments used and is
can be computationally
efficiently
therefore straight forward to use in practice.
The
small sample properties of the procedures developed are very promising.
mators are equivalent to conventional
OLS
The
esti-
even when the innovations are iid and strictly
dominate OLS when the innovations are conditionally heteroskedastic. These results hold
for sample sizes as small as 250 observations and are therefore relevant for macroeconomic
time
series.
17
A. Appendix
Lemma
-
Lemmas
A.l. Under Assumption (A-1) the following statements hold:
a) for each
m 6 N+\ {1} m fixed,
the vector
,
-^
Yl^=i i^t^t-i, ,£t£t-m] =^
N{0,Q) with
a{l)+a^
n=
(T{m)
+ (7^
Leram.a A. 2. Let In,yy (A) be the periodogram of {yi,. ,yn} and /„_££
odogram of {ei ,...,£„} Assume et satisfy Assumption {A — 1) and that yt
.
.
(A) is the peri-
=
YI'jLq i^j^t-j
with spectral density ^gyyiPot^) such that Yl'jLo |'0j| \j\ < o°- Let (.) be any continuous
tt] —^M. with absolutely summable Fourier coefficients {zk,—oo <k < oo}
<;
function on [— IT,
then for any
77,
P(y^
as
n
e
,
>
J
In,yy (A)
<;
{\)
dX -
^J
/„,,, (A)
<e
dX >r/
gyy{pQ, X)q (A)
00.
Proof. The proof is the same
minor modifications.
Lemma
A. 3.
(Billingsley)
pose that for each
m
Xr,
p.
388, with
Let Xmn, Yn be random variables defined on {Q,,J^,P)
Xm
—^
1
and Davis (1987),
as the proof in Brockwell
as
—
n
00 and that
>
Xm
—^
X
as
m
.
Sup-
-^ 00. Suppose
further that
hmsup
lim
for each positive
Lemma
e.
A. 4. Let
Then Yn
^>-
X
as
P {\Xmn - i^n| >
n —f
00.
In,yy{X) be the periodogram of {yi,.
riodogram of {£!,...,£„}. Suppose the
Y^'jLoi'j^t-j with spectral density
any continuous even function on
=
e}
satisfy
et
.
.
,yn} and
Assumption {A —
[—n,Tr]
such that
fc=i
and f^^
q (A) QyyiPo,
X)dX
=
0,
then
In,yy (A) ? (A)
dA
T
with bk
A iV
0,
\
= Ejl-oo 7yy {k - j) Zj.
18
I)
is
the pe-
and that
yt
=
that Yl'jLo
°°- ^^^ "^(O ^^
l^il bl <
with Fourier coefficients {zk, —00 < k < 00}
^gyyiPoA) such
—> M
/n,ee (A)
4
^
t,_i
fc=i
Qfet^
,
B. Appendix
-
Proofs
Proof of Lemma 4.1
— —T
Y^°^_
J
<
First
oo-
we show that /„
Next note
(A)
for a typical
€ Li
which follows from J^
[—tt,7t]
\fa (A)|
element k
—n
oo
/jr
-•^
oo
oo
oo
j=0 l=-oo
j=0 /=-oo
oo
oo
I
/ oo
^
I
oo
gJA(j+fc)
such that the result follows.
Proof of Proposition 4.2 Using equation
can be written as Re[In,zy
W] =
+4>{e-'^)li.
(4)
In,yy (A) Re[h^{(f),
(-A)
u;,
i-X)
Rl^
and the
A)](?!)
(A)
definition for
h^{(f).,
A), /n,2y (A)
+ /^ (A) In,ee (X) + Re[Rn (A)]
+
l^
(-A)
where
|<^(e'^)|' |i?^_^ (A)|'
+i?i,^(-A)i?2(A)
and
a;,(A)
The two remainder terms R^ ^
quences
uct i?^
Un,j (A)
g
= rne"^>e(A)+<^(A).
and R^
(A)
(A)
are defined next.
First define the se-
= {ipQ,ip-y,...} and Tpik) = {^o/<^fc! V'l/^i+A:) •••}• Then let the inner prod= -^Y^'jLoSj^~^^''Un,j
be defined for a sequence s = {so,si,...} where
= Yl7=i-j £te*'^* — Y17=i ^te'^*. We can now write the remainder term R^ (A) in a
Tp
W
(A)
compact form
Rl
where R^
-
(A)
'
=
(A)
=
-l^
(A) 0(e-^^)Hi,^(A)
^(uW' ...,R^
[i?-^
+ a (X)* © i?;^_- (A)
such that i?^ (A)
r, ,(A)]
is
a
p x
vector. This leads
1
'
to
Re[In,zyiX)]dX
=
/
+
In,yyiX)Re[h%(j),X)]dX(f>
/"
Ir^
(A) /„,.. (A)
+
dA
J —ir
Alternatively one has also Re[In,yz (A) a (—A)]
with RP, (A)
=
(l)-\e-^^)uj, (A)
Rl i-X)
+
r Re[i?„ (A)]dA.
(1)
J —TV
=
Rj, (A)
19
In,yy (A) Re[/i^((?!>, A)]
R^ (-A) Now
.
if
-I-
Re[i?^ (A)
/Re [i?^
a(— A)]
(A)] <r(A)dA
=
dX
=
^/^),
Op (n
(f)
+
where
has absolutely summable Fourier coefficients,
<;(A)
follows that
it
(f)
=
Opin^^f^) and
-1
In,yy{X)Re[h''{cj,,X)]dX
Vn[j)-(j)j
x-v/n
r
(A) In,.e (A) c?A
Ir,
Since l^ (A) G C'^
[-7r,7r]
and
(?!>(e^'^)
G
J —TV
C'^ [-tTjTt]
/
/•IT
^^(A)J„,eaA)tiA^7V
V^J
where the matrix Ym^i
'^i
W
{^n
>
s'^')
follows
it is
Lemma
(A. 4) that
A
(A)
{^n
,
i\l
,
has typical element
e*'^')
=
J2'^J+kvl'l'4-,J%,J+\k-l\-
j=0
k,l
(4.1)
from
OO
1=1
Lemma
it
0,^a,(z,(A),e'^^)(z,(A),e'^^)
J^ai{l^iX),e'^')(lr,{X),e
Using
T Re[i?„ {X)]d\
+
J — TT
easy to show that
5^a,(/,(A),e^^A(z,(A),e'^^-
faiX-Ov{<l>,Ov{^,X) d^dX + -,
/
77
a'*
a
A)
77 (<^,
A)
dX
^/n X^tt -^ (A) dX = Op (1) We discuss the four different types
of remainder terms separately. Since Z^ (A) and Z^ (A) (j){e^^) have absolutely summable
The
result then follows
if
.
Fourier coefficients and are uniformly bounded on [— 7r,7r] the proof of
be applied to the remainder terms involving
R\ ^
Lemma
To show that -Jnj^^(f)~^{e^^)u)^{—X)R'^{X)dX
=
Op(l)
is
it
enough to show that
element by element
Ey/n [%-'ie^^)u,{-X)Rl,,{X)dX
0.
J —IT
Using the definition of
/?^ (A)
EV^
one has
r r'
(e^^)'^e
J — IT
<
(- A) Rlf,
(A)
dX
E^ Jr l^^k{X)ct>'\e'^)<p{e-'^)u,{-X)R\^^{X)dX
—IT
+Ey/n
£ r'
(f")
-. (-A) n- V2e-A'=iii
20
(A. 2) can
(A) alone.
(A)
-
^^^
dX
(e ''^) and applying
term can be analyzed by setting q {X) = l^ (A) (^ ^ (e*"^)
the proof of Lemma (A. 2). Next look at a typical element k of the second term
The
first
(f)
i^i
EV^
^
/
^ -^
Then
for j,
Z,
and r
oo
oo
J
,,
n
EE E E S^'.
(='- -
-"--) --"<'-'-'*"<iA
j=0 1=0 r=l t=l "'=+J
fixed such that
l<r — k + l<n
1/2
E
fc+/
Now summing
i,
over
'-"j-n-k+l
m gives
j,
U^) u>, i-X) n-V2 Y^ ^.e-'^Wnj
This holds true
1/2
'
for all
k
=
l,...,p.
(A) ? (A)
dX
<sup24/vv2j;j:
^,-^/
j=0 /=0
"fc+j
lJl-0.
Next consider
r Rli-X)Rl,{X)dX
J—TT
Vfc(A)0(e-'^)|i?i(A)|'dA
r R\ (-A) n-1/2 V Ae-^(^+'=)C/„,,
+Ey/n
^
V-TT
such that the
first
OLj
term goes to zero by the proof of
(A)
dA
+k
Lemma
The second term then
(A. 2).
is
OO
E^
/
y J^e-Mi+fc)c/^
Ri (_A) n-1/2
.
(A)
dX
/TT
< En
(fA
-TT
<
n-^l''
'^
/TT
1=0 j=0
EE
"j+'=
E\Un,l{->^)\\UnjW\dX
/=o j=0
OO
.
\
1/2
""
1/2
< n
(^it^n,,(A)r
0!j+k
/TI"
< n-V2
CO
OO
ElV'dmin(/,n)^/2E
'^
Z=0
j=0
21
mm{j,n)^^^dX^O.
dX
The remaining terms can be shown
to go to zero
by the same arguments and the proofs
are omitted. This completes the proof of the proposition
Lemma
=
^ Z7=i+p+i £ti<t>)^t-M, ""^ = Ea* {4>) with e^cp) defined
satisfying Assumptions (A-1) and (B-1) and any fixed S < oo it follows that
(2), £t
B.l. For a*
(</.)
sup
maxt;ar(n^/^
Proof. Let yt
=
,yt-p\ and
[yt,
m,^
and a"^
=
ai^^{n
=
— — p)/n.
et {(pf Et-i
4>q
=
=
fi'n,y...y{qi,
94=0
=
sup
\ai
sup
'^Y^\^g^---(J}g^\\il'^yy{qi,...,qi,l)\.
<l>^N6{<l>o)
n'^ Y17=i+p+l
{(f))
-
q^=Q
'
--yt-i-qi
-
Eyt-q^
--yt-i-q^.
Then
ai^^l
(2)
54=0
can be bounded by
s
fmax(|(?!)i_o|
^
it
yt-qi
p
/
so that
write
^'Yl^i^'" ^giyt-qiyt-q2yt-l-q3yt-l-q4-
p
(2)
we can
1
-,94,0
<
Expression
0(1).
p
gi=0
define
=
vec{ct)(f)'y{yty't0yt-iy't_i)vec{(t)(f)')
p
Now
af^\)
vec{(l)cj)')'E{ytyt®yt-iy't-i)vec{(j)(j)')
=
((P)'^
-
Then
=[l,(t)']'.
</>
Letting
I
|a,* {(())
in
+ ^)
p
p
XI
'
'
'
J
^91=0
'
remains to show var
n
|An,y...2/(9ij
XI
|An,2/...2/(9i>
•••>94,0|
94=0
•)94)0|
=
0{n~^)
for all
/.
Now
var
\jj!n^y„,y\
is
n
X X
"~^
4
COv{yt-q^
yt-l-q^,ys-qs
--ys-l-qs)-
t=l+l+p s=l+k+p
We
"~^
can substitute
n
n
X
Yl
for the definition of yj
oo
^
'4'j^t-j
which leads to
oo
X X
t=l+l+ps=l+k+pji=0
=
^Jl
^J8^°^(^*-9l-Jl
^t-l-qi-H,£s~qs.-h
'
'
'
^s-k-qs-js)
J8=0
(3)
The
covariances can be represented by eighth and lower order cumulants by considering
the following matrix
X{t,sJ,qi -ji,...,q8-J8)
=
^t-qi-jl
^t-q2-h
^t-l-q3-J3
^t-l~q4-J4
^s—qz—jb
^s-qe—je
^s-l—qr-jr
^s-l-qs—js
22
.
with typical element Xij where reference to the time indices
Brillinger (1981),
Theorem
Then from
suppressed.
is
2.3.2,
4
4
coz;(JJ
j=l
Xi
,
J,
n
JJX2j) = X]
ctim(Xij,z,j G Vs)
Vs€v
V
j=l
where cum{Xij,i,j € Vs) is the joint cumulant of all the Xij with indices
the sum is over all indecomposable partitions v of the table
(1,1)
(1,4)
(2,1)
(2,4)
i,j
S Vs and
•
A
given in Brillinger (1981), p.20. We note that
cumulants are zero. Second cumulants are nonzero only if the time indices of et
definition of indecomposable partitions
all first
By
coincide.
indecomposability of partitions there
from both rows
n
-1
is
in
X
in
each product Ylv
is
one cumulant with elements
at least
£v^'^''^i-^i,j^'''i3
^
Es Et co'u(n)=i ^i,j, 11^=1 ^2,j) = Es cou(nl=i ^l,j, nj=i
dices for
^Y
^s)-
strict stationarity
^2j) where the time
in-
have been normalized to zero. The summability assumption (B-1) implies that
t
Es<^^''^(nt=i ^ij,n7=i-^2,j) < oo uniformly in I, j\,...,js- This shows that the term (3)
is of order n~^ uniformly in I as had to be shown.
Proof of Proposition
for
any
ry,
5.1 For the
first
part of the proposition
we have
to
show that
A^^ ((^g) is
contained
>
e
lim
P{ sup
hr,
>
(0„,A)-/i(0o,A)
77)
<
e.
Ae[-7r,7r]
This holds
there
if
a ^ and a neighborhood
is
A^^
lim
P(
Consistency of
considered.
4>
hn{(l),X)
implies P{(p
€
-h{(po,X)
A^^ {4>o))
of
(pQ
such that
—
1
>
>
rj)
+
lim
P{(j)
so that only the
^ Ng
first
{(f)o))
<
e.
term needs to be
From
hni(f),X)
Re
<
sup
sup
{(J)q)
parameter space and
in the interior of the stationary region of the
-
h{(f)Q,X)
f^(A)</>(e-^)
l^ (A) </.(e-'^)
-
Re hfi{X)Ue-^^)
/^,o (A)
-iA>
M^~n
+
1
i^(-X)^{e'^)-l^^o{-X)Me'^)
and
l^{X)cp{e-^^j-l^,o{X)Me-'^)
<
kW-h,o{X)
<t>{e-^^)
+||Z^,o(A)||
23
0(e-^)-(/)o(e-^^)
,
it is
enough to show that
while
\(l){e'-^)
To estabhsh
element
aj^tj)
=
^ip,k
—
(e''^)|
(f>Q
sup
sup
A
4>€Ns{4>o)
<
(A)
—
ltp,kfi
Eel{4i)el_A(j)).
sup
on Ns
57/2
sup^^sup^^g^y^^^^^^
(A)
-
li,
Z,/,
(A)
Now
let ctj
/^,o (A)
=
=
Op (1)
Ee1e1_j,
Then, using the definition of
sup
'i/',fe
(A)
—
Z^,fc,o
=
l^fi (A)
Op (1)
by uniform continuity of ^q
{4>o)
—
-
(A)
bj^k
it is
{^^^) o^i
enough to look
denotes the
A;-th
[-t'",'?'']
at a typical
element of
bj
and
l^^k (A)
(A)
Ae[-7r,7r](;6eiVi(<^o)
sup
n—p—1
oo
j=i
j=i
sup
n—p—l
<
sup
sup
Yl
(^j(^) ^hk,4>-(^j,lhk,4>
+ (^jibj,k,ct>-aj^bj^kfi)e
'^^
(4)
j=i
+ sup
A
j>n—p
We note that sup;^ Er>n-p cq\ks>e-''^ <
sup,,
afn-'/' T.T>n-pj"'
\hkfi\
=
o{n-''').
Next
n—p—\
n—p—l
<
(5)
n—p—l
+
E
3
It is
therefore
enough to show that
(5)
and
(6)
{oc-\bi^k,^-b,,kfl))e-''^^
(6)
=1
go to zero uniformly on [— 7r,7r] as 6
—
>
0.
First consider (6).
n—p—l
sup
{(f)Q,n))dn
Ae[-7r,7r]
<
sup
Ae[-7r,7r]
n—p—l
1
+-4
(^
sup
(7)
A£[-7r,7r]
24
where, for
E Ng
cf)
{(po)
the finite Fourier approximation of
,
77 {(f>,
^) converges uniformly
such that
n—p—1
sup
yZ
Ae[-7r,7r]
J
=
sup
ihk,4>
-
^j,kfi) e'
i\j
=l
(77 {cp,
X)-fj
(</>o,
A))fc
-
Y,
AG[-7r,7r]
<
ibj^k,4>
-
bj,k,o)
e-'^'
]=n—p
sup
\7]{4>,X)-f]((f)Q,X)\f.
+
V
rap
2
\bj^k,4
AGf-TT.TT
<
+2
e
^
sup
\bj,k,<p\
4>eNs{4>o)j=n-p
Then, letting
ctj
=
a~^ — a~^ the
term
first
in (7)
dominated by
is
n—p—1
<
sup
\Vk{4>,iA -'?fc(0o>M)M/^
Ag[— 7r,7r] J —TT
j=i
n—p—1
<
J—n
„_,
j=—n+p+l
,
<
The constant
continuity of
from which
1
Ce{5).
n—p—1
C is bounded by X^,=i
7)
it
< 00. From uniform
< Ejli Qj (-^o)
Wo)
— f] {(f>Q, ^) < e for some 5 >
(cj),
fj) on [— vr, tt] x A^^ ((/)q) we have \ri {(p, fi)
follows that J_^ ri{(j),fx) — f] ((f)Q,fj,)\i^dfi < 2'ne. The constant C can be
[.
bounded by (27r^°l^
goes to zero as
aij
n —
>
dj
00.
((/>o)~
+ 0-4)
independent of n and sup^g^^(^^) Ejln-p
Next we consider
(5). First
l^j.fc.^l
look at
Pj>-^jl
From
aj_0
a.s.
and
(j),
=
E[4>{L)yt)^ {(f){L)yt-j)'^
But, since
(f){L)yt
4>{L)yt has
=
(f){L)(f)Q
it
> a^ > since otherwise (j){L)yt =
an ARMA{p,p) process with parameters 0q
follows that a^^^
{L)et
is
nonzero variance contradicting
(«j,0)~^
-
(^J^
< Ci
(f>{L)yt
\aj,4,
-
=
a.s.
aj\
some constant Ci. Since Eej < 00 we can uniformly bound
C2 is a finite constant depending on (f>Q and Eef. Then
for
Then
\aj^^
n—p—1
n—p—1
j=l
25
—
aj\
by 6C2 where
^
1
where Y^=i
<
\^oM,<t'\
oo on
Ne
{(po)-
Now
turn to the
two terms of
first
(4)
n—p—l
n—p—l
<
Now
"j
(</•)
- ",i <
^
=
.
"
^
^
1
- ajj <
{4>)
\aj
Y^
a^-,,^)
("j (0)
l^i
where a^
sup
{(f))''^
|aj
(</))
\aj
{(f))
-
aj^^-^
aj, <^|
^j,k,4>\
and
——
- a]J + J'+P
n
the expected value of a*
^^.(^ is
-
((^)
|aj-
Then
.
n-p—l
n—p—l
w— p—
n
First note, that since Qj^^
replace
{ctj
((f))
is
bounded away from zero and aj
{(f>)~
<
cn^/^~'",
we can
aj^^)"^ by n^^'^~'" times a constant that does not affect the argument.
Then
the second term
n—p—l
^-i/2-i;
^
{j
+p)
\aj^^\ \bj^k,4>\
-^
n
as
^ CO
3=1
uniformly on
Ng
((^q)
The
•
term now
first
is
shown
to go to zero in probability
by looking
at
P{
The term
n}''^~^
X]?=f~
^n-p-l
The
last
term
is
ni/2-
sup
.
^i i^)
~
Y"-^~
^?<4
\^J i^)
l^j,fc,</>l
-
^U
\hk,4>\
>
n)
dominated by
^^
n—p—l
^
zero with probability tending to one
which follows form maxj P(n-^/^ sup j,g^(0
the first term is bounded by
n
)('-''*i d>
maxj P{sup^^j^^/^
if
~ '^1 d>^ >
s—^n-p-l
n"'
ry)
—
sup
>
0.
\
a^
By Markov's
|a*(<^)-a^^^|
4>&Ns{ct>o)
1/2
<
^^YT.
^
\hk,<l>\\^ar{n^^^i
26
sup
\a* {^)
-
-
al^\))
<
d„)
—
>
inequality
.
The
^
from
result then follows
nmaxfar(
This
shown
is
It
Lemma
l^j.fc.^l
sup
it is
|q;*
(0)
0{n-'')
—
if
= O (1)
a"^|)
.
(B.l).
remains to show that P(p(/i„(^„,
part
first
in
=
^
YTjZl
A), /i(<;6o)'^))
enough to show that sup;^g[_^
,r]
>
^^
^)
dhn{4>ni
Using the result from the
0.
^)/dX - dh
A)
(</)q,
/dX
=
Op (1)
Since
implies that
e C^[-7r,n]
f]{(l),X),^
V/i
from Folland (1984), Theorem
follows
~
Bernstein's Theorem, ^\j\
<
8.22e, that {ij^hj
\bj {(p)\
<
=
oo and d^r]{(j>,-T^)^ /dX^
{(j))
=J
d^i]{(i),T:)^
/dX^
(^T]((f),X)) e-'^^dX.
it
By
oo such that
^(^j)6,(</>)e'^^->^r,(</>,A)
uniformly on [— 7r,7r]
the
first
Since
that
.
Using these
part can be applied to the
{Ti.,p) is
PChn
and noting that
facts,
first
lr]{X)
€ C^ [— vr,7r]
,
the proof of
derivative.
P (p
a complete metric space
(
/i„(^„, A),
^((J^iq,
A)
]
>
6)
^
implies
eH) ^i.m
Proof of Theorem
the proof of
Lemma
5.2
(A. 2)
We
start
by obtaining an expression
for
Rn
(A) in (6).
we have
oo
^y
=
(A)
ct>o'
(e~'^) ^. (A)
+ n'^/^ ^V'^-e-^^^'C/^,-
(A)
j=0
Letting
R^
(A)
In,yy{X)
=
n.-^/^
Yl'^o ^je'^^^Unj (A) we can write
=
UJy{X)Uy{-X)
=
^y
=
In,yy (A)
(A)
^y (-A) a (A)
a (A)
4>o
(/.q
+ Uy (A) u, (-A) + cjy (A) 0o
+ In,ee (A) <Po'
+u,{X)-p^Ri{-X) +
00 (e
In the
same way we
In,yy (A)
also
=
[^"^)
+ ^e
(e'^)
(-A) hi
< (-A)
(A)
\Ri{-X)\'
^-^j
have
In,yy (A)
a (-A)
<^o
+ ^n,.. (A) <t>o^
27
(e^^)
+ o;, (A) R], (-A)
From
Let
Rn
Next write
= Re
(A)
(-A)
(A) i?i
a;,
=
A)
/iQ (</>0,
Re
In,yy (A)
+In,ee (A)
while a
filter
based on a different parameter
=
In,yy{\)h{(p,X)
(A)
Zr,,0
</>
(-A)
l^^Q
+
/i0|j
that ^/n
(<?!),
j Rn
coefficients.
= Re
A)
(A)
(-A)
(A)
dX
(11)
is
<;
Next
[l^p
In,yy{X)Re
=
Op(l) for
P
limsup
ra—oo
which
in turn follows
lim sup
[e'^j a (-A)
(pQ
/iQ (</'o,
(po
A) i?n (A)
{- X)
l^
(j)
+
(e'^^ a
h
{- X)
(t>Q
(^, A) Rr, (A)
<;
for all
sup
\h&n,pihM)<6
i?,
>
e
||f„ {h)
-
there exists a 5
Vn
>
(/io)||
<
'J?
>
such that
^
J
from
P
n->oo
as
(pQ
i^''^)]
all
if
(-A)I'
Then by the proof of Lemma (^.2) it follows
continuous (A) with absolutely summable Fourier
(j){e'^)(t)Q^
estabhshed
|i?i
expressed as
is
+In,ee (A) h^, (0, X)
with
+
(A)
based on the true parameter
In,yy (A) h^ {cpQ, A) for the filter
In,yy (A)
%^i?i
+ ^, (-A)
sup
^/n
<e
In,esW{h4,,{(f>,X)-lr^fiW)dX
[
(8)
J — IT
\h&H,p{h,ho)<6
and
P
lim sup
n-xx,
For
(9) define
sup
all
(A) [h
-
ho]
<£.
dX
(9)
the open neighborhood
ns = {h:
Now for
Rn
^/n
\hGn,p{h,ho)<6
h €
TCs
it is
[-7r,7r]
^RP \h
en,p{h,ho) < 6}
the case that h (A) € C^[— 7r,7r].
.
By Bernstein's theorem this implies
— ho is also continuous and
uniformly bounded with absolutely summable Fourier coefficients such that (9) follows
absolute summability of the Fourier coefficients of h. In turn h
from the proof of Lemma (A. 2).
Next turn to (8) The main idea of the proof
by parts is used to separate h from In^ee (A)
.
is
taken from Robinson where integration
-
<
r In,eeW{h4,,{<P,X)-lr^fiW)dX
sup y/n
h&Hs
J —TV
sup y/n
h&Hs
J —IT
+
sup
r
Vn
{In,,e (A)
-
EIn,ee (A)) {h^,
(</>,
A)
-
lr,,0
EIn,ee{X){h<t>,{(l),X)-lr,fiiX))dX
28
(A))
dX
Since EIn,ee (A)
note that J^^
=
Ir^fi
cr^ it
(A)
dA
sup y/n
heUs
r
and /^^
sup \/n
+
-
h,p^
{In,,e (A)
{(f),
EIn,ee (A))
{h^o
(</),
(V
^)
(<^'
{4>, Tt)
-
-
(A))
^'/.O
for
/i
/^,0 (tI"))
/
J —-K
(-^n.ee (A^)
Q
PIT
/
'dX
remains to show that y/n J^^ J_^
ity.
Let /^gg(^)
=
—
/„_££ (/i)
"
^^^"^
^'''°
—
{^n,e£ (/^)
ting Hg
=
=
^,
In,ee
we
(m
+
71")
The
/j,,o (7r)||
are uniformly
~
'^^
E^n,6e
(/^)) C^MC^A
dX.
(fJ-))
d^ dX
is
bounded
in probabil-
fj.<X
/Lt> A
we work with y/n J^''
n-l
n
-
tt)
(10)
/q (/„,ee
(f^)
-
Eln^ee
(a^)) C^Ai
dX. Let-
show that
first
27r
£^/n,e£ {lA) df^^X
Eln^ee (yu)and define the function
1
(yu)
-
-E/„,ee (ju)) C?A
(</>,
^"^"'^^
/
EIn,ee
t{X,h)
Since /„,££
dA
/-A
^^'^° ^^' '^^
d^
It
"
||/i^g
(<?!',
sup y/n
hens
(A))
/^,o
{In,ee (m)
/
A) - lr,fl (A))|| and
ein Tis both ||^ (/i^^,
bounded by C6 for some constant C < oo such that
Now
-
A)
Now
r\
-Q^
sup y/n
h&He
0.
(V,
Q
J^^ (h^^ {(p, A) - Irjfl (A)) dX.
Next use integration by parts
is a'^
=
dX
A)
J — IT
TT
<
term
follows that the last
=
V
„27r
r(A, /xjl°,ee (Ms)
-
/
^(A, M)^n,e£
(A^)
^/^
dX
=
Op(l).
inner integral can be split into a part
"-1
"
f27r
-^0
5=1
"-1
/•27r(s+l)/n
= X]
[(7"(A,AXs)
/
-
T{X,ll))In,s6
ifJ-s)
+ r{X,IJ.)(In,,£
(/J-s)
-
In,ee
{^J'))]
dfM
r2iT/n
+
T(X,n)In,eei^)dlX
(11)
Jo
and o-2(^ ^^Zj^ r(A, /xj 27ri
< An
implying that
/(f"
t(A, n)dn). If M^ < A for all s
2TTt/n < 27r/n such that
"-1
/r27r
— y]^(A,/xJ-/
27r
n<T
Vni
<
i
then
<A—
29
T{X,fj,)diJ,
=
0(n-i/2)
27r(i
+
1)
> An and
1
uniformly in
A.
Using Markov's inequality we look at each term of (11) separately. First
"-1
u
yfnE
r27r(s+l)/n
{t^K Ms)
- t{X, lj))In,ee
{f^s) ^/^
l2TTs/n
"-1
IV —
<
/•27r(s+l)/n
p
Wn
sup
2ns/n
where sup^ \T{X,^g) —
=
t(A,/x)|
"-1
|r(A, Us)
-
t(A,
/x)
a'^dfj.
I
fJ-
A ^
if
+
[27rs/n, 27r(s
l)/n]
and
1
otherwise. Therefore
^
/•27r(s+l)/n
27r
727rs/n
V^l
A^
uniformly in A. Also
/•27r/n
^/nE
<a'
T{X,^)In,te{lj)d^
/
JO
where the bound
is
again uniform in
"-1
\fnE
"-1
2tt
^/n
A. Finally, since E{In,ee {^Js)
~
In,ee ifj))
=
0,
r2n(s+l)/n
T{X,jl){In,ee
~(
J2-n
s=l •'27rs/ra
(mJ "
-^n.ee (Ai))c^M
r27r{s+l)/n
g—l J2-Ks/n
where, from Brillinger (1981), p. 417, nvar(In,e£
2iTs/n < fi < 2n{s + l)/n. This shows that
n-l
uniformly in
A.
is
=
We
bounded
0{n~^) uniformly on
=
0{n-'/^)
can therefore consider
n
J —IT
is
In,ee{fJ'))
n
2n
which
—
j.2n
27r
y/nE
{/J^s)
in probability
n-l
^^(A,Ms)^°,£e(Ats) dX
s=l
by Markov's inequality if sup;^
bounded. From Brillinger (1981), Theorem
^ Yl^=i
5.10.1,
2
n-l
nE
nE
— ^T(A,/xJ/°_^^(AiJ
s=l
Jo
+
/
/
Jo
o/o
/e..e(/^l,Ai2>-j"l)c?/^l<^M2+C>(n
30
^)
'''(^^
f^s)^n,ee (Ms)
where the error is uniform in A. Then f^^{X)
bounded under assumption (A-1).
From Brillinger (1981), Theorem 5.10.1 and
=
cr'^
5.10.2,
and
it
/^..^ (/ii,/X2,
— /^i)
is
uniformly
follows immediately that
VnE
J —TI
is
bounded such that the second term
in (10)
is
the proof.
31
small in probability on Tis. This completes
References
Andrews, D.W.K. (1994) Asymptotics
for
semiparametric econometric models via sto-
chastic equicontinuity. Econometrica, 62:43-72.
Bickel, P.J. (1982)
On
adaptive estimation. The Annals of Statistics, 10(3):647-671.
BoUerslev, T. (1986) Generalized autoregressive conditional heteroskedasticity.
Journal
of Econometrics, pages 307-327.
Brillinger,
D.R. (1981) Time Series, Data Analysis and Theory, Expanded Edition. Holden-
Day, Inc.
Engle, R.F. (1982) Autoregressive conditional heteroscedasticity with estimates of the
variance of united kingdom inflation. Econometrica, pages 987-1007.
Folland, G.B. (1984) Real Analysis
-
Modern Techniques and Their
Applications.
John
Wiley and Sons.
Hannan,
The asymptotic theory
E.J. (1973)
of linear time series models.
Journal of
Applied Probability, 10:130-145.
Hansen, B.E. (1995) Regression with nonstationary
volatility.
Econometrica, 63(5):1113-
1132.
Hansen, L.P. (1985)
A method for calculating bounds on the asymptotic covariance matri-
ces of generalized
method
of
moments
estimators. Journal of Econometrics, 30:203-
238.
Hansen, L.P.
&c
Heaton, J.C.
moment
conditional
& Ogaki,
M. (1988)
restrictions.
Efficiency
bounds implied by multiperiod
Journal of the American Statistical Association,
83:863-871.
Hansen, L.P.
&;
linear time
editors,
Singleton, K.J. (1991)
series
Computing semiparametric
efficiency
bounds
for
models. In William A. Barnett, James Powell, and George Tauchen,
Nonparametric and Semiparametric Methods in Econometrics and
Statistics,
pages 387-412.
Hansen, L.P. & Singleton, K.J. (1996) Efficient estimation of linear asset-pricing models
with moving average errors. Journal of Business and Economic Statistics, pages
53-68.
Hayashi, F.
&
Sims, C. (1983) Nearly efficient estimation of time series models with
predetermined, but not exogenous, instruments. Econometrica, 51(3):783-798.
Hidalgo,
J.
(1992) Adaptive estimation in time series regression models with heteroskedas-
ticity of
unknown
Kreiss, J. P. (1987)
On
form. Econometric Theory, 8:161-187.
adaptive estimation in stationary
of Statistics, 15(1):112-113.
32
ARMA
processes.
The Annals
Kuersteiner,
tors.
G.M.
(1997). Linear covariance matrix lower
bounds
for
time series estima-
Manuscript.
Linton, O. (1993) Adaptive estimation in
ARCH models.
Econometric Theory, 9(4):539-
569.
Milhoj, A. (1985)
The moment
structure of arch processes.
Scandinavian Journal of
Statistics, 12:281-292.
Nelson, D.B. (1990) Stationarity and persistence in the
GARCH(1,1) model. Econometric
Theory, 6:318-334.
Newey, W.K. (1990)
Efficient instrumental variables estimation of nonlinear models.
Econometrica, 58:809-837.
Robinson, P.M. (1987) Asymptotically efficient estimation in the presence of heteroskedasticity of unknown form. Econometrica, 55:875-891.
Robinson, P.M. (1988) Root-n-consistent semiparametric regression. Econometrica, 56:931954.
Robinson, P.M. (1991) Automatic frequency domain inference on semiparametric and
nonparametric models. Econometrica, 59:1329-1363.
Steigerwald, D.G. (1994) Efficient estimation of models with conditional heteroscedasticity.
Technical report, Department of Economics, University of California Santa
Barbara.
Optimal instrumental variable estiprocess. IEEE Transactions on Automatic
Stoica, P. &: Soderstrom, T. &: Friedlander B. (1985)
mates of the ar parameters of an
ARMA
Control, 30:1066-1074.
33
Table
Model
1:
Relative efRciency of
yt
:
=
(jnjt-i
+et
OLS
et
1'^
"-.ML
1'^
<Pn
9n
4>n
n
=
for
ARCH(l)
uth^'
'•iML
n =:512
== 256
innovations
ht
=
0.1
+ 7..eU
1'^
~,ML
(Pn
<Pn
n
=
1024
<!>
7i
0.0
0.0
0.9969
(0.0023)
(0.0058)
(0.0017)
(0.0041)
(0.0011)
(0.0029)
0.0
0.5
0.9918
0.5158
0.9947
0.4144
0.9984
0.3801
(0.0015)
(0.0158)
(0.0013)
(0.0134)
(0.0011)
(0.0126)
0.9619
0.4458
0.9631
0.3978
1.0294
0.9966
1.0126
0.9980
1.0107
0.0
0.9
0.9502
0.4432
(0.0031)
(0.0138)
(0.0043)
(0.0131)
(0.0046)
(0.0123)
0.5
0.0
0.9937
1.0272
0.9985
1.0087
0.9971
1.0106
(0.0030)
(0.0054)
(0.0021)
(0.0039)
(0.0016)
(0.0032)
0.5
0.5
0.9388
0.4761
0.9375
0.4515
0.9616
0.3658
(0.0056)
(0.0144)
(0.0054)
(0.0144)
(0.0082)
(0.0121)
0.5
0.9
0.9361
0.4083
0.9171
0.4478
0.9106
0.4759
(0.0078)
(0.0131)
(0.0096)
(0.0137)
(0.0108)
(0.0141)
1.0046
1.0190
1.0046
1.0096
1.0012
1.0025
(0.0040)
(0.0054)
(0.0029)
(0.0036)
(0.0022)
(0.0029)
0.5084
0.8958
0.4416
0.8590
0.3715
0.7
0.0
0.7
0.5
0.9071
(0.0070)
(0.0155)
(0.0081)
(0.0139)
(0.0076)
(0.0121)
0.7
0.9
5.5958
0.3998
0.8274
0.4470
0.8274
0.4470
(0.1854)
(0.0134)
(0.0120)
(0.0138)
(0.0120)
(0.0138)
1.0391
1.0097
1.0152
1.0101
1.0085
1.0022
(0.0069)
(0.0047)
(0.0053)
(0.0035)
(0.0037)
(0.0024)
0.5420
0.8522
0.4570
0.8308
0.4663
0.9
0.0
0.9
0.5
0.9209
(0.0101)
(0.0154)
(0.0094)
(0.0135)
(0.0094)
(0.0142)
0.9
0.9
0.7281
0.4352
0.5951
0.3486
0.4878
0.4247
(0.0123)
(0.0151)
(0.0113)
(0.0119)
(0.0107)
(0.0139)
is defined as Sfy/Spi^g or S]i^j^/SQj^g, respectively,
where S^ is the estimated variance of the estimator 4>- Numbers in parenthesis
are asymptotic standard deviations of the variance ratio. Results are
Relative Efficiency
based on 3000 replications.
34
Table
Model:
yt
=
(f)yt-i
Means and Medians
2:
+ et
et
= uthj
ht
1'^
"lOLS
= 0.1+7i£?_i
'-.ML
<Pn
4>n
<t>
7i
0.0
Mean
Median
Mean
Median
Mean
Median
-0.0007
-0.0006
-0.0006
-0.0005
-0.0007
-0.0007
0.0
0.5
-0.0025
-0.0014
-0.0024
-0.0016
-0.0009
-0.0009
0.0
0.9
0.0006
0.0000
0.0009
-0.0002
0.0011
0.0009
0.4978
0.4994
0.4969
0.4980
0.4977
0.4992
0.5
0.5
0.5
0.4927
0.4935
0.4908
0.4919
0.4987
0.4995
0.5
0.9
0.4770
0.4830
0.4694
0.4780
0.4945
0.4984
0.6962
0.6969
0.6948
0.6957
0.6962
0.6970
0.7
0.7
0.5
0.6933
0.6950
0.6913
0.6933
0.6979
0.7000
0.7
0.9
0.6779
0.6888
0.6733
0.6847
0.6945
0.6992
0.8965
0.8981
0.8947
0.8962
0.8965
0.8982
0.9
0.5
0.8948
0.8977
0.8933
0.8957
0.8982
0.8992
0.9
0.9
0.8850
0.8925
0.8860
0.8921
0.8984
0.8999
0.9
Samp
e size is 512.
Results are based on 3000 replications.
Relative Efficiency
0.95
0.9
0.85
0.8
0.75
AR
0.7-
0.2
0.4
0.6
0.8
Asymptotic efficiency of OLS relative to the IV estimator as a function of the
parameter
Generating mechanisms considered are from bottom to top: 7^ = .5, 7^ = .4,
Figure
1:
cj).
7i
=
.3
and 7^
=
.2.
35
0.65
0.70
0.75
0.80
0.85
OLS
Figure
2:
0.90
0.95
1.00
1.05
1.10
ARCH-ML
IV
Empirical density of parameter estimates for an AR(1) model with
the errors have
no
ARCH
effects
36
4>
=
.9
when
0.65
0.70
0.75
0.80
OLS
Figure
3:
0.85
0.90
0.95
1.00
1.05
1.10
ARCH-M
IV
Empirical densities of estimated
AR parameters
37
when
cj)
=
.9
and 7^
=
.9
7 U
/
b
Date Due
Lib-S6-6r
MIT LIBRARIES
3 9080 01972 0991
Download