Document 11074785

advertisement
^^
Of
tbc^
-''
HD28
.M414
^1
ALFRED
P.
WORKING PAPER
SLOAN SCHOOL OF MANAGEMENT
THE SOURCES AND NATURE OF LONG-TERM MEMORY
IN
THE BUSINESS CYCLE
by
Joseph G. Haubrich and Andrew W. Lo
Latest Revision: August 1989
Working Paper No. 3062-89-EFA
MASSACHUSETTS
INSTITUTE OF TECHNOLOGY
50 MEMORIAL DRIVE
CAMBRIDGE, MASSACHUSETTS 02139
lAUG
23 1989
THE SOURCES AND NATURE OF LONG-TERM MEMORY
IN
THE BUSINESS CYCLE
by
Joseph G. Haubrich and Andrew
W. Lo
Latest Revision: August 1989
Working Paper No. 3062-89-EFA
AU6 2 5ig89
RECaVED
THE SOURCES AND NATURE OF LONG-TERM MEMORY
IN
THE BUSINESS CYCLE
Joseph G. Haubrich* and Andrew
First Draft:
Latest Revision
We
May
:
W. Lo **
1988
August 1989
examine the stochastic properties of aggregate macroeconomic time
series
from the
standpoint of fractionally integrated models, and focus on the persistence of economic
shocks.
We
develop a simple macroeconomic model that exhibits long-term dependence,
a consequence of aggregation in the presence of real business cycles.
We
derive the re-
between properties of fractionally integrated macroeconomic time series and those
of microeconomic data, and discuss how fiscal policy may alter their stochastic behavior.
To implement these results empirically, we employ a test for fractionally integrated time
series based on the Hurst-Mandelbrot rescaled range. This test is robust to short-term
dependence, and is applied to quarterly and annual real GNP to determine the sources
and nature of long-term dependence in the business cycle.
lation
'Department of Finance, Wharton School, University of Pennsylvania.
** Sloan School of Management, Massachusetts Institute of Technology and
NBER.
We thank the editors and the referee for comments and we are grateful to Don Andrews, Pierre Perron, Fallaw Sowell, and
seminar participants at Columbia University, the NBER Summer Institute, the Penn Macro Lunch Group, and the University
of Rochester for useful discussions Research support from the National Science Foundation (Grant No. SES-8520054), the
Rodney L. White Fellowship at the Wharton School (Lo), the John M. Olin Fellowship at the NBER (Lo), the Batterymarch
Fellowship (Lo), and the University of Pennsylvania Research Foundation (Haubrich) is gratefully acknowledged.
1.
Introduction.
Questions about the persistence of economic shocks currently occupy an important
Most controversy has centered on whether aggregate time
place in macroeconomics.
are better approximated by fluctuations around a deterministic trend, or by a
series
random
walk plus a stationary or temporary component. The empirical results from these studies
are mixed, perhaps because the time series representation of output
wider
clciss
of processes than previously considered.
may
belong to a
much
In particular, earlier studies
have
ignored the class of fractionally integrated processes, which serves as a compromise between
the two stochastic models usually considered, and which also exhibits an interesting type
of long-range dependence. This
new approach
also accords well with the classical
business cycle program exemplified by Wesley Claire Mitchell,
trends and cycles at
Economic
all
and depression, a "rhythmical fluctuation
crises take place
Even
because the data evince
in activity."
Recurrent
roughly every 3 to 5 years and thus seem part of a non-
Studying such cycles
century macroeconomics.
cult
of
frequencies.
ing periods of prosperity
periodic cycle.
who urged examination
does not proceed smoothly: there are good times, bad times, alternat-
life
downturns and
NBER
in detail
has been the main activity of twentieth
so, isolating cycles of these frequencies
many
has been
other cycles of longer and shorter duration.
diffi-
Wesley
Mitchell (1927, p. 463) remarks "Time series also show that the cyclical fluctuations of
most
[not
sorts:
tions."
all]
economic processes occur
secular trends, primary
combination with fluctuations of several other
in
and secondary, seasonal variations, and
irregular fluctua-
Properly removing these other influences has always been controversial.
No
less
an authority than Irving Fisher (1925) considered the business cycle a myth, akin to runs
of luck at
Monte
Carlo. In a similar vein, Slutzky (1937) suggested that cycles arose from
smoothing procedures used to create the data.
A
similar debate
is
now taking
The standard methods
place.
of removing a linear or
exponential trend assume implicitly that business cycles are fluctuations around a trend.
Other work
similar to
changes.
[e.g.
Nelson and Plosser (1982)] challenges this and posits stochastic trends
random
walks, highlighting the distinction between temporary and permanent
Since empirically the cyclical or temporary component
fluctuation in the trend
like Fisher's
[as in
2.7
myth. This
the case of a
component
is
important
random
[the
random walk
for forecasting
1
-
small relative to the
part], business cycles look
more
purposes because permanent changes
walk] today have a large effect
-
is
many
periods later, whereas
8.89
temporary changes
The
large
[as in
stationary fluctuations around a trend] have small future effects.
random walk component
of aggregate output.
also provides evidence against
theoretical models
Models that focus on monetary or aggregate demand disturbances
cannot explain much output variation; supply side
as a source of transitory fluctuations
or other models
some
must be invoked
[e.g.
Nelson and Plosser (1982), Campbell and
Mankiw
(1987);.
However, the recent studies posit a misleading dichotomy.
sus
random walks, they overlook
who have
Kuznets,
earlier
In stressing trends ver-
work by Mitchell [quoted above], Adelman, and
stressed correlations in the data intermediate between secular trends
and transitory fluctuations.
In the language of the interwar
Kondratiev, Kuznets and Juglar cycles.
NBER,
they are missing
In particular, in the language of
modern time
has ignored
series analysis, the search for the stochastic properties of aggregate variables
the
clciss
of fractionally integrated processes.
These stochastic processes, midway between
white noise and a random walk, exhibit a long-range dependence no
can mimic, yet lack the permanent
effects of
an
ARIMA
of explaining the lower frequency effects, the long swings
Adelman
finite
ARMA
model
They show promise
process.
emphasized by Kuznets (1930),
(1965) or the effects which persist from one business cycle to the next. Since long
memory models
they
may
for a
more
exhibit dependence without the permanent effects of an
ARIMA
process,
not be detected by standard methods of fitting Box-Jenkins models. This calls
direct investigation of this alternative class of stochastic processes.
This paper examines the stochastic properties of aggregate output from the standpoint of fractionally integrated models.
reviewing
its
main
We
properties, advantages,
introduce this type of process in Section
and weaknesses. Section
macroeconomic model that exhibits long-term dependence. Section
for fractional integration in
Though
dence.
2.7
conclude
3 develops a simple
employs a new
time series to search for long-term dependence
related to a test of Hurst's and Mandelbrot's,
We
4
it
is
2,
test
in the data.
robust to short-term depen-
in Section 5.
-2-
8.89
Review of Fractional Techniques
2.
A random
walk can model time
in Statistics.
The
but non-periodic.
series that look cyclic
first
differences of that series [or in continuous time, the derivative] should then be white noise.
This
ries
an example of the
is
makes
it
"rougher," whereas
macroeconomic time
that
common
intuition that differencing [differentiating] a time se-
summing
[integrating]
series look like neither a
some compromise
or hybrid between the
it
makes
it
"smoother."
random walk nor white
random walk and
noise, suggesting
integral
its
Many
may
be useful.
Such a concept has been given content through the development of the fractional calculus,
i.e.,
differentiation
between
and
1
and integration
may
the ordinary integral;
than white noise.
to non-integer orders.
be viewed as a
it
filter
The
that smooths white noise to a lesser degree than
yields a series that
rougher than a random walk but smoother
is
Granger and Joyeux (1980) and Hosking (1981) develop the time
implications of fractional differencing in discrete time.
ries
fractional integral of order
have employed fractionally difference time
series,
se-
Since several recent studies
we provide only
a brief review of their
properties in Sections 2.1 and 2.2.
2.1.
Fractional Differencing.
Perhaps the most intuitive exposition of fractionally differenced time
infinite-order autoregressive
and moving-average representations. Let Xt
(l-I)'^X,
where
d
=
tf is
white noise, d
then Xt
is
is
=
series
is
via their
satisfy:
(2.1)
et
the degree of differencing, and
L
denotes the lag operator.
random walk
if
d
white noise, whereas Xt
is
a
=
1.
If
However, as Granger
and Joyeux (1980) and Hosking (1981) have shown, d need not be an
integer.
From
the
binomial theorem, we have the relation:
The idea
'
lay
dormant
only arisen
its
an old one, dating back to an oblique reference by Leibnit in 1695, but the subject
Liouville, and Riemann, developed it more fully Extensive applications have
this century; see, for example, Oldham and Spanier (1974). Kolmogorov (1940) was apparently the first to notice
of fractional differentiation
until the 19th century
in
ie
when Abel,
applications in probability and statistics.
^See, for example, Diebold and Rudebusch (1989) and Sowell (1989a, b) for further details.
2.7
- 3 -
8.89
—
(1-^)'
where
(^)
is
=
E(-l)'(^)^'
(2-2)
the binomial coefficient extended to non-integer values of d in the natural way.
By manipulating
(2.1) mechanically,
Xt also
ha.s
an infinite-order
particular, expressing the binomial coefficients in terms of the
=
X.
The
when d
is
d.
have:
(2.3)
less
For example, Granger and Joyeux (1980) and Hosking (1981) show
than
^,
Although the specification
to fractional
The
ARIMA
AR
applications,
and
and
Xt
MA
is
in (2.1)
models
is
stationary;
is
when d
is
greater than
— j, Xt
is
invertible.
a fractional integral of pure white noise, the extension
clear.
representations of fractionally differenced time series have
many
illustrate the central properties of fractional processes, particularly long-
term dependence. The
MA
coefficients Bj^ tell the effect of a
indicate the extent to which current levels of the process
this
we
function,
pl|^.
B, =
B(L)..
gamma
representation. In
particular time series properties of Xt depend intimately on the value of the differ-
encing parameter
that
=
(.-L)-^<,
MA
shock k periods ahead, and
How
depend on past values.
dependence decays furnishes valuable information about the process. Using
fast
Stirling's
approximation, we have:
Bu
for large k.
Comparing
this
-—
«
(2.4)
with the decay of an AR(l) process highlights a central
feature of fractional processes: they decay hyperbolically, at rate k
exponential rate of p
for
an AR(l). For example, compare
function of the fractionally differenced series
0.9X(_i
+
£(.
Although they both have
autocorrelation function decays
^See Hosking (1981)
2.7
1
—
L)^^^'^^Xt
Figure
=
Cf
1
rather than at the
the autocorrelation
with the AR(l) Xt
first-order autocorrelations of 0.90, the
much more
functions of these two processes. At lag
(1
in
,
rapidly. Figure
=
AR(l)'s
2a plots the impulse-response
the MA-coefficients of the fractionally differenced
for further details.
- 4 -
8.89
series
and
and the AR(l) are 0.475 and 0.900 respectively;
at lag
series
lags. Alternatively, v/e
may
autoregressive parameter will, for a given lag, yield the
fractionally differenced series (2.1). This value
in
and 0.349,
100 they are 0.048 and 0.000027. The persistence of the fractionally differenced
apparent at the longer
is
at lag 10 they are 0.158
when d —
Figure 2b for various lags
must be very
ask
of an AR(l)'s
same impulse-response
simply the
is
what value
A;-th
root of Bf^ and
is
as the
plotted
For large k, this autoregressive parameter
0.475.
close to unity.
These representations also show how standard econometric methods can
methods
fractional processes, necessitating the
of Section 4.
fail
to detect
Although a high order
ARMA
process can mimic the hyperbolic decay of a fractionally differenced series in finite samples,
the large
number
of parameters required
usual Akaike or Schwartz Criteria.
pattern with a single parameter
Hudak
explicitly fractional process, however, captures that
Granger and Joyeux (1980) and Geweke and Porter-
d.
(1983) provide empirical support for this by showing that fractional models often
out-predict fitted
The
lag
represents
ARMA
models.
polynomial B{L) provides a metric for the persistence of Xf.
GNP, which
forecast of future
given by (2.1).
Rudebusch (1989)
unexpectedly this year.
falls
GNP? To
polynomial C{L) that
is
An
would give the estimation a poor rating from the
address this issue, define
satisfies the relation (1
How much
should this change a
as the coefficients of the lag
Cj^
— L)Xt =
Suppose Xt
C{L)et, where the process Xt
One measure used by Campbell and Mankiw
(1987) and Diebold and
is:
lim B,,
=
J^C,,
=
C(l)
(2.5)
.
k=0
For large k, the value of Bj^ measures the response of Xt^k ^o ^" innovation at time
natural metric for persistence."*
and asymptotically there
is
From
(2.4),
no persistence
the autocorrelations die out very slowly.
case), but also for j
<
d
<
1,
when
it is
immediate that
for
<
<i
<
in a fractionally differenced series,
This holds true not only for d
the process
is
<
1,
C(l)
t,
=
a
0,
even though
^ (the stationary
nonstationary.
*But sec Cochrane (1988), Quah (1988), and Sowell (1989c) for opposing views.
'There has been some confusion in the hterature on this point. Geweke and Porter-Hudak (1983) argue that C(l) > 0.
They correctly point out that Granger and Joyeux (1980) have made an error, but then incorrectly claim that C(l) ;= l/r{d).
If our equation (2 4) is correct, then it is apparent that C(l) =
(which agrees with Granger (1980) and Hosking (1981)].
Therefore, the focus of the conflict lies in the approximation of the ratio r(/c + d]/r{k + 1) for large k. We have used Stirling's
approximation. However, a more elegant derivation follows from the functional analytic definition of the gamma function as
2.7
- 5 -
8.89
From
these calculations,
apparent that the long-range dependence of fractional
is
it
processes relates to the slow decay of the autocorrelations, not to any permanent effect.
This distinction
is
important; an IMA(l,l) can have small but positive persistence, but
the coefficients will never mimic the slow decay of a fractional process.
The long-range dependence
of fractionally differenced time series forces us to
some conclusions about decomposing time
modify
"permanent" and "temporary" com-
series into
ponents. Although Beveridge and Nelson (1981) show that non-stationary time series
always be expressed as the
sum
of a
random walk and a
may
stationary process, the stationary
component may exhibit long-range dependence. This suggests that the temporary compo-
may be
nent of the business cycle
practical purposes, closer to
The presence
what we think
if
Nelson and Plosser (1982) argue that deterministic detrending
the cylical component
(1986) shows that the
or trend part
for all
of as a long non-periodic cycle.
overstate the importance of the business cycle
walk. However,
is,
of fractional differencing has yet another implication for the Beveridge-
Nelson (BN) approach.
may
transitory only in the mathematical sense and
BN
decomposition
if
secular
movements
follow a
an AR(l) with a nearly unit root, McCallum
is
will assign too
and understate the contribution
of the
much
variance to the permanent
temporary component.^ But suppose
A
decom-
differencing the series the appropriate [fractional]
number
the temporary component follows the fractionally differenced time series (2.1).
position in the spirit of
BN,
i.e.,
random
of times, leaves no temporary component. However, taking first-differences of the series as
Plosser and Schwert (1978) suggest will obviously overstate the importance of the cyclical
component. For example,
order, the temporary
may
reverse
we
difference (2.1) once, instead of the appropriate fractional
component becomes
(fractional) order leaves
noise
if
— L)^~
(1
e^.
Differencing by the appropriate
no temporary component. Therefore, the presence of fractional
McCallum's
result.
the Bolution to the following recursive relation
[see, for
example, lyanaga and Kawada (1980 Section 179.A)):
r(i+l)
=
xT{x)
and the conditions;
r(i)
=
1
lin,
n
That an AR(1) with
2.7
coefficient 0.99
is in
— ao
ii^^
any sense "temporary"
- 6 -
=
1
n'Tln)
is
arguable.
8.89
Spectral Representation.
2.2.
The spectrum,
or spectral density [denoted /(w)] of a time series specifies the contri-
bution each frequency makes to the total variance. Granger (1966) and Adelman (1965)
have pointed out that most aggregate economic time series have a "typical spectral shape"
where the spectrum increases dramatically
as
w —
>
0].
Most
as the frequency approaches zero [/(cj)
—
>
oo
of the power or variance seems concentrated at low frequencies, or long
periods. However, pre-whitening or differencing the data often leads to "over-differencing"
component", and often replaces the peak by a dip at
or "zapping out the low frequency
The spectra
Fractional differencing yields an intermediate result.
exhibit peaks at
the
random
more
[unlike the flat
walk's.
A
We
persistence.
spectrum of an
fractional series hais a
0.
of fractional processes
ARM A process], but ones not so sharp as
spectrum richer
low frequency terms, and
in
by calculating the spectrum of fractionally integrated
illustrate this
white noise, and also present several formulas needed later on.
Given Xt
=
(1
— L)~
if,
the series
is
noise input, so that the spectrum of Xt
=
T.
1
The
identity
|l
— zf =
2[l
^^"^
^dT
— 2r
^TT
—
cos(t<;)]
fiu)
output of a linear system with a white
is:
a2
^(^)
clearly the
2
=
e-*",
implies that for small
=
u)
=
o'
E{ei].
we have:
c^^.
cu;-2^
(2.7)
This approximation encompasses the two extremes of a white noise
process]
and a random walk. For white
walk, d
=
1
and the spectrum
is
noise, d
=
0,
and
f{u>)
inversely proportional to
(2.6)
o;'^.
=
c,
A
a finite
ARMA
while for a
random
[or
class of processes of
current interest in the statistical physics literature, called l// noise, matches fractionally
integrated noise with d
''See Chatfield (1984,
2.7
=
I.
Chapters 6 and
9).
- 7 -
8.89
A
3.
Simple Macroeconomic Model with Long-Term Dependence.
Over half
to learn
if
a century ago,
more about economic
we approach
Models
oscillations at large
p. 230)
wrote that "We stand
and about business cycles
the problem of trends as theorists, than
empirical work."
theory.
Wesley Claire Mitchell (1927
if
we
in particular,
confine ourselves to strictly
Indeed, gaining insights beyond stylized facts requires guidance from
of long-range
dependence may provide organization and discipline
construction of economic paradigms of growth and business cycles.
They can guide
in
future
research by predicting policy effects, postulating underlying causes, and suggesting
ways to analyze and combine data. Ultimately, examining the
the
new
facts serves only as a prelude.
Economic understanding requires more than a consensus on the Wold representation
GNP;
demands a
it
falsifiable
model based on the
tastes
of
and technology of the actual
economy.
Thus, before testing
for long-range
dependence, we develop a simple model of economic
equilibrium in which aggregate output exhibits such behavior.
It
presents one reason that
macroeconomic data might show the particular stochastic structure
shows that models can
also
for
which we
restrict the fractional differencing properties of
time
test.
It
series, so
that our test holds promise for distinguishing between competing theories. Furthermore,
the majcimizing model presented below connects long-rajige dependence to central economic
concepts of productivity, aggregation, and the limits of the representative agent paradigm.
3.1.
A
Simple Real Model.
One
will
plausible
mechanism
for generating long-range
mention here and not pursue,
integrated process.
is
dependence
in
output, which we
that production shocks themselves follow a fractionally
This explanation for persistence follows that used by Kydland and
Prescott (1982). In general, such an approach begs the question, but in the present case
evidence from geophysical and meteorological records suggests that
important shocks have long run correlation properties.
for
example, find long-range dependence
many
economically
Mandelbrot and Wallis (1969),
in rainfall, riverflows,
earthquakes and weather
[measured by tree rings and sediment deposits].
A more
them
satisfactory
model explains the time
despite white noise shocks.
data by producing
This section develops such a model with long-range
dependence, using a linear quadratic version of the
2.7
series properties of
- 8 -
real business cycle
model of Long and
8.89
Plosser (1983) and aggregation results due to Granger (1980). In our multi-sector model
the output of each industry (or island) will follow an
N
sectors will not follow an
AR(l) process. Aggregate output with
AR(l) but rather an ARMA(A'',A'^-1). This makes dynamics
with even a moderate number of sectors unmanageable. Under
fairly general conditions,
however, a simple fractional process will closely approximate the true
Consider a model economy with
many goods and
and consumption plan. The
a production
version of the real business cycle model.
U =
A'^
^/9'u(C() where Ct
goods
is
u{Ct)
where
l
is
assume
B
total
may be
process
either
is
j-th entry S^^t of the
et is
t
+
i
at time
a lifetime utility function of
function u{Ct)
is
given by:
- \c[BCt
(3.1)
and
it is
Sti
=
later,
we
face a resource constraint:
Yt
(3.2)
matrix St denotes the quantity of good j invested
assumed that any good
may be consumed
Yjt
or invested.
determined by the random linear technology:
a (vector)
1.
yielding a diagonal
might occur,
for
A
=
ASt
+
(3.3)
et
random production shock whose
The matrix A
term dependence we
value
is
realized at the beginning of
consists of the input-output parameters a^,.
restrict A's form.
Thus, each sector uses only
its
To focus on
own output
matrix and allowing us to simplify notation by defining
Cj
=
long-
as input,
a^^.
This
example, with a number of distinct islands producing different goods. To
further simplify the problem,
2.7
C[i
+
NxN
Yt
period
The agent has
C[BCt = Yi ^ti^ff "^^^ agents
consumed or saved, thus:
z,
t,
chooses
agent inhabits a linear quadratic
to be diagonal so that
where the
where
who
an A^xl vector of ones. In anticipation of the aggregation considered
output Yt
Output
infinitely lived
utility
=
Ct
in
a representative agent
an A^xl vector denoting period-t consumption of each of the
our economy. Each period's
in
ARMA specification.
all
commodities are perishable and capital depreciates
- 9 -
at a
8.89
rate of 100 percent. Since the state of the
period's output and productivity shock,
economy
it is
in
each period
is
fully specified
useful to denote that vector Zt
=
by that
[V/
e[]'.
Subject to the production function and the resource constraints (3.2) and (3.3), the
agent maximizes expected lifetime
utility:
Y,0'-\{Yt-Sti)
{5,}
'
where we have substituted
maps
{St)
'
'
"-
consumption
for
naturally into a dynamic
T
=
(3.4)
t
in (3.4)
using the budget equation (3.2). This
programming formulation, with a value function
V [Zt]
and
optimality equation:
V[Zt)
=
U^.^[u[Yt-Sti) + 0E\V{Zt+^)\Zt]\
\St]
With quadratic
utility
and
K
(3.5)
.
)
linear production,
it
is
straightforward to show that V[Zi)
is
quadratic, with fixed constants given by the matrix Riccati equation that results from
Given the value function, the
the recursive definition of the value function.
first
order
conditions of the optimality equation (3.5) yield the chosen quantities of consumption and
investment/savings and have closed-form solutions.
The simple form
of the optimal consumption
and investment decision
the quadratic preferences and the linear production function.
Two
rules
comes from
qualitative features bear
emphasis. First, higher output today will increase both current consumption and current
investment, thus increasing future output. Even with 100 percent depreciation, no durable
commodities, and
i.i.d.
serial correlation.
Second, the optimal choices do not depend on the uncertainty present.
production shocks, the time-to-build feature of investment induces
This certainty equivalence feature
The time
series of
is
clearly
an artifact of the linear-quadratic combination.
output can now be calculated from the production function
and the optimal consumption/investment decision
rules.
(3.1)
Quantity dynamics then come
from the difference equation:
y,(+l
*See Sargent (1987, Chapter
2.7
1) for
-
^^-^
—
^y,t
+ K, +
e,t+i
(3.6)
an excellent exposition.
- 10 -
8.89
or
Yu^l
where
Pj
and X, are
summarized by
=
fixed constants.
(3.7)
is
a^,
which
in
The key
That
effect dies off at a rate that
mimics business
However, aggregate output, the sum across
It
sum
has an ARMA(A^,A'^
—
of
A'^
all
is
two
series,
Xt
an ARMA(2,1) process. Simple induction
With over
six million registered businesses in
America (CEA, 1988), the dynamics can be incredibly
unmanageably huge. The common response
different firms (islands)
of
independent AR(l) processes with distinct parameters
representation.
1)
sum
well-known that the
is
each AR(l) with independent error,
then implies that the
ters
depends on the
show such dependence, which we now demonstrate by applying the aggre-
gation results of Granger (1980,1988).
Yt,
dynamics
Higher output today
for a single industry or island neither
cycles nor exhibits long-run dependence.
and
qualitative property of quantity
turn depends on the underlying preferences and technology.
The simple output dynamics
sectors, will
(3.7)
£,(^1
that output Y^t follows an AR(l) process.
implies higher output in the future.
parameter
+ K, +
<^^Y^t
rich,
to this
and the number
problem
is
of
parame-
to pretend that
many
have the same AR(l) representation for output, which reduces the
dimensions of the aggregate
ARMA
An
autoregressive parameters.
This "canceling of roots" requires identical
process.
alternative approach reduces the scope of the problem by
ARMA process approximates a fractionally integrated process, and thus
the many ARMA parameters in a parsimonious manner. Though we consider
showing that the
summarizes
the ca^e of independent sectors, dependence
Consider the case of
related
A'^
for sector I's
output
sumptions on the
Y^°'
=
X^j_i Vit
.
ttj's
Y^^.
easily handled.
sectors, with the productivity shock for each serially uncor-
and independent across
the productivity coefficient
is
a^.
islands.
Furthermore,
This implies differences
One
of our key results
is
let
the sectors differ according to
in Qj,
the autoregressive parameter
that under
some
distributional as-
aggregate output Vj" follows a fractionally integrated process, where
To show
this,
we approach
this
problem from the frequency domain and
apply spectral methods which often simplify problems of aggregation.^ Let /(w) denote
the spectrum [spectral density function] of a
random
variable,
and
let z
=
e~*'^.
From
the definition of the spectrum as the Fourier transform of the autocovariance function, the
*See Theil (1954).
2.7
- 11 -
8.89
spectrum of
yj^
is:
a'
^•(")
Similarly,
=
u^^^-
''•''
independence implies that /(w), the spectrum of
The
of the individual Fit's.
q:,'s
the
Y^°', is
sum
of the spectra
mezisure an industry's average output for given input.
This attribute of the production function can be thought of as a drawing from nature, as
can the variance of the productivity shocks
for
e^t
each sector. Thus,
it
makes sense
to
think of the Oj's as independently drawn from a distribution G[a) and the Oj's drawn from
F{a). Provided that the
density of the
£,j
sum can be
shocks are independent of the distribution of
the distribution F{a)
ARMA(m,m —
1)
the spectral
written as:
/H
If
Qj's,
is
process.
=
^£[^2]
J
discrete, so that
A more
/
.
2n
--J-^dF(a)
\l
takes on
it
(3.9)
— azr
m
(<
values, F^" will be an
A'^)
general distribution leads to a process no finite
ARMA
model can represent. To further specify the process, take a particular distribution
Beta distribution.
in this case a variant of the
In particular, let
a
for
F,
be distributed as
Beta(p,g), which yields the following density function for a:
dF{a)
=
~
^(^''^
{
~
(3.10)
otherwise
with p
>
requires a
and
little
q
>
0.^^
Obtaining the Wold representation of the resulting process
more work. As a
first
step,
expand the spectrum
(3.10) by long division.
Substituting this expansion and the Beta distribution (3.10) into the expression for the
spectrum and simplifying [using the relation
z
+
z
=
2cos(a;)] yields:
'"
Granger (1980) conjectures that the particular distribution
"
For a discussion of the variety of shapes the Beta distribution takes as p and
2.7
is
not essential but only proves the result for Beta distributions.
- 12 -
q vary, see
Johnson and Kott (1970)
8.89
/M
Then
=
y^
^2
+ 2^a^os(Mj^(^c.2p-l(l-a2)9-2da.
the coefficient of cos(A;u;)
(3.11)
is:
2a^
c,2p+^-l(i
_ ^2)9-2^^
(3.12)
_
/:
Since the spectral density
the
A:-th
autocovariance of
simplifies to
coefficients
0{p
+
is
the Fourier transform of the autocovariance function, (3.12)
Yf'^.
k/2,q —
is
Furthermore, the integral defines a Beta function, so (3.12)
Dividing by the variance gives the autocorrelation
l)//?(p, g).
which reduce to
*'
r(p)
r(p
+
1
+ 0+1)
which, again using the result from Stirling's approximation T{a
proportional
(for large lags) to
integrated process of order d
—
Thus, aggregate output
k^~^.
1
—
f"
+
k)/T{b
+
A;)
«
k°'~",
is
Yf^ follows a fractionally
Furthermore, as an approximation for long
this does not necessarily rule out interesting correlations at higher, e.g.
lags,
business cycle,
frequencies. Similarly, co-movements can arise as the fractionally integrated income process
may
induce fractional integration
in
maximizing model given tastes and
In principle,
all
other observed time series.
This has arisen from a
technologies.-^
parameters of the model may be estimated, from the distribution of
production function parameters to the variance of output shocks. Though to our knowledge
no one has
explicitly estimated the distribution of production function parameters,
people have estimated production functions across industries. ^^
studies disaggregates to 45 industries.
One
many
of the better recent
For our purposes, the quantity closest to a,
is
the
''Two additional points are worth emphasizing First, the Beta distribution need not be over (0,1) to obtain these results,
only over (a,l) Second, it is indeed possible to vary the a.'s so that a, has a Beta distribution.
'^Leontief, in his classic study (1976) reports own-industry output coefficients for 10 sectors:
how much an extra
unit of food
food production These vary from 0.06 (fuel) to 1.24 (other industries).
'* Jorgenaon, GoUop and Fraumeni
(1987).
will increase
2.7
- 13 -
8.89
value-weighted intermediate product factor share. Using a translog production function,
this gives the factor share of inputs
These range from a low of 0.07
in
coming from
TV
radio and
industries, excluding labor
and
capital.
advertising to a high of 0.811 in petroleum
and coal products. Thus, even a small amount of disaggregation reveals a large dispersion
and suggests the
3.2.
plausibility
and significance of the simple model presented
in this section.
Welfare Implications.
Taking a policy perspective
of national income.
that they
live in
People
raises a natural question
about the fractional properties
Does long-term dependence have welfare implications? Do agents care
such a world?
who must
predict output or forecast sales will care about the fractional nature
of output, but fractional processes can have normative implications as well.
Lucas (1987),
regimes.
let
We
this section estimates the welfare costs of
can decide
if
the typical household
economic
people care whether their world
consume
under
different
fractional. For concreteness,
C(, evaluating this via a utility function:
oo
.l-CT
U
is
instability
Following
^
1
0'
l-ai
-E\Cl
-c
(3.14)
t=0
Also assume:
InCt
=
(l
+
A)^^^LSt
(3.15)
fc=0
where
With
rjj
Tji
=
Ine^.
The
A term measures compensation for variations in the process (^{L).
normally distributed with
between two processes
4>
and
xJj
mean
and variance
1,
the compensating fraction A
is:
oo
1
+
exp
A
Ln-o)^(4-4)
(3.16)
k=0
Evaluating this using a
2.7
realistic
(7
=
5,
again comparing an AR(l) with p
- 14 -
=
0.9
and
8.89
fractional process of order 1/4,
those
is
in
we
Lucas because the process
find that A
is
in logs
=
—0.99996
[this
number
looks larger than
rather than in levels]. ^^ For comparison, this
the difference between an AR(l) with p of 0.90 and one with p of 0.95. This calculation
provides a rough comparison only.
model generating the processes,
When
as only
it
feasible, welfare calculations
will correctly
should use the
account for important specifics,
such as labor supply or distortionary taxation.
'^
We
calculate this using (2.4) and the Hardy-Littlewood approximation for the resulting
Titchmarsh, 1951, sec
2.7
Riemann Zeta Function,
following
4 11
- 15 -
8.89
R/S
4.
The
Analysis of Real Output,
results of Section 3
dependence
show that simple aggregation may be one source of long-term
in the business cycle.
memory and apply
statistic first
it
to real
In this section
GNP. The technique
we employ a method
is
for detecting long
based on a simple generalization of a
proposed by the English hydrologist Harold Edwin Hurst (1951), which has
subsequently been refined by Mandelbrot (1972, 1975) and others.^
Our
generalization
of Mandelbrot's statistic [called the "rescaled range" or "range over standard deviation"
or R/S] enables us to distinguish between short
made
in
a sense to be
precise below.
We
define our notions of short
4.1. In Section 4.2
term dependence
and long memory and present the
we present the empirical
in log-linearly
results for real
under two
null
To develop
a
method
of detecting long
used concepts of short-term dependence
is
Section 4.3;
several
less
we
find long-
dependence
in
Monte Carlo experiments
results in Section 4.3.
Statistic.
between long-term and short-term
(1956), which
we perform
and two alternative hypotheses and report these
The Rescaled Range
tinction
GNP
test statistic in Section
detrended output, but considerably
the growth rates. To interpret these results,
4.1.
and long run dependence,
is
a measure of the decline
by successively longer spans of time.
memory, we must be
statistical
precise about the dis-
dependence. One of the most widely
the notion of "strong-mixing" due to Rosenblatt
in statistical
dependence of two events separated
Heuristically, a time series
maximal dependence between any two events becomes
trivial as
is
strong-mixing
if
the
more time elapses between
them. By controlling the rate at which the dependence between future events and those of
the distant past declines,
limit
it is
possible to extend the usual laws of large
numbers and central
theorems to dependent sequences of random variables. Such mixing conditions have
been used extensively by White (1982), White and Domowitz (1984), and Phillips (1987)
for
example, to relax the assumptions that ensure consistency and asymptotic normality of
various econometric estimators.
We
adopt this notion of short-term dependence as part of
our null hypothesis. As Phillips (1987) observes, these conditions are satisfied by a great
many
stochastic processes, including
Moreover, the inclusion of a
moment
all
Gaussian finite-order stationary
ARMA
models.
condition also allows for heterogeneously distributed
'*See Mandelbrot and Taqqu (1979) and Mandelbrot and Wallis (1968, 1969a-c).
2.7
- 16 -
8.89
sequences [such as those exhibiting heteroscedasticity], an especially important extension
in
GNP.
view of the non-stationarities of real
In contract to the "short
phenomena
natural
memory"
of
weakly dependent
often display long-term
memory
in
[i.e.,
strong-mixing] processes,
the form of non-periodic cycles.
This has lead several authors, most notably Mandelbrot, to develop stochastic models
that exhibit dependence even over very long time spans.
series
The
fractionally integrated time
models of Mandelbrot and Van Ness (1968), Granger and Joyeux (1980), and Hosking
(1981) are examples of these. Operationally, such models possess autocorrelation functions
that decay at
much
slower rates than those of weakly dependent processes, and violate
To
the conditions of strong-mixing.
detect long-term dependence [also called "strong
dependence"], Mandelbrot suggests using the range over standard deviation (R/S)
also called the "rescaled range,"
The R/S
river discharges.
from
statistic
mean, rescaled by
its
which was developed by Hurst (1951)
its
is
statistic,
in his studies of
the range of partial suras of deviations of a time series
standard deviation. In several seminal papers, Mandelbrot
demonstrates the superiority of R/S to more conventional methods of determining long-run
dependence [such as autocorrelation analysis and spectral
In testing for long
that
is
is
memory
in
analysis].
output, we employ a modification of the
R/S
statistic
robust to weak dependence. In Lo (1989), a formal sampling theory for the statistic
obtained by deriving
theorem.
We
its
limiting distribution analytically using a functional central limit
use this statistic and
its
asymptotic distribution for inference below.
Let Xt denote the first-difference of log-GNP; we assume that:
Xt
where n
is
= ^ + ^
(4.1)
an arbitrary but fixed parameter. Whether or not Xt exhibits long-term
depends on the properties of
{ft}-
As our
null hypothesis
memory
H, we assume that the sequence
of disturbances {e^} satisfies the following conditions:
{A\)
E\et\
=
for all
t.
*'See Mandelbrot {1972, 1975), Mandelbrot and Taqqu (1979), and Mandelbrot and Wallis (1968, 1969a-c).
'*Thi8 Btatietic is asymptotically equivalent to Mandelbrot's under independently and identically distributed observations,
however Lo (1989) shows that the original R/S statistic may be significantly biased toward rejection when the time series is
short-term dependent. Although aware of this bias, Mandelbrot (1972, 1975) did not correct for it since his focus was on the
relation of the R/S statistic's logarithm to the logarithm of the sample site, which involves no statistical inference; such a
relation clearly is unaffected by short-term dependence
2.7
- 17 -
8.89
{A2)
suptE\\et\^]
{A3)
a^
{A4)
{et}
=
<
oo for
limn-
is
some
/?
>
2.
strong-mixing with mixing coefficients
°°
>
and a
exists
-ni^U^if
a;,
0.
that satisfy 19
l_i
k=l
Condition [Al)
is
standard. Conditions {A2) through {A4) are restrictions on the maximal
degree of dependence and heterogeneity allowable while
still
permitting some form of the
law of large nimibers and the [functional] central limit theorem to obtain.
we have not assumed
marginal distributions of
less
than
moments
et
such as those
may
the disturbances
2,
[e.g.
Although condition
stationarity.
still
coefficients decline faster
moments
of
than l/k. However,
must decline
rules out infinite variance
(.42)
family with characteristic exponent
exhibit leptokurtosis via time-varying conditional
conditional heteroscedasticity]. Moreover, since there
conditions {A2) and {A4), the uniform
absolute
in the stable
all
if
For example,
orders [corresponding to
restrict
et
is
a trade-off between
bound on the moments may be relaxed
than {A4) requires.
we
Note that
/?
to have finite
—
>
oo],
if
we
require to
if
the mixing
et
have
then a^ must decline
moments only up
finite
feister
to order 4, then aj^
faster than l/k"^.
Conditions {Al)
—
{A4) are satisfied by
many
of the recently proposed stochastic
models of persistence, such as the stationary AR(l) with a near-unit root. Although the
distinction
between dependence
of degree, strongly
series that
in the short versus long
dependent processes behave so
diff"erently
our dichotomy seems most natural.
strongly dependent processes are either
sums do not converge
graphically, their behavior
is
marked by
may appear
to be a matter
from weakly dependent time
For example, the spectral densities of
unbounded
in distribution at the
runs
same
or zero at frequency zero. Their partial
rate as weakly dependent series.
cyclic patterns of all kinds,
some that
And
are virtually
indistinguishable from trends. '^^
Consider a sample Xi, X2, ...,Xn and
let
the modified re-scaled range statistic, which
Xn
we
denote the sample
shall call
Qn,
is
mean ^
^
X,. Then
given by:
'*For the precise definition of strong-mixing, and for further details, see Rosenblatt (1956), White (1984), and the papers
Eberlein and Taqqu (1986).
'°See Herrndorf (1985).
moments
are not required.
Note that one of Mandelbrot's (1972) arguments in favor of R/S analysis is that finite second
This is indeed the case if we are interested only in the almost sure convergence of the statistic.
However, since we wish to derive its limiting distribution
^'See Mandelbrot (1972) for further details
2.7
in
for
purposes of inference, a stronger
- 18 -
moment
condition
is
needed.
8.89
Qn
(4.2)
^n{g)
- -
- -
;=1
3=1
where
^E(^;-^n)' + ^E'^;(9){ i: (X.-^n)(X._,-X,)
=
lig)
-
and a^ and
^2
+ 2^u;j(9)7y
^. are the usual
w,(g)
Xj from
its
of squared deviations of X., but also
those suggested by
always positive.
q
9+1
<
n.
its
Theorem
^,1(9) involves
weighted autocovariances up to lag
Newey and West
Qn
mean, normalized by an estimator
The estimator
partial sum's standard deviation divided by n.
is
1
sample variance and autocovariance estimators of X.
range of partial sums of deviations of
u)j{q) are
=
(4.3)
I
(1987),
and
is
the
of the
not only sums
q;
the weights
yields an estimator o^iq)- that
4.2 of Phillips (1987) demonstrates the consistency o( an{q)
under the following conditions:
<
00 for
some /?>
{A2')
supt
{A5)
As n increases without bound, q
E\\et\'^^]
2.
also increases without
bound such that
q
~
o(nV4).
The
choice of the truncation lag g
[but at a slower rate than] the
becomes large
dramatically.
relative to the
However,
autocorrelations
may
q
is
a delicate matter.
sample
number
size,
d„(9)
ie
2.7
q
must increase with
Monte Carlo evidence suggests that when
of observations, asymptotic approximations
cannot be chosen too small otherwise the
not be captured. The choice of q
must therefore be chosen with some consideration
^^ See, for
Although
also an estimator of the spectral density function of Xi
example, Lo sind MacKinlay (1989)
-19-
at
is
clearly
may
q
fail
effects of higher-order
an empirical issue and
of the data at hand.
frequency lero, using a Bartlett window.
8.89
If
the observations are independently and identically distributed with variance o^, our
normalization by dn{^)
deviation estimator Sn
asymptotically equivalent to normalizing by the usual standard
is
=
~
[^ IZi(-^j
p
-X'n)
resulting statistic, which
we
Max
Qr
Qn,
call
and Mandelbrot (1972):
precisely the one proposed by Hurst (1951)
is
The
•
(4.4)
l</t<
To perform
we
require
its
with the standardized re-scaled range F^
statistical inference
distribution.
Although
finite-sample distribution
its
sample approximation has been derived
(A3)
(-45). In particular,
variable
we
call
V
,
in
is
=
not apparent, a large-
Lo (1989) under assumptions {Al),
(^2'),
the limiting distribution of V^, which corresponds to a
has the following
and
c.d.f.
Qn/y/^,
and
random
p.d.f.:
oo
=
Fv{v)
+ 2^(l-4fc2v2)e-2(H^
1
(4.5)
Jb=l
(4.6)
k=l
Using Fy
,
critical values
may
readily be calculated for tests of any significance level.
most commonly used values are reported
computed using fy\
the
it
is
mean and standard
in
Table
1.
The moments
V
V
are also easily
= ^,
thus
are approximately 1.25 and 0.27 respectively.
The
straightforward to show that E\V]
deviation of
of
The
= y/^ and
^^[V^]
distribution and density functions are plotted in Figure 3. Observe that the distribution
positively
is
skewed and most of
its
mass
falls
between | and
Although Fy completely characterizes the rescaled range
pothesis of short-range dependence,
pendence
is
considerably different.
its
2.
statistic
under the
null hy-
behavior under the alternative of long-range de-
Lo (1989) shows that
for the fractionally differenced
alternative (2.1), the normalized rescaled range V^ diverges to infinity in probability
d G (0,2) and converges to zero
against (2.1), a test for long-term
'*In fact, Lo (1989) shows that such a test
2.7
is
in probability
when d G (— j,0).
memory based on Vn
is
consistent against a considerably
-20
when
Therefore, at least
consistent.
more general
class of alternatives.
8.89
4.2.
Empirical Results for Real Output.
We
GNP
apply our test to two time series of real output: quarterly postwar real
from
1947:1 to 1987:4, and the annual Friedman and Schwartz (1982) series from 1869 to 1972.
The
results are reported in Table 2.
These
the classical rescaled range V^ which
row of numerical entries are estimates
first
not robust to short-term dependence.
is
eight rows are estimates of the modified rescaled range Vn{q) for values of q
Recall that q
zero.
bieis
first
and
computed
is
column of nxmierical
short-term dependence for the
of
q.
The
for the
Friedman and Schwartz
with values of q from
beyond
4 are used,
to
8.
l]-
log-GNP cannot be
third
rejected for any value
supports the null hypothesis.
When we
we no
The
results
GNP,
log-linearly detrend real
column of numerical
entries in Table 2
show
be rejected for log-linearly detrended quarterly output
rejections are
weaker
for larger q
not surprising
is
When
values
longer reject the null hypothesis at the 5 percent level of
Friedman and Schwartz time
series,
we only
reject
with the
range and with Vn{l).
classical rescaled
tests
~
[(^n/^n(9))
from estimating higher-order autocorrelations.
significance. Finally, using the
The
statistic also
That the
to 4.
1
since additional noise arises
of q
1
entries in Table 2 indicate that the null hypothesis of
The
may
that short-term dependence
•
series are similar.
considerably.
diff"er
as 100
first-diff"erence of
range
classical rescaled
the results
from
parentheses below the entries for Vn{q) are estimates of the percentage
in
of the statistic Vn,
The
The next
the truncation lag of the estimator of the spectral density at frequency
is
Reported
of
on the
diff"erenced series
ought to be particularly striking, especially since
the classical version [Vn(0)] rejects too often, on evidence of merely short term dependence.
Even over-differencing
its first diff"erence,
not a problem:
and the
The detrended
cance disappears
is
series
when
test
more
may
2.7
not adequately control for short term dependence.
suggested by Nelson and
Kang
(1981).
GNP
is
In
pre-
Their results are
show that inappropriate detrending introduces a great deal
into low frequencies.
Taken
dence.
of dependence, but even there the signifi-
significant autocorrelations in log-linearly detrended
particularly cogent since they
power
ought to pick that up as well.
only a year of lags are included. Partly due to decreasing size and
cisely the spurious periodicity
of
the series has a fractional component, so will
show more evidence
power, at these lags the test
addition, finding
if
as a
whole then, our results accept the
null hypothesis of short
Equivalently, they reject the alternative of long term dependence;
- 21 -
term depen-
GNP
has no
8.89
fractional
component. This finding deserves some notice simply because of the many pre-
From the path-breaking
vious attempts (direct or indirect) to resolve the issue.
work
of
early
Mandelbrot and Wallis (1969b) and Adelman (1963) to the sophisticated recent
estimates of Diebold and Rudebusch (1989), econometricians have not conclusively established the existence or non-existence of long
term dependence.
and Rudebusch have rather large standard
of Diebold
In particular the estimates
errors, leaving
precise conclusions about the existence of fractional differencing.
tighter distribution,
allow
more
and
in
Though by no means the
economy, whether
test statistic has a
how
final
work on the subject, our
term dependence.
tests present strong evidence against long
is
Our
to reach
conjunction with the size and power results reported below,
definite conclusions.
Our main concern
them unable
these results improve our ability to model the aggregate
purely statistical sense of estimating the right stochcistic process
in the
or in the theoretical sense of developing the appropriate equilibrium structure.
perspective, finding only short term dependence simplifies
From
either
some matters while complicating
others.
From
the standpoint of obtaining a correct statistical representation, our results sub-
stantially circumscribe the class of stochastic processes
simplifies the problem,
whether the goal be to measure the persistence of economic shocks,
assess the welfare costs of
economic
instability, or derive the correct theory.
models of trend (perhaps stochastic) and
tic
structure of
needed to characterize GNP. This
GNP. On
finite
ARMA
The standard
adequately capture the stochas-
the other hand, rejecting fractional processes removes a simple
explanation to the problems plaguing this area.
The
intriguing possibility that the con-
tending sides had each misread the true, fractional, nature of
GNP
unfortunately
is
false,
and the disagreement must have another, perhaps deeper source.
These
results also confirm the unit root findings of
Campbell and Mankiw (1987),
Nelson and Plosser (1982), Perron and Phillips (1987) and Stock and Watson (1986).
cannot reject the null hypothesis of a
As mentioned
diff"erence stationary short
stationary noise model of
GNP
to reject the null hypothesis
observe that
if
log-GNP
* Of courBe, this may be
Monte Carlo experiments
2.7
term dependent process.
before, the significant autocorrelations appearing in detrended
the spurious periodicity suggested by Nelson and
yt
is
is
Kang
We
GNP
indicate
(1981). Moreover, the trend plus
not contained in our null hypothesis, hence our failure
also consistent with the unit root model. ^^
were trend stationary,
To
see this,
i.e.:
the result of low power against stationary but near-integrated processes, and must be addressed by
- 22 -
8.89
=
yt
where
€(
=
rjt
-
T]t
is
stationary white noise, then
T]i^i.
But
Q +
/3<
+
its first-difference
this innovations process violates
(4.7)
rjt
Xt
is
simply Xt
=
+
our assumption (A3) and
et
where
therefore
is
not contained in our null hypothesis.
From
the perspective of a theorist, the tests in Table 2 provide
more information than
usual because they also serve to reject a particular model. Recall the section's main point;
moving
to a multi-sector
test provides
no evidence
model can produce qualitatively
for those
does not qualitatively affect
its
different
output dynamics. Our
dynamics. The American economy's multi-sector nature
output dynamics. For
many
questions, single sector models
provide a rich enough environment. Once again this represents a simplification;
us to avoid a broad and
As
difficult class of
sectors, rejecting the predictions of a multi-sector
tential puzzle. Currently the strong
allows
models.
before, the results also complicate the theoretical picture. Because the
economy has many
it
American
model presents a po-
assumptions needed to produce long term dependence
take the edge off any puzzle, but these are only sufficient conditions. Necessary conditions,
unknown.
or broader sufficient conditions, are
We
don't
know enough about what
pro-
duces fractional differencing or how closely actual conditions match those requirements to
worry about not finding
it.
If
broader conditions produce long term dependence - and
if
actual sector dynamics meet those conditions - then a genuine puzzle arises.
Related research
GNP
may
also turn
up puzzles, making the
lack of fractional processes in
anomalous. For example Haubrich (1989) finds fractional processes
across several countries.
The output
in
consumption
results here therefore rule out several natural expla-
nations, such as a constant marginal propensity to
consume out
of fractionally differenced
income.
From another
tempts.
many
perspective, our results impose a discipline on future modelling at-
As Singleton (1988) points out, dynamic macro models make predictions over
frequencies: seasonal, cyclical,
the facts on long term dependence.
and longer. Their predictions must now conform to
Future dynamic models - multi-sector or otherwise
- must respect this constraint, or bear the defect of having already been rejected.
restrictive this will be
is
unknown.
To conclude that the data support the
reject
2.7
it
is,
How
null hypothesis
of course, premature since the size
- 23 -
and power
because our statistic
fails to
of our test in finite samples
is
8.89
yet to be determined.
We
perform
Monte Carlo experiments and report the
illustrative
results in the next section.
4.3.
The
Size
and Power of the
To evaluate the
trative
the
size
and power
Monte Carlo experiments
number
Test.
of our test in finite samples,
for a
sample
of quarterly observations of real
we perform
size of 163 observations,
GNP
several illus-
corresponding to
growth from 1947:2 to
We
1987:4.'^^
simulate two null hypotheses: independently and identically distributed increments, and
increments that follow an
the
mean and standard
ARMA(2,2)
process.
deviation of our
standard deviation of our quarterly data
To choose parameter values
{\
-
4>iL
-
using nonlinear least squares.
=
set:
^l
+
{\
i.i.d.
deviates to
match the sample mean and
null hypothesis,
fix
7.9775 x 10~^ and 1.0937 x 10~^ respectively.
ARMA(2,2)
for the
4>2L^)yt
random
we
Under the
simulation,
+ eiL +
we estimate
tt~WN[0,ol)
d2L^)it
The parameter estimates
the model:
(4.8)
are [with standard errors in paren-
theses]:
4>i
=
0.5837
Oi
=
(0.1949)
4>2
=^
(0.1736)
- 0.4844
$2
=
(0.1623)
ti
=
- 0.2825
0.6518
(0.1162)
0.0072
(0.0016)
a1
=
0.0102
Table 3 reports the results of both null simulations.
It is
apparent from the
"I.I.D. Null"
Panel of Table 3 that the 5 percent test based on
the classical rescaled range rejects too frequently.
^* All simulations were performed in double precision on a
each experiment was comprised of 10,000 replications.
VAX
2.7
- 24 -
The
8700 using the
5 percent test using the modified
IMSL
10.0
random number generator DRNNOA;
8.89
=
rescaled range with q
number
estimator d^{q)
it
the size of a 5 percent test based on the classical rescaled range
is
34 percent, whereas the corresponding size using the modified
is
4.8 percent.
As
As the
apparent that modifying the rescaled range by the spectral density
is
critical;
is
size.
becomes more conservative. Under the ARMA(2,2)
of lags increcises to 8, the test
null hypothesis,
nominal
3 rejects 4.6 percent of the time, closer to its
before, the test
R/S
statistic
becomes more conservative when q
is
=
with q
5
increased.
Table 3 also reports the size of tests using the modified rescaled range when the
lag length q
is
chosen optimally using Andrews' (1987) procedure. This data-dependent
procedure entails computing the first-order autocorrelation coefficient p(l) and then setting
the lag length to be the integer- value of
Under the
null,
i.i.d.
percent; under the
may
77-^(l-p2)2
=
&
Andrews' formula
ARMA(2,2)
significantly different
formula
K^
^
A^n
Mn, where.
(4-9)
yields a 5 percent test with empirical size 6.9
alternative, the corresponding size
from the nominal value, the empirical
is
4.1 percent.
size of tests
not be economically important. In addition to
Although
based on Andrews'
optimality properties, the
its
procedure has the advantage of eliminating a dimension of arbitrariness
in
performing the
test.
Table 4 reports power simulations under two fractionally differenced alternatives:
L)
£(
=
function
rji
where d
7e(fc)
of
e^
is
=
1/3,-1/3.
given by:
k)
2
Realizations of fractionally differenced time series
[of
multiplying vectors of independent standard normal
factorization of the 163 x 163 covariance matrix
calibrate the simulations,
a"^
is
2.7
-^
[j
=
l,...,fl),
random
q is
an integer and
- 25 -
Mn
^
variates by the Cholesky-
entries are given
chosen to yield unit variance
where
,
length 163] are simulated by pre-
GNP
'^In addition, Andrews' procedure requires weighting the autocovariances by
-
l^
1
/
,
whose
multiplied by the sample standard deviation of real
1
—
Hosking (1981) has shown that the autocovariance
T(l-2d)T{d +
and West's (1987)
(1
1
-
€t's,
by
(4.10).
the {e^} series
is
To
then
growth from 1947:1 to 1987:4,
j^
(j
=
1,
.
.
.
,
[A/„]) in contrast to
Newey
need not be.
8.89
and to
this series
period.
The
added the sample mean of
is
resulting time series
is
real
GNP
growth over the same sample
used to compute the power of the rescaled range; Table
4 reports the results.
For small values of
power against both
tests
q,
based on the modified rescaled range have reasonable
fractionally differenced alternatives. For example, using one lag the 5
=
percent test against the d
alternative this test hcis 81.1 percent power.
power
declines.
Note that
d=
1/3 alternative has 58.7 percent power; against the
tests baised
on the
powerful than those using the modified
R/S
As the
lag length
classical rescaled
increased, the test's
is
range
however,
statistic. This,
—1/3
is
is
significantly
more
value
when
of
little
distinguishing between long-term versus short-term dependence since the test using the
classical statistic also has
power against some stationary
Finally, note that tests using
against the d
d
—
= —1/3
ARMA
processes.
Andrews' truncation lag formula have reasonable power
alternative but are considerably weaker against the
more
relevant
1/3 alternative.
The simulation evidence
in
Tables 3 and 4 suggest that our empirical results do indeed
support the short-term dependence of
null hypothesis does not
alternatives.
additional
test's size
little
is
finite-order
Of
seem
GNP
with a unit root.
to be explicable by a lack of
course, our simulations were illustrative
Monte Carlo experiments must be performed
and power
is
Our
failure to reject the
power against long-memory
and by no means exhaustive;
before a
full
assessment of the
complete. Nevertheless our modest simulations indicate that there
empirical evidence in favor of long-term
memory
in
GNP
growth
rates.
the direct estimation of long-memory models would yield stronger results and
is
Perhaps
currently
being investigated by several authors.'^®
5.
Conclusion.
This paper hzs suggested a new approach to the stochastic structure of aggregate
output.
Traditional dissatisfaction with the conventional methods - from observations
about the typical spectral shape of economic time
periods - calls for such a reformulation.
series, to
the discovery of cycles at
all
Indeed, recent controversy over deterministic
versus stochastic trends and the persistence of shocks underscores the difficulties even
modern methods have
of identifying the long
run properties of the data.
^'See, for example, Diebold and Rudebusch (1989), Sowell (I989a,b), and Yajima (1985,1988).
2.7
- 26 -
8.89
Fractionally integrated
random
processes provide one explicit approach to the prob-
lem of long-term dependence; naming and characterizing
studying the problem
it
Controlling for
from trends and to
late business cycles
extent that
scientifically.
Jissess
its
this aspect
the
is
first
step in
presence improves our ability to
the propriety of that decomposition.
explains output, long-term dependence deserves study in
own
its
iso-
To the
right. Fur-
thermore, Singleton (1988) has recently pointed out that dynamic macroeconomic models
often link inextricably predictions about business cycles, trends,
too linked
in a
is
and seasonal
effects.
So
long-term dependence: a fractionally integrated process arises quite naturally
dynamic
linear
model via aggregation. This model not only predicts the existence of
fractional noise, but also suggests the character of
its
parameters. This
on the nature of long-term dependence
leads to testable restrictions
in
clciss
of models
aggregate data, and
also holds the promise of policy evaluation.
Advocating a new
were intractable.
class of stochastic processes
would be a
fruitless task
if its
In fact, manipulating such processes causes few problems.
members
We
con-
structed an optimizing linear dynamic model that exhibits fractionally integrated noise,
and provided an
explicit test for such long-term
and Mandelbrot gives us a
R/S
statistic possesses
trative
statistic
dependence. Modifying a
statistic of
Hurst
robust to short-term dependence, and this modified
a well-defined limiting distribution which
computer simulations indicate that
this test
we have
tabulated. Illus-
has power against at least two specific
alternative hypotheses of long-memory.
Two main
conclusions arise from the empirical work and
First, the evidence does not
term dependence
support long-term dependence
null hypothesis occur only
in
Monte Carlo experiments.
GNP.
Rejections of the short-
with detrended data, and
is
consistent with
the well-known problem of spurious periodicities induced by log-linear detrending. Second,
since a trend-stationary
may
model
is
not contained in our null hypothesis, our failure to reject
also be viewed as supporting the first-difference stationary
additional result that the resulting stationary process
is
model of GNP, with the
weakly dependent at most. This
supports and extends the conclusion of Adelman that, at least within the confines of the
available data, there
2.7
is little
evidence of long-term dependence in the business cycle.
- 27 -
8.89
References
Adelman, Irma (1956): "Long Cycles: Fact or Artefact?" American Economic Review
55,
444-463.
Andrews, Donald (1987): "Heteroskedasticity and Autocorrelation Consistent Covariance
Matrix Estimation," Working Paper, Cowles Foundation, Yale University.
Beveridge, Stephen, and Charles R. Nelson (1981): "A New Approach to Decomposition
of Economic Time Series into Permanent and Transitory Components with Particular
Attention to Measurement of the 'Business Cycle'," Journal of Monetary Economics 7,
151-174.
Billingsley, Patrick (1968):
New
Convergence of Probability Measures.
York: John Wiley
and Sons.
-,
Weak Convergence
(1971):
Regional Conference Series
in
of Measures: Applications in Probability,
Applied Mathematics.
S.I. A.M.:
CBMS-NSF
Philadelphia.
Campbell, John Y., and N. Gregory Mankiw (1987): "Are Output Fluctuations Transitory?" Quarterly Journal of Economics 102, 857-880.
Chatfield, Christopher (1984): The Analysis of
tion, Chapman and Hall, New York.
Cochrane, John (1988): "How Big
Economy 96, 893-920.
is
the
Time
Series:
Random Walk
in
An
Introduction, Third Edi-
GNP?," Journal
of Political
Council of Economic Advisors (1988): Economic Report of the President, U.S. Government
Printing Office.
Diebold, Francis X., and Glenn D. Rudebusch (1989): " Long Memory and Persistence
Aggregate Output" to appear in Journal of Monetary Economics.
in
Eberlein, Ernst and Murad Taqqu (1986): Dependence in Probability and Statistics. Progress
in Probability and Statistics 11. Birkhauser: Boston.
Mark Kac (1946): "On Certain Limit Theorems in the Theory of ProbaBulletin of the American Mathematical Society 52, 292-302.
Erdos, Paul and
bility,"
-,
(1947): "On the Number of Positive Sums of Independent
Bulletin of the American Mathematical Society 53, 1011-1020.
Feller,
Random
Variables,"
William (1951): "The Asymptotic Distribution of the Range of Sums of Independent
Variables," Annals of Mathematical Statistics 22, 427-432.
Random
Fisher, Irving (1925):
the
American
"Our Unstable Dollar and the So-called Business Cycle", Journal
Friedman, Milton, and Anna
the United
Kingdom,
J.
NBER
Schwartz (1982): Monetary Trends in The United State
Monograph, University of Chicago Press, Chicago.
Geweke, John, and Susan Porter-Hudak (1983):
2.7
of
Statistical Association 20, 179-202.
- 28 -
"
The Estimation and Application
of
Long
8.89
Memory Time
series
Models" Journal of Time Series Analysis
Granger, Clive \V. J. (1969):
Econometrica 37, 150-161.
of
4,
221-238.
"The Typical Spectral Shape of an Economic Variable,"
(1980): "Long Memory Relations and the Aggregation of Dynamic Models" Journal
Econometrics 14, 227-238.
(1988): "Aggregation of Time Series Variables-A Survey" Federal Reserve
of Minneapolis Institute for Empirical Macroeconomics, Discussion paper 1.
Bank
and Roselyne Joyeux (1980): "An Introduction to Long-Memory Time
Models and Fractional Differencing", Journal of Time Series Analysis 1,14-29.
Series
-,
Old and New
Haubrich, Joseph G. (1989): "Consumption and Fractional-Differencing:
Anomalies," Rodney L. White Center Working Paper No. 26-89.
Herrndorf, Norbert (1984): "A Functional Central Limit
Sequences of
,
Random
Theorem
for
Weakly Dependent
Variables," Annals of Probability 12, 141-153.
(1985): "A Functional Central Limit Theorem for Strongly Mixing Sequences of
Variables," Z. Wahrscheinlichkeitstheorie verw. Gebiete, 541-550.
Random
Hosking, J.R.M. (1981): "Fractional Differencing," Biometrika 68, 165-176.
Hurst, Harold E. (1951): "Long Term Storage Capacity of Reservoirs,"
the American Society of Civil Engineers 116, 770-799.
Transactions of
lyanaga, Shokichi, and Yukiyosi Kawada, editors (1977): Encyclopedic Dictionary of Mathematics, Mathematical Society of Japan, MIT Press, Cambridge.
Jorgenson, Dale W., Frank M. Gollop, and Barbara M. Fraumeni (1987): Productivity and
U.S. Economic Growth, Harvard Economic Studies vol 159, Harvard University Press,
Cambridge.
Kolmogorov, Andrei N. (1940): "Wienersche Spiralen und einige andere interessante Kurven im Hilberteschen Raum," C. R. Doklady Acad. Sci. URSS (n.s.) 26, 115-118.
Kuznets, Simon (1965); Economic Growth and Structure, Norton,
New
York.
Kydland, Finn, and Edward C. Prescott (1982): "Time to Build and Aggregate Economic
Fluctuations", Econometrica 50, 1345-1370.
Leontief, Wassily W. (1976): The Structure of the American Economy 1919-19S9, second
enlarged edition. International Arts and Sciences Press Inc., White Plains.
Lo,
Andrew W.
(1989):
"Long-Term Memory
in
Stock Market Prices," Working Paper,
Sloan School of Management, M.I.T.
and A. Craig MacKinlay (1989): "The Size and Power of the Variance Ratio Test in
A Monte Carlo Investigation," Journal of Econometrics 40, 203-238.
Finite Samples:
Long, John B.
Political
Jr.
Lucas, Robert E.
2.7
and Charles
Economy
Jr.
I.
Plosser (1983):
"Real Business Cycles", Journal of
91, 39-69.
(1987):
Models of Business Cycles, Yrjo Jahnsson Lectures, Basil
- 29 -
8.89
Blackwell,
New
York.
Mandelbrot, Benoit (1972): "Statistical Methodology for Non-Periodic Cycles: From the
Covariance to R/S Analysis," Annals of Economic and Social Measurement 1, 259-290.
"Limit Theorems on the Self-Normalized Range for Weakly and Strongly
-, (1975):
Dependent Processes," Z. Wahrscheinlichkeitstheorie verw. Gebiete 31, 271-285.
-, and Murad Taqqu (1979):
"Robust R/S Analysis of Long-Run Serial Correlation,"
Bulletin of the International Statistical Institute 48, Book 2, 59-104.
-, and John Van Ness (1968), "Fractional Brownian Motion, Fractional Noises and
Applications," S.I. A.M. Review 10, 422-437.
,
and James Wallis (1968):
Resources Research
4,
"Noah, Joseph and Operational Hydrology,"
"Computer Experiments with Fractional Gaussian Noises. Parts
-, (1969a):
Water Resources Research 5, 228-267.
-,
(1969b):
Research
-,
5,
"Some Long Run Properties
of Geophysical Records,"
1, 2, 3,"
Water Resources
321-340.
(1969c): "Robustness of the Rescaled
Long Run
Water
909-918.
Statistical
Range R/S
in
the Measurement of Noncyclic
Dependence," Water Resources Research
5,
967-988.
McCallum, Bennett T. (1986): "On 'Real' and 'Sticky Price' Theories
Cycle", Journal of Money, Credit, and Banking 18, 397-414.
of the Business
NBER
Mitchell, Wesley Claire (1927): Business Cycles: the problem and its setting,
ies in Business Cycles No.l National Bureau of Economic Research, New York.
Nelson, Charles R. and Heejoon
Detrended Time
Series,"
Kang
(1981):
Stud-
"Spurious Periodicity in Inappropriately
Econometrica 49, 741-751.
Nelson, Charles R. and Charles L Plosser (1982): "Trends and Random Walks in Macroeconomic Time Series: some evidence and implications", Journal of Monetary Economics
10, 139-162.
Newey, Whitney K. and Kenneth D. West (1987): "A Simple Positive Definite, Heteroscedasticity and Autocorrelation Consistent Covariance Matrix," Econometrica 55,
703-705.
Oldham, Keith
New
B.,
and Jerome Spanier
(1974):
The Fractional Calculus, Academic Press,
York.
Perron, Pierre and Peter C. B. Phillips (1987): "Does
Letters 23, 139-145.
Phillips, Peter C. B. (1987):
55, 277-301.
"Time
GNP
Have a Unit Root?" Economics
Series Regression with a Unit Root,"
Econometrica
and G. William Schwert (1978): "Money, Income, and Sunspots: Measuring Economic Relationships and the Effects of Differencing", Journal of Monetary
Economics 4, 637-660.
Plosser, Charles L,
2.7
- 30 -
8.89
Quah, Dannv
Series?"
(1987):
NBER
"What Do We Learn From Unit Roots
in
Macroeconomic Time
Working Paper No. 2450.
Rosenblatt, Murray (1956): "A Central Limit Theorem and a Strong Mixing Condition,"
Proceedings of the National Academy of Sciences 42, 43-47.
Sargent,
Thomas
J.
(1987):
Dynamic Macroeconomic Theory, Harvard University
Press,
Cambridge.
Kenneth J. (1988): "Econometric Issues in the Analysis of Equilibrium Business
Cycle Models," Journal of Monetary Economics 21, 361-386.
Singleton,
Slutzky, Eugene (1937): "The Summation of
Processes", Econometrica 5, 105-146.
Random Causes
as the Source of Cyclic
Sowell, Fallaw. (1989a): "Fractional Unit Root Distributions," Working Paper,
School of Industrial Administration, Carnegie-Mellon University.
Graduate
(1989b): "Maximum Likelihood Estimation of Fractionally Integrated Time Series
Models," Working Paper, Graduate School of Industrial Administration, CarnegieMellon University.
,
(1989c): "The Deterministic Trend in Real GNP," Working Paper, Graduate
-,
School of Industrial Administration, Carnegie-Mellon University.
Stock, J. and
147-151.
M. Watson
Theil, Henri (1954):
(1986): "Does
GNP
Have a Unit Root?" Economics
Letters 22,
Linear Aggregation of Economic Relations, North-Holland, Amster-
dam.
Titchmarsh, E.C. (1951):
Oxford.
White, Halbert (1980):
and a Direct Test
,
The Theory of the Riemann
Z eta- Function,
Claredon Press,
"A Heteroscedasticity-Consistent Covariance Matrix Estimator
for Heteroscedasticity,"
Econometrica 48, 817-838.
(1984): Asymptotic Theory for Econometricians.
John Wiley and Sons, New York.
and I. Domowitz (1984): "Nonlinear Regression with Dependent Observations,"
Econometrica 52, 143-162.
,
Yajima, Yoshihiro (1985): "On Estimation of Long-Memory Time Series Models," Australian Journal of Statistics, 303-320.
-,
(1988): "On Estimation of a Regression
Errors," Annals of Statistics 16, 791-807.
2.7
- 31 -
Model with Long-Memory Stationary
8.89
C-t:l
Creoteo b> Aoo'e**
o
w
C'eoteo b* Anorcf w
_C
Ap*
CRrt-Ej g.
Table
P{V <
v)
la. Fractiles of the Distribution
Fv{v).
Table
R
GNP;
2.
GNP
from 1947:1 to
1987:4, and Ay*^ indicates the first-differences of the logarithm of real GNP. y^^^ and Ay^^ are
defined similairy for the Friedman and Schwartz series. The classical rescaled range V„, and the
'S analysis of real
y^^^
indicates log-linearly detrended quarterly real
modified rescaled range Vn{q) are reported.
Table
3.
R/S statistic under i.i.d and ARMA(2,2) null hypotheses for the first-difference of
log-GNP The Monte Carlo experiments under the two null hypotheses are independent and consist of 10,000 replications
each Parameters of the i.i.d. simulations were chosen to match the sample mean and variance of quarterly real GNP growth
rates from 1947 1 to 1987:4, parameters of the ARMA(2,2) were chosen to match point estimate* of an ARMA(2,2) model
fitted to the same data set Entries m the column labelled "q' indicate the number of lags used to compute the R/S statistic;
corresponds to Mandelbrot's classical rescaled range, and a non-integer lag value corresponds to the average (across
a lag of
replications) lag value used according to Andrew's (1987) optimal lag formula Standard errors for the empirical sire may be
computed using the usual normal approximation; they are 9.95 x 10~*, 2.18 x 10~^, and 3.00 x 10"^ for the 1, 5, and 10
Finite sample distribution of the modified
real
percent tests respectively.
I.I.D. Null
Hypothesis:
n
Table*.
Power of the modified R/S statistic under a Gaussian fractionally differenced alternative with differencing parameters d =
1/3, —1/3 The Monte Carlo experiments under the two alternative hypotheses are independent and consist of 10,000 replications
each Parameters of the simulations were choeen to match the sample mean and variance of quarterly real GNP growth rates
from 19471 to 1987:4 Entries in the column labelled 'q" indicate the number of lags used to compute the R/S statistic; a
lag of
corresponds to Mandelbrot's classical rescaled range, and a non-integer lag value corresponds to the average (across
Andrew's (1987) optimal lag formula.
replications) lag value used according to
1/3:
n
59^0
1)31
\
Date Due
DEC
kiV
1990
16 1993
WP''''
Lib-26-67
MIT
3
iiBRARiE"; nnpL
i
TDSD DD57DM35
5
Download