Differences-in-Differences and A (Very) Brief Introduction to

advertisement
Unit Roots & Forecasting
Methods of Economic
Investigation
Lecture 20
Last Time

Descriptive Time Series



Processes
Estimating with exogenous serial correlation
Estimating with endogenous processes
Today’s Class

Non-stationaryTime Series



Unit Roots and Spurious Regressions
Orders of Integration
Returning to Causal Effects


Impulse Response Functions
Forecasting
Random Walk Processes

Definition:




Et[xt+1] = xt that is today’s value of X is the
best predictor of tomorrow’s value.
This looks very similar to our AR(1) processes,
where φ = 1.
Autocovariances of a random walk are not well
defined in a technical sense, but imagine AR(1)
process with φ1: we have nearly perfect
autocorrelation for any two time periods.
persistence dies out too slowly so most of
variance will largely be due to very lowfrequency “shocks.”
Permanence of Shocks in Unit Root
An innovation (a shock at t ) to a stationary
AR process dies out eventually (the
autocorrelation function declines eventually
to zero).
 A shock to a random walk is permanent

xt  xt 1   t
 xt  2   t   t 1
t
 x0    i 1
i 1

Variance is increasing over time

Var(xt) = Var(x0) + tσ2
Drifts and Trends

Deterministic trend




yt = δt + xt + εt
xt is some stationary process
yt is “trend” stationary
It’s easy to add a deterministic trend to a
random walk
yt  yt 1    ut
t
 y0  t   ui 1
i 1
Orders of Integration

A series is integrated of order p if a p
differences render it stationary.


If a time series is integrated and differencing
once renders the time series stationary, then it
is integrated of order 1 or I(1).
If it is necessary to difference twice before a
time series is stationary, then it is I(2), and so
forth.
Integrated Series

If a time series has a unit root, it is said to be
integrated.

First differencing the time series removes the unit root. E.g.
in the case of a random walk
yt = yt-1 + ut, ut ~ N(0, σ2)
Δyt = ut


the first difference is white noise, which is stationary.
For an AR(p) a unit root implies
1 – β1L – β2L2 – ... – βpLp = (1 – L) (1 – λ1L – λ2L2 ... – λpLp-1) = 0

and as a result first differencing also removes the
unit root.
Non-stationarity

Nonstationarity can have important
consequences for regression modelsand
inference.



Autoregressive coefficients are biased
t-statistics have non-normal distributions even
in large samples
Spurious regression
Problem: Spurious Regression

imagine we now have two series are
generated by independent random walks,

Suppose we run yt on xt using OLS, that is
we estimate yt = α + βxt + νt.

In this case, you tend to see ”significant”
β because the low-frequency changes
make it seem as if the two series are in
some way associated.
Unit Root Tests

Standard Dickey-Fuller test appropriate
for AR(1) processes



Many economic and financial time series have
a more complicated dynamic structure than is
captured by a simple AR(1) model.
Said and Dickey (1984) augment the basic
autoregressive unit root test to accommodate
general ARMA(p, q) models with unknown
orders and
Called the augmented Dickey-Fuller (ADF) test
ADF Test – 1
The ADF test tests the null hypothesis that
a time series yt is I(1) against the
alternative that it is I(0), assuming that
the dynamics in the data have an ARMA
structure.
 The ADF test is based on estimating the
test regression

p
yt  xt  yt 1   j yt  j   t
j 1
Deterministic
variables
Potential unit root
Other serial
correlation
ADF Test - 2

To see why:
yt   xt  1 yt 1   2 yt  2  ...   p yt  p   t
 xt  (1   2  ..   p ) yt 1   2 ( yt  2  yt 1 )  ...
  p ( yt  p  ...  yt 1 )   t

Subtract yt-1 from both sides and define
Φ = (α1+ α2+…+ αp – 1) and we get
p
yt  xt  yt 1    j yt  j   t
j 1

Test Φ= 0 against alternative Φ<0


Use special DF upperbound and lowerbound
Under alternative, test statistic is not normally distributed
Estimating in Time Series


Non-stationary time series can lead to a lot of
problems in econometric analysis.
In order to work with time series, particular in
regression models, we should therefore transform
our variables to stationary time series first.




First differencing removes unit roots or trends. Hence,
difference a time series until it is I(0).
Differencing too often is less of a problem since a
differenced stationary series is still stationary.
Regressions of one stationary variable on another is less
problematic.
Although observations may not be independent, we
can expect regression to have similar properties as
with cross sectional data.
Impulse Response Function

One of the most interesting things to do with an
ARMA model is form predictions of the variable
given its past.



we want to know what is Et(xt+j )
Can do inference with Vart(xt+j)
The impulse response function is a simpel way to
do that



Follow te path that x follows if it is kicked by unit shock
characterization of the behavior of our models.
allows us to start thinking about “causes” and “effects”.
Impulse Response and MA(∞)

1. The MA(∞) representation is the same thing as
the impulse response function.

i.e.
yt   j  t  j
j 0

The easiest way to calculate an MA(∞)
representation is to simulate the impulseresponse function.

The impulse response function is the same as
Et(xt+j) − Et−1(xt+j).
Causality and Impulse Response

Can either forecast or simulate the effect of a given shock

Try to pick a shock time/level to simulate and try to replicate
observed data


Know a shock happened in time time t


Issue of whether that shock is what really happened
See if observed change (more on this next time)
Granger causality implies a correlation between the current
value of one variable and the past values of others



it does not necessarily imply that changes in one variable
“causes” changes in another.
Use a F-test to jointly test for the significance of the lags on
the explanatory variables, this in effect tests for ‘Granger
causality’ between these variables.
Can visually see correlation in impulse response functions
Source: Cochrane, QJE (1994)
Next Time

Estimating Causality in Time Series



Some additional forecasting stuff
Testing for breaks
Regression discontinuity/Event study
Download