Short and Long Run Causality Measures: Theory and Inference

advertisement
Short and Long Run Causality Measures: Theory
and Inference
Jean-Marie Dufour y
Université de Montréal
Abderrahim Taamoutiz
Université de Montréal
First version: October 20, 2004
This version : May 18, 2006
Preliminary
The authors thank Kilian Lutz, Benoit Perron, Nadège-Désirée Yameogo, the participants of 45th Annual Meeting of the Société canadienne de science économique at
Charlevoix (2005), Joint Statistical Meetings of the American Statistical Association at
Minneapolis (2005), 22nd CESG Conference at Vancouver (2005), and CIREQ Time Series
Conference at Montréal (2005) for their helpful comments.
y
Canada Research Chair Holder (Econometrics). Centre interuniversitaire de recherche
en économie quantitative (CIREQ), Centre interuniversitaire de recherche en analyse des
organisations (CIRANO), and Département de sciences économiques, Université de Montréal. Mailing address: Département de sciences économiques, Université de Montréal,
C.P. 6128 succursale Centre-ville, Montréal, Québec, Canada H3C 3J7. TEL: 1 514
343 2400; FAX: 1 514 343 5831; e-mail: jean.marie.dufour@umontreal.ca. Web page:
http://www.fas.umontreal.ca/SCECO/Dufour .
z
Centre interuniversitaire de recherche en économie quantitative (CIREQ), Centre interuniversitaire de recherche en analyse des organisations (CIRANO), et Département de sciences économiques, Université de Montréal. Adresse: Département de sciences
économiques, Université de Montréal, C.P. 6128 succursale Centre-ville, Montréal, Québec,
Canada H3C 3J7. E-mail: abderrahim.taamouti@umontreal.ca.
i
ABSTRACT
The concept of causality introduced by Wiener (1956) [32] and Granger
(1969) [15] is de…ned in terms of predictability one period ahead. Recently, Dufour and Renault (1998) [9] generalized this concept by considering
causality at a given (arbitrary) horizon h, and causality up to any given horizon h, where h is a positive integer and can be in…nite (1 h 1). This
generalization is motivated by the fact that, in the presence of an auxiliary
variable Z, it is possible to have the situation where the variable Y does not
cause variable X at horizon 1, but causes it at horizon h > 1. In this case,
this is an indirect causality transmitted by the auxiliary variable Z.
Another related problem consists in measuring the importance of causality between two variables. Existing causality measures have been established
only for the horizon 1 and fail to capture indirect causal e¤ects. This paper
proposes a generalization of such measures for any horizon h. We propose
nonparametric and parametric measures for feedback and instantaneous effects at any horizon h. Parametric measures are de…ned in terms of impulse
response coe¢ cients of vector moving average representation (V M A). By
analogy with Geweke (1982) [11], we show that it is always possible to de…ne a measure of dependence at horizon h which can be decomposed into
a sum of feedback measures from X to Y; from Y to X; and instantaneous
e¤ect at horizon h. We propose a new approach to evaluate these measures
based on the simulation of a large sample from the process of interest. We
also propose a valid nonparametric con…dence intervals, using the bootstrap
technique. Finally, we suggest an empirical application which illustrates the
evolution in time of the degree of causality between macroeconomic and
…nancial variables.
Keywords: Causality; auxiliary variable; indirect e¤ect; short-run causality
measure; long-run causality measure; instantaneous causality; dependence;
predictability; estimation by simulation; valid bootstrap con…dence intervals.
ii
Contents
1 Introduction
1
2 Motivation
3
3 Framework
5
4 Nonparametric causality measures
7
5 Parametric causality measures
12
5.1 Parametric measures in the context of V ARM A(p; q) process 12
5.2 Characterization of causality measures in the context of V M A(q)
process . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
6 Estimation
23
7 Evaluation by simulation of causality measures
29
8 Con…dence intervals
33
9 Empirical illustration
40
10 Conclusion
42
i
1
Introduction
In this paper we propose measures of causality at di¤erent horizons which
can help to detect the well known indirect causality that appears only at
higher horizons. We also propose a valid nonparametric con…dence intervals
for these measures, using the bootstrap technique.
Our causality measures can be applied in di¤erent contexts and may help
to solve some puzzles of the economic and …nancial literature. For example
they can improve the well known debate on long-term predictability of stock
returns.
In the present paper we applied the causality measures at di¤erent
horizons to illustrate the evolution in time of the degree of causality between macroeconomic and …nancial variables. Taamouti, Garcia, and Dufour (2006), …rst, use these measures to quantify and compare the strength
of dynamic leverage and volatility feedback e¤ects in high-frequency equity
data. Second, they measure the dynamic impact of bad and good news on
volatility. Finally, using simulations they assess the ability of these measures
to detect the symmetric property of some parametric volatility models as
GARCH and nonlinear GARCH models.
Recall that the concept of causality introduced by Wiener (1956) [32] and
Granger (1969) [15] is now a basic notion for studying dynamic relationships
between time series. This concept is de…ned in terms of predictability at
horizon 1 of a variable X from its own past, the past of another variable
Y; and possibly a vector Z of auxiliary variables. Following Granger (1969)
[15], we de…ne temporal causality from X to Y one period ahead as follows:
Y causes X in the sense of Granger if observations of Y up to time (t 1)
can help to predict X(t) given the past of X and Z up to time (t 1). More
precisely, we say that Y causes X in the sense of Granger if the variance of
the forecast error of X obtained by using the past of Y is smaller than the
variance of the forecast error of X obtained without using the past of Y .
The theory of causality has generated a considerable literature. In the
context of the bivariate ARM A models, Kang (1981) [20], derived necessary and su¢ cient conditions for noncausality. Boudjellaba, Dufour, and
Roy (1992, 1994) ([3], [4]) developed necessary and su¢ cient conditions of
noncausality for multivariate ARM A models. Parallel to the literature on
the noncausality conditions, some authors developed tests for the presence of
causality between time series. The …rst test is due to Sims (1972) [30] in the
context of bivariate time series. Other tests were developed for V AR models
[ see Pierce and Haugh (1977) [27], Newbold (1982) [25], Geweke (1984) [12]
] and V ARM A models [ see Boudjellaba, Dufour, and Roy (1992, 1994) ([3],
1
[4])].
Recently, Dufour and Renault (1998) [9] generalized the concept of causality in the sense of Granger (1969) [15] by considering causality at a given
(arbitrary) horizon h and causality up to horizon h, where h is a positive
integer and can be in…nite (1 h 1) [ see also Sims (1980) [31], Hsiao
(1982) [18], and Lütkepohl (1993) [24] ]. Such generalization is motivated
by the fact that, in the presence of an auxiliary variable Z, it is possible to
have the variable Y not causing variable X at horizon 1, but causing it at
a horizon h > 1. In this case, this is an indirect causality transmitted by
the auxiliary variable Z. Dufour and Renault (1998) [9] introduced necessary and su¢ cient conditions of noncausality between vectors of variables at
any horizon h for stationary and nonstationary processes with …nite second
moments.
The analysis of Wiener-Granger distinguishes among three types of causality: two unidirectional causalities (called feedbacks) from X to Y; from Y
to X; and an instantaneous causality due to the e¤ect of the present value
of variables on the other. In practice, it is possible that these three types of
causality coexist, hence the importance of …nding means to measure their
degree and to determine the most important one. Unfortunately, existing
tests of causality fail to accomplish this task, they only inform us about
the presence or not of causality. Geweke (1982) [11] extended the causality
concept by de…ning measures of the degrees of feedback and instantaneous
e¤ects, that can be decomposed in time and frequency …eld. Gouriéroux,
Monfort and Renault (1987) [14] proposed causality measures based on the
Kullback information. Polasek (1994) [28] showed how causality measures
can be calculated using Akaike Information Criterion (AIC). Polasek (2000)
[29] introduced new causality measures in the context of univariate and multivariate ARCH models and their extensions based on a Bayesian approach.
Existing causality measures have been established only for a one period
horizon and fail to capture indirect causal e¤ects. This paper proposes a
generalization of such measures for any horizon h. We propose nonparametric and parametric measures for feedback and instantaneous e¤ects at any
horizon h. Parametric measures are de…ned in terms of impulse response
coe¢ cients of V M A representation. By analogy with Geweke (1982, 1984)
([11], [13]), we show that it is always possible to de…ne a measure of dependence at horizon h which can be decomposed into a sum of feedback
measures from X to Y; from Y to X 1 ; and instantaneous e¤ect at horizon
1
See de…nition of the feedback from Y to X in the beginning of this introduction.
Similar de…nition for the feedback from X to Y can be obtained.
2
h. We propose a new approach to evaluate these measures based on the
simulation of a large sample from the process of interest. We also propose
a valid nonparametric con…dence intervals, using the bootstrap technique.
The plan of the paper is as follows. Section 2 provides the motivation behind an extension of causality measures at horizon h > 1. Section 3 presents
the framework allowing the de…nition of causality at di¤erent horizons. In
section 4 we propose nonparametric short run and long run causality measures. In section 5 we give a parametric equivalent, in the context of linear
stationary invertible processes, of the causality measures suggested in section 4. Thereafter, we characterize our measures in the context of …nite
moving average models V M A(q) with …nite order q. In section 6 we discuss
di¤erent estimation approaches. In section 7 we suggest a new approach to
estimate these measures based on the simulation. In section 8 we establish
the asymptotic distribution of measures and the asymptotic validity of their
nonparametric bootstrap con…dence intervals. Section 9 is devoted to an
empirical application and the conclusion relating to the results is given in
section 10.
2
Motivation
The causality measures we propose in this paper are an extension of those
developed by Geweke (1982,1984) ([11], [13]) and others [see introduction].
The existing causality measures quantify the e¤ect of a vector of variables
on another at one period horizon. The contribution of such measures is
particularly limited in the presence of auxiliary variables, since it is possible
that a vector Y causes another vector X at horizon h strictly higher than
1 even if there is no causality at horizon 1. In this case we speak about
an indirect e¤ect induced by the auxiliary variables Z: Clearly causality
measures de…ned for horizon 1 are unable to quantify this indirect e¤ect.
This paper aims …rst to propose causality measures at di¤erent horizons to
quantify the degree of short and long run causality between a vectors of
random variables. Second we propose causality measures that detect and
quantify the indirect e¤ects due to the presence of an auxiliary variable:
To illustrate the importance of such causality measures we consider the
following examples.
Example 1 We suppose that the information set contains only two vari0
ables X and Y . We consider a bivariate stationary process (X; Y ) which
follows a V AR(1) model:
3
X(t + 1)
Y (t + 1)
=
0:5
0:4
0:7
0:35
X(t)
Y (t)
+
"X (t + 1)
"Y (t + 1)
.
(1)
X(t + 1) is given by the following equation:
X(t + 1) = 0:5 X(t) + 0:7 Y (t) + "X (t + 1):
(2)
Since the coe¢ cient of Y (t) in equation (2) is equal to 0:7, we can conclude
that Y causes X in the sense of Granger. However, this does not give
any information about the existence or not of causality for horizons strictly
higher than 1 nor on its degree (weak or strong). To study the causality at
horizon 2 [ see Dufour and Renault (1998) [9] ] let us consider the system
(1) at time (t + 2), we have
X(t + 2)
Y (t + 2)
0:53 0:595
0:34 0:402
0:5 0:7
+
0:4 0:35
"X (t + 2)
+
.
"Y (t + 2)
=
X(t)
Y (t)
"X (t + 1)
"Y (t + 1)
(3)
X(t + 2) is given by:
X(t + 2) =
0:53 X(t) + 0:595Y (t)
(4)
+0:5"X (t + 1) + 0:7"Y (t + 1)
+"X (t + 2):
The coe¢ cient of Y (t) in equation (4) is equal to 0:595, so we can conclude
that Y causes X at horizon 2. But, how can we measure the importance of
this “long run” causality? Existing measures cannot answer this question.
Example 2 We suppose now that the information set contains not only
the two variables of interest X and Y but also an auxiliary variable Z: We
0
consider a trivariate stationary process (X; Y; Z) which follows a V AR(1)
model:
0
1 2
X(t + 1)
@ Y (t + 1) A=4
Z(t + 1)
0:60 0
0
0:40
0
0:60
0
"X (t + 1)
+@ "Y (t + 1)
"Z (t + 1)
4
0:80
0
0:10
1
A.
30
1
X(t)
5@ Y (t) A
Z(t)
(5)
X(t + 1) is given by:
X(t + 1) = 0:6 X(t) + 0:8 Z(t) + "X (t + 1):
(6)
Since the coe¢ cient of Y (t) in equation (6) is 0, we can conclude that Y
does not cause X at horizon 1. If we consider the model (5) at time (t + 2),
we get:
0
1 2
X(t + 2)
0:60 0:00 0:80
@ Y (t + 2) A=4 0:00 0:40 0:00
Z(t + 2)
2 0:00 0:60 0:10
0:60 0:00 0:80
+4 0:00 0:40 0:00
00:00 0:60 10:10
"X (t + 2)
+@ "Y (t + 2) A.
"Z (t + 2)
32 0
1
X(t)
5 @ Y (t) A
Z(t) 1
30
"X (t + 1)
5@ "Y (t + 1) A
"Z (t + 1)
(7)
X(t + 2) is given by:
X(t + 2) =
0:36 X(t) + 0:48Y (t)
(8)
+0:56 Z(t) + 0:6"X (t + 1)
+0:8"Z (t + 1) + "X (t + 2):
The coe¢ cient of Y (t) in equation (8) is equal to 0:48, which implies that
Y causes X at horizon 2: This shows that the absence of causality at h = 1
does not exclude the possibility of a causality at horizon h > 1. This indirect
e¤ect is transmitted by the variable Z:
0:6
0:8
Y ! Z ! X;
0:48 = 0:60 0:80, where 0:60 and 0:80 are the coe¢ cients of the one period
e¤ect of Y on Z and the one period e¤ect of Z on X; respectively: So,
how can we measure the importance of this indirect e¤ ect? Again, existing
measures cannot answer this question.
3
Framework
The notion of noncausality considered here is de…ned in terms of orthogonality conditions between subspaces of a Hilbert space of random variables
with …nite second moments. We denote L2 L2 ( ; A; Q) the Hilbert space
of real random variables de…ned on a common probability space ( ; A; Q);
with covariance as inner product.
5
We consider three multivariate stochastic processes fX(t) : t 2 Zg ; fY (t) : t 2 Zg ;
and fZ(t) : t 2 Zg, with
0
X(t) = (x1 (t); :::; xm1 (t)) ;
0
Y (t) = (y1 (t); :::; ym2 (t)) ;
xi (t) 2 L2
(i = 1; :::; m2 );
2
(i = 1; :::; m3 );
yi (t) 2 L
0
Z(t) = (z1 (t); :::; zm3 (t)) ;
(i = 1; :::; m1 );
2
zi (t) 2 L
where m1
1; m2
1; m3
0; and m1 + m2 + m3 = m: We denote
Xt = fX (s); s
tg, Yt = fY (s), s
tg; and Zt = fZ(s), s
tg the
¯
¯
¯
information sets which contain all the past and present values of X, Y; and
Z; respectively: We denote It the information set which contains Xt , Yt ; and
¯ ¯
Zt . It At , for At =Xt ;Yt ; Zt ; contains all the elements of It except those of
¯
¯ ¯ ¯
At . These information sets can be used to predict the value of X at horizon
h, denoted X(t + h), for all h 1.
For any information set Bt , we denote by P (xi (t + h) j Bt ) the best
linear forecast of xi (t + h) based on the information set Bt ; u(xi (t + h) j
Bt ) = xi (t + h) P (xi (t + h) j Bt ) the corresponding prediction error,
and 2 (xi (t + h) j Bt ) the variance of this prediction error: In other words,
the best linear forecast of X(t + h) is denoted by P (X(t + h) j Bt ) =
0
(P (x1 (t + h) j Bt ); :::; P (xm1 (t + h) j Bt )) : U (X(t + h) j Bt ) = (u(x1 (t + h) j
0
Bt ) ; :::; u(xm1 (t + h) j Bt )) is the corresponding vector of prediction error,
and (X(t+h) j Bt ) its covariance matrix: Each component P (xi (t+h) j Bt )
of P (X(t + h) j Bt ); for 1
i
m1 ; is then the orthogonal projection of
xi (t + h) on the subspace Bt :
Dufour and Renault (1998) [9] provide the following de…nition of noncausality at horizon h and up to horizon h, for h 1.
De…nition 1 (Dufour and Renault (1998) [9]): For h
1; where h
positive integer,
(i) Y does not cause X at horizon h given It Yt , denoted Y 9 X j It
¯
h
if:
P (X(t + h) j It Yt ) = P (X(t + h) j It ) 8t > w;
¯
where w represents a “starting point”2 ;
(ii) Y does not cause X up to horizon h given It Yt , denoted Y 9
¯
(h)
It Yt ; if Y 9 X j It Yt for k = 1; 2; :::; h;
¯
¯
k
2
is a
Y;
¯t
X j
The “starting point”w is not speci…ed. In particular w may equal 1 or 0 depending
on whether we consider a stationary process on the integers (t 2 Z) or a process fX(t) :
t 1g on the positive integers given initial values preceding date 1.
6
(iii) Y does not cause X at any horizon given It
It
Y ; if Y 9 X j It
¯t
k
Yt ; denoted Y 9 X j
¯
(1)
Yt for all k = 1; 2; :::
¯
This de…nition corresponds to an unidirectional causality from Y to X.
It means that Y causes X at horizon h if to predict X(t + h); the past of Y
improves the forecast based on the information set It Yt . For more details,
¯
the reader can refer to Dufour and Renault (1998) [9].
An alternative de…nition can be expressed in terms of the covariance
matrix of the forecast errors.
De…nition 2 For h 1; where h is a positive integer,
(i) Y does not cause X at horizon h given It Yt , if:
¯
det (X(t + h) j It Yt ) = det (X(t + h) j It ) 8t > w;
¯
where det (X(t + h) j At ); At = It ; It Yt ; represents the determinant of
¯
the covariance matrix of the forecast error of X(t + h) given At :
(ii) Y does not cause X up to horizon h given It Yt ; if:
¯
(X(t + k) j It Yt ) = (X(t + k) j It )
¯
8 t > w and k = 1; 2; :::; h;
(iii) Y does not cause X at any horizon given It
Yt ; if:
¯
(X(t + k) j It Yt ) = (X(t + k) j It )
¯
8 t > w and k = 1; 2:::
4
Nonparametric causality measures
In this section and for the remainder of this paper we consider an information
set It which contains two vector valued random variables of interest X and
Y; and an auxiliary valued random variable Z: In other words, we suppose
that It = H[Xt [Yt [Zt , where H represents a subspace of the Hilbert space,
¯ ¯ ¯
possibly empty, containing time independent variables (e.g. the constant in
a regression model).
The causality measure we adopt in this paper is an extension of the
measure adopted by Geweke (1982, 1984) ([11], [13]). The motivations remain the same as those enumerated in Geweke (1982) [11], among which
the most important are: 1) it is nonnegative and 2) it is 0 only when there
is no causality. Speci…cally, we propose the following causality measures at
horizon h 1.
7
De…nition 3 For h 1; where h is a positive integer, a causality measure
from Y to X at horizon h, called the INTENSITY of the causality from Y
to X at horizon h; denoted C(Y !XjZ ); is given by:
h
C(Y !XjZ ) = ln
h
det (X(t + h) j It Yt )
¯
;
det (X(t + h) j It )
Similarly we can de…ne a causality measure from X to Y at horizon h,
called the INTENSITY of the causality from X to Y at horizon h; denoted
C(X!Y jZ ); by:
h
C(X!Y jZ ) = ln
h
det (Y (t + h) j It Xt )
¯
:
det (Y (t + h) j It )
Remark 1 For m1 = m2 = m3 = 1,
2
C(Y !XjZ ) = ln
h
and
(X(t + h) j It Yt )
¯
;
2 (X(t + h) j I )
t
2
C(X!Y jZ ) = ln
h
(Y (t + h) j It Xt )
¯
:
2 (Y (t + h) j I )
t
Then C(Y !XjZ ) quantify the degree of the causal e¤ect from Y to X at
h
horizon h given the past of X and Z: In terms of predictability, it measures
the quantity of the information brought in the past of Y that can improve the
forecast of X(t + h): Following Geweke (1982) [11], this measure can be also
interpreted in terms of proportion of reduction in the variance of forecast
error of X(t + h) by taking into account the past of Y . This proportion is
equal to:
2
100 [
2
(X(t + h) j It Yt )
(X(t + h) j It )
]%=100 [1-exp(-C(Y !XjZ ))]%.
¯
2 (X(t + h) j I
h
Yt )
t
¯
We can rewrite the conditional casualty measures given by de…nition 3
in terms of unconditional casualty measures [see Geweke (1984) [13]]:
C(Y !XjZ ) = C(Y Z!X )
h
h
where
8
C(Z!X );
h
det (X(t + h) j It Yt Zt )
;
¯
¯
det (X(t + h) j It )
h
det (X(t + h) j It Yt Zt )
C(Z!X ) = ln
:
¯
¯
det (X(t + h) j It Yt )
h
¯
C(Y Z!X ) = ln
C(Y Z!X ) and C(Z!X ) represent the unconditional causality measures from
0
h
0 0
h
(Y ; Z ) to X and from Z to X; respectively. Similarly we have
C(X!Y jZ ) = C(XZ!Y )
h
h
C(Z!Y );
h
where,
det (Y (t + h) j It Xt Zt )
,
¯
¯
det (Y (t + h) j It )
det (Y (t + h) j It Xt Zt )
:
C(Z!Y ) = ln
¯
¯
det (Y (t + h) j It Xt )
h
¯
CXZ!Y
= ln
h
We de…ne an instantaneous causality measure between X and Y at horizon h as follows.
De…nition 4 For h
1; where h is a positive integer, an instantaneous
causality measure between Y and X at horizon h, called the INTENSITY of
the instantaneous causality between Y and X at horizon h; denoted C(X$Y jZ );
h
is given by:
C(X$Y jZ ) = ln
h
det (X(t + h) j It ) det (Y (t + h) j It )
:
det (X(t + h); Y (t + h) j It )
where det (X(t+h); Y (t+h) j It ) represents the determinant of the covari0
ance matrix of the forecast error of the joint process X ; Y
given the information set It .
0
0
at horizon h
Remark 2 If we take m1 = m2 = m3 = 1; then
det (X(t + h); Y (t + h) j It )
= 2 (X(t + h) j It ) 2 (Y (t + h) j It )
(cov(X(t + h); Y (t + h) j It ))2 :
So the instantaneous causality measure between X and Y at horizon h can
be written as follows,
9
(cov(X(t + h); Y (t + h) j It ))2
)
2 (X(t + h) j I ) 2 (Y (t + h) j I )
t
t
1
;
2 (X(t + h); Y (t + h) j I )
t
C(X$Y jZ ) = ln (1
h
= ln
1
1
where (X(t+h); Y (t+h) j It ) represents the correlation coe¢ cient between
X(t + h) and Y (t + h) conditional on the information set It . This instantaneous causality measure is higher when this coe¢ cient becomes higher.
We can also de…ne a measure of dependence between X and Y at horizon
h. This will enable us to check if, at horizon h; the processes X and Y must
be considered together or if they should be treated independently given the
information set It Yt .
¯
De…nition 5 For h
1; where h is a positive integer, a measure of dependence between X and Y at horizon h, called the INTENSITY of the
dependence between X and Y at horizon h, denoted C (h) (X; Y j Z), is given
by:
C(h) (X,YjZ)=ln
:
det (X(t + h) j It Yt ) det (Y (t + h) j It
¯
det (X(t + h); Y (t + h) j It )
Xt )
¯
We can easily show that the measure of dependence at horizon h is equal
to the sum of feedbacks measures from X to Y , from Y to X, and of the
instantaneous causality measure at horizon h. We have
C (h) (X; Y j Z) = C(X
!Y jZ )
h
+ C(Y
!XjZ )
h
+ C(X$Y jZ ):
(9)
h
It is possible to build a recursive formulation of causality measures. This
one will depend on the predictability measure introduced by Diebold and
Kilian (1998) [5].
Diebold and Kilian (1998) [5] propose a predictability measure based on
the ratio of expected losses of short and long run forecasts:
P (L;
t ; j; k)
=1
E(L(et+j;t ))
;
E(L(et+k;t ))
where t is an available information set at time t, L is a loss function, j
and k represent respectively the short and the long-run, et+s;t = X(t + s)
P (X(t + s) j t ), s = j; k; is the forecast error at horizon (t + s):
10
This predictability measure can be constructed according to the horizons
of interest and it allows for general loss functions and for univariate or
multivariate information sets. In this paper we suppose that the loss function
L is a quadratic function,
L(et+s;t ) = e2t+s;t ; for s = j; k;
and that t is a multivariate information set. We have the following recursive
formulation.
Proposition 1 Let h1 , h2 be two di¤ erent horizons. For h2 > h1
C(Y
!XjZ )-C(Y !XjZ )
h1
= ln [1
h2
PX (It
Y ; h1 ; h2 )]
¯t
ln [1
1,
PX (It ; h1 ; h2 )];
(10)
where PX ( : ; h1 ; h2 ) represents the predictability measure for variable X,
PX (It
det (X(t + h1 ) j It
det (X(t + h2 ) j It
Y ; h1 ; h2 ) = 1
¯t
det (X(t + h1 ) j It )
:
det (X(t + h2 ) j It )
PX (It ; h1 ; h2 ) = 1
Similarly, we have:
C(X
!Y jZ )-C(X !Y jZ )
h1
h2
= ln [1
Y)
¯t ;
Y)
¯t
PY (It
X ; h1 ; h2 )]
¯t
ln [1
PY (It ; h1 ; h2 )];
(11)
where PY ( : ; h1 ; h2 ) represents the predictability measure for variable Y ,
PY (It
det (Y (t + h1 ) j It
det (Y (t + h2 ) j It
X ; h1 ; h2 ) = 1
¯t
det (Y (t + h1 ) j It )
:
det (Y (t + h2 ) j It )
PY (It ; h1 ; h2 ) = 1
Corollary 1 For h
C(Y
X)
¯t ;
X)
¯t
2;
!XjZ )
h
= C(Y
!XjZ )
+ ln[1
1
ln[1
PX (It
PX (It ; 1; h)]
Y ; 1; h)]:
¯t
Similarly, we have
C(X
!Y jZ )
h
= C(X
!Y jZ )
+ ln[1
1
ln[1
PY (It
PY (It ; 1; h)]
X ; 1; h)]:
¯t
This corollary follows immediately from proposition 1 above.
11
For h2
h1 , the function Pk ( : ; h1 ; h2 ), k = X; Y; represents the measure of short-run forecast relative to the long-run forecast, and C(k !ljZ )
h1
C(k
!ljZ ); for l 6= k and l; k = X; Y; represents the di¤erence between the
h2
degree of short run causality and that of long run causality. By Diebold
and Kilian (1998) [5], Pk ( : ; h1 ; h2 )
0 means that the series is highly
predictable at horizon h1 relative to h2 ; and Pk ( : ; h1 ; h2 ) = 0; means that
the series is nearly unpredictable at horizon h1 relative to h2 .
In particular cases, the interpretation of (10) is easy. For example, consider the case where PX (It ; h1 ; h2 ) PX (It Yt ; h1 ; h2 ), which means that
¯
the knowledge of the past of Y can help to improve the forecast of X at
horizon h1 relative to horizon h2 . Then the unidirectional e¤ect from Y
to X is stronger at horizon h1 than at horizon h2 . In other words, if h1
represents the short-run and h2 represents the long-run, then the causality
from Y to X is stronger in the short-run than in the long-run. Moreover, if
PX (It ; h1 ; h2 ) = PX (It Yt ; h1 ; h2 ), then we have a constant unidirectional
¯
e¤ect at horizon h1 relative to h2 .
5
Parametric causality measures
Now we consider a more speci…c set of linear invertible processes which
includes as special cases V AR, V M A; and V ARM A models with …nite
order. Under this set we show that it is possible to obtain a parametric
expression for short run and long run causality measures in terms of impulse
response coe¢ cients of V M A representation.
This section is divided into two subsections. In the …rst one we calculate parametric measures of short run and long run causality in the context of autoregressive moving average model. We suppose that the process
0
0
0
0
fW (s) = (X (s); Y (s); Z (s)) ; s tg (hereafter unconstrained model) is a
V ARM A(p; q) model, where p and q can be in…nite. The process fS(s) =
0
0
0
(X (s); Z (s)) ; s
tg (hereafter constrained model) can be deduced from
the unconstrained model using corollary 6.1.1 of Lütkepohl (1993) [24]. This
model follows a V ARM A(p; q) model with p mp and q (m 1)p + q.
In the second subsection we provide a characterization of those parametric
measures in the context of V M A(q) model, where q is …nite.
5.1
Parametric measures in the context of V ARM A(p; q) process
Without loss of generality, let us consider the discrete vectorial process of
0
0
0
0
zero mean fW (s) = (X (s); Y (s); Z (s)) ; s tg de…ned on L2 and charac12
terized by the following autoregressive moving average representation:
W (t) =
p
X
j W (t
j=1
=
p
X
j=1
+
2
4
2
j) +
'j u(t
j) + u(t)
j=1
(12)
30
XXj
XY j
XZj
Y Xj
YYj
Y Zj
ZXj
ZY j
ZZj
'XY j
'Y Y j
'ZY j
'XZj
'Y Zj
'ZZj
'XXj
4 'Y Xj
j=1
'ZXj
q
X
q
X
1
X(t j)
5 @ Y (t j) A
Z(t j)
30
1 0
1
uX (t j)
uX (t)
5 @ uY (t j) A +@ uY (t) A,
uZ (t j)
uZ (t)
with
E [u(t)] = 0;
Or compactly
h
i
0
E u(t)u (s) =
u
0
for s = t
:
for s 6= t
(L)W (t) = '(L)u(t);
where
2
XX (L)
XY (L)
XZ (L)
Y X (L)
Y Y (L)
Y Z (L)
ZX (L)
ZY (L)
ZZ (L)
'XX (L)
'(L)=4 'Y X (L)
'ZX (L)
'XY (L)
'Y Y (L)
'ZY (L)
Pp
'XZ (L)
'Y Z (L)
'ZZ (L)
(L)=4
2
ii (L)
= Imi
ik (L)
=
j=1
Pp
'ik (L) =
j
ikj L ;
j=1
'ii (L) = Imi +
Pq
j
iij L ,
Pq
j
j=1 'iij L ,
j
j=1 'ikj L ;
for i 6= k, and i; k = X; Y; Z:
13
3
5,
3
5,
We suppose that u(t) is orthogonal to the subspace of the Hilbert space
fW (s); s
(t 1)g and that u is a symmetric positive de…nite matrix.
Under stationarity, the V M A(1) representation of model (12) is given by:
W (t) =
(L)u(t);
(13)
where
(L) =
(L)
1
'(L) =
1
X
j
jL
=
1
X
j=0
j=0
2
4
XXj
XY j
XZj
Y Xj
YYj
Y Zj
ZXj
ZY j
ZZj
3
5Lj ;
and
0
= Im :
From the previous section, measures of dependence and feedbacks are de…ned in terms of covariance matrices of the constrained and unconstrained
forecast errors3 . So to calculate these measures we need to know the structure of the constrained model(imposing noncausality). This can be deduced
from the structure of the unconstrained model (12) using the following
proposition and corollary.
Proposition 2 (Lütkepohl (1993, pages 231-232) [24]): Linear Transformation of a V M A(q) process): Let u(t) be a K-dimensional white noise
process with nonsingular covariance matrix u and let
W (t) =
+
q
X
j u(t
j) + u(t);
j=1
be a K-dimensional invertible V M A(q) process. Furthermore let F be an
(M K) matrix of rank M: Then the M -dimensional process S(t) = F W (t)
has an invertible V M A(q) representation:
S(t) = F +
q
X
j "(t
j) + "(t);
j=1
where "(t) is M -dimensional white noise with nonsingular covariance matrix
M ) coe¢ cient matrices and q q:
" ; the j ; j = 1; :::; q; are (M
3
Here we mean the forecast errors of the constrained and unconstrained models.
14
Corollary 2 (Lütkepohl (1993, pages 232-233) [24]): Linear Transformation of a V ARM A(p; q) process): Let W (t) be a K-dimensional, stable, invertible V ARM A(p; q) process and let F be an (M K) matrix of rank M:
Then the process S(t) = F W (t) has a V ARM A(p; q) representation with
p
q
Kp;
(K
1)p + q:
Remark 3 If we suppose that W (t) follows a V AR(p) = V ARM A(p; 0)
model, then the linear transformation S(t) = F W (t) has a V ARM A(p; q)
representation with p Kp and q (K 1)p:
Suppose that we are interested in measuring the causality from Y to X
at given horizon h. So we need to apply corollary (2) to de…ne the structure
0
0
of process fS(s) = (X (s); Z (s))0 ; s tg. If we left-multiply equation (13)
by the adjoint matrix of the matrix (L), denoted (L) , then we get
(L)
(L)W (t) =
(L) '(L)u(t);
(14)
where (L) (L) = det [ (L)]. Since the determinant of (L) is a sum of
products involving one operator from each row and each column of (L),
the degree of the AR polynomial, here det [ (L)] ; is at most mp. We write
det [ (L)] = 1
1L
p
mp:
:::
p
pL ;
where
It is easy to check that the degree of the operator (L) '(L) is at most
p(m 1) + q. So equation (14) can be written as follows:
det [ (L)] W (t) =
(L) '(L)u(t):
(15)
This equation is another stationary invertible V ARM A representation of
process W (t); called the …nal equation form.
0
0
The model of process fS(s) = (X (s); Z (s))0 ; s
tg can be obtained
by taking F equal to:
F =
Im1
0
0
0
15
0
Im3
:
Pre-multiplying (15) with F results in
det [ (L)] S(t) = F (L) '(L)u(t):
(16)
The right-hand side of equation (16) is a linearly transformed …nite order
V M A process which, by proposition 2, has a V M A(q) representation with
q p(m 1) + q . We have the following constrained model,
det [ (L)] S(t) = (L)"(t) =
XX (L)
XZ (L)
ZX (L)
ZZ (L)
"(t)
(17)
with
E ["(t)] = 0;
where
h
i
0
E "(t)" (s) =
for s = t
for s 6= t;
"
0
ii (L)
= Imi +
ik (L)
=
Pq
Pq
j
j=1 ii;j L ;
j
j=1 ikj L ;
for i 6= k and i; k = X; Z;
Note that, in theory the coe¢ cients ikj , i; k = X; Z; j = 1; ::::q; and elements of the covariance matrix " can be computed from coe¢ cients ikj ,
'ikl ; i; k = X; Z; Y ; j = 1; :::; p; l = 1; :::; q; and elements of the covariance
matrix u . This is possible by solving the following system:
" (v)
=
u (v),
v = 0; 1; 2; :::;
(18)
where " (v) and u (v) are the autocovariance functions of the processes
(L)"(t) and F (L) '(L)u(t); respectively. For large numbers m, p; and q;
the system (18) can be solved by using optimization methods. In the next
two sections we will discuss other approaches to estimate the constrained
and unconstrained models.
The following example shows how one can calculate the theoretical parameters of the constrained model in terms of those of unconstrained model
in the context of a bivariate V AR(1) model.
16
Example 3 We consider the following bivariate V AR(1) model:
X(t)
Y (t)
=
XX
XY
YX
YY
X(t
Y (t
=
X(t
Y (t
1)
1)
1)
1)
uX (t)
uY (t)
+
(19)
+u(t)
In a reduced form
X(t)
Y (t)
(L)
where
1
(L)=
= u(t);
XX L
XY
Y XL
1
L
YY
;
L
with
E [u(t)] = 0; 8
h
0
E u(t)u(s)
i
<
=:
u
2
uX
=
cov(uX (t); uY (t))
2
uY
cov(uY (t); uX (t))
0
for s = t;
for s 6= t:
We suppose that all roots of det( (z)) = det(I2 z) are outside of the unit
circle. Under this assumption the model (19) has the following V M A(1)
representation:
X(t)
Y (t)
=
1
X
j=0
=
1
X
uX (t
uY (t
j
j=0
XX;j
XY;j
Y X;j
Y Y;j
j)
j)
uX (t
uY (t
j)
j)
;
where
j
=
0
= I2 :
j 1
=
j
; j = 1; 2; :::;
If we suppose that we are interested in determining the model of marginal
process X(t), then by corollary (2) and for F = [1; 0] we have,
det( (L))X(t) = [1; 0] (L) u(t);
where,
(L) =
1
YYL
Y XL
XY L
1
17
XX L
;
and
det( (L)) = 1
(
YY
+
XX )L
(
2
XX Y Y )L :
Y X XY
Thus,
X(t) ( Y Y + XX )X(t 1) ( Y X XY
= XY uY (t 1)
1)+ uX (t):
Y Y uX (t
XX Y Y )X(t
2)
The right-hand side, denoted $(t); is the sum of an M A(1) process and a
white noise process. By proposition 2, $(t) has an M A(1) representation,
$(t) = "X (t) + "X (t 1). To calculate parameters and V ar("X (t)) = 2"X
in terms of parameters of the unconstrained model, we have to solve system
(18) for v = 0 and v = 1;
8
V ar ["X (t) + "X (t 1)] = V ar [uX (t)
1) + XY uY (t
>
Y Y uX (t
>
<
>
>
:
E [("X (t) + "X (t 1))("X (t
+ XY uY (t 1))(uX (t 1)
()
8
<
:
(1 +
2
2
"X
=
)
2
"X
= (1 +
YY
1) + "X (t 2))] = E[(uX (t)
2) + XY uY (t 2))];
Y Y uX (t
Y Y uX (t
2 ) 2
uX
YY
+
2
XY
2
uY
2
XY
uY uX
and
2 .
"X
YY
2 :
uX
Here we have two equations and two unknown parameters
parameters must satisfy constraints j j< 1 and 2"X > 0:
These
The V M A(1) representation of model (17) is given by:
1
S(t) = det [ (L)]
(L)"(t) =
1
X
j "(t
j)
(20)
j=0
=
1
X
j=0
XXj
XZj
ZXj
ZZj
where
0
= I(m1 +m2 ) :
18
"X (t
"Z (t
j)
j)
;
1)] ;
1)
To quantify the degree of causality from Y to X at horizon h; we …rst need
to consider the unconstrained and constrained models of process X. The
unconstrained model is given by the following equation:
X(t) =
1
X
XXj uX (t
j=1
1
X
+
1
X
j) +
XY j uY (t
j)
j=1
XZj uZ (t
j) + uX (t);
j=1
and the constrained model is given by:
X(t) =
1
X
XXj "X (t
j) +
j=1
1
X
XZj "Z (t
j) + "X (t):
j=1
Second, we calculate the covariance matrices of the unconstrained and constrained forecast errors of X(t + h). From equation (13) the forecast error
of W (t + h) is given by:
enc [W (t + h) j It ] =
h 1
X
i u(t
+h
i);
i=0
associated to the covariance matrix
(W (t + h)
j
=
It ) =
h 1
X
i
V ar [u(t)]
0
(21)
i
i=0
h 1
X
i
u
0
i:
i=0
So, the unconstrained forecast error of X(t + h) is given by:
enc (X(t + h) j It ) =
+
h 1
X
h 1
X
XXj uX (t
+h
j) +
j=1
XZj uZ (t
h 1
X
XY j uY (t
j=1
+h
j) + uX (t + h)
j=1
associated to the unconstrained covariance matrix
(X(t + h) j It ) =
h 1
X
i=0
19
[enc
X
i
u
0
nc0
i eX ];
+h
j)
where
enc
X =
Im1
0
0
:
Similarly the forecast error of S(t + h) is given by:
ec [S(t + h) j It
Yt ] =
¯
associated to the covariance matrix
(S(t + h) j It
h 1
X
i "(t
+h
i);
i=0
Yt ) =
¯
h 1
X
i
"
0
i:
i=0
Consequently, the constrained forecast error of X(t + h) is given by:
ec (X(t+h) j It Yt ) =
¯
h 1
X
XXj
j=1
h 1
X
"X (t+h j)+
"Z (t+h j)+"X (t+h)
XZj
j=1
associated to the constrained covariance matrix
(X(t + h) j It
Yt ) =
¯
h 1
X
Im1
0
ecX
i
"
0
i
0
ecX ;
i=0
where
ecX =
:
Finally, we have the following result.
Theorem 1 Under assumptions 12, 13; and for h 1; where h is a positive
integer,
#
"
Ph 1 c
0
0
det( i=0
eX i " i ecX )
C(Y !XjZ ) = ln
;
P
0 nc0
h
det( hi=01 enc
e
)
u
i
i
X
X
where
enc
X =
Im1
0
0
;
ecX =
Im1
0
:
Example 4 If we consider a bivariate V AR(1) process [see example (3) ],
then one can calculate the causality measure at any horizon h using only the
unconstrained parameters. For example, the causality measure at horizons
1 and 2 are given by:
20
(1 +
C(Y
!X )=ln[
2
YY
)
2
uX
2
XY
+
2
uY
+
1
C(Y
ln[
!X )=
2
4
2
YY
2[(1+
4
2
uX +[(1+ Y Y
)
2
2
2
XX ) uX + XY
2
2
uX + XY
2
uY
][(1+
2
uY
2
YY
)
q
q
((1+
((1 +
YY
2
2
uX + XY
2
2
YY
)
2
uX
2
XY
+
2
uY
4
2
YY
X
2
YY
4
uX
2
2
XY
2
uY
)2 4
YY
2
2
uX ]
2
YY
4
uX ]
The last two formulas are obtained under assumptions cov(uX (t); uY (t)) = 0
and ((1 + 2Y Y ) 2uX + 2XY 2uY )2 4 2Y Y 4uX 0:
Now we will calculate the parametric measure of instantaneous causality at given horizon h. We know from section 4 that this measure is de…ned only in terms of covariance matrices of unconstrained forecast errors.
The covariance matrix of the unconstrained forecast error of joint process
0
0
X (t + h); Y (t + h)
0
is given by:
(X(t + h); Y (t + h) j It ) =
h 1
X
G
i
u
0
iG
i=0
where
G=
Im1
0
0
Im2
0
0
:
We also have,
(X(t + h) j It ) =
h 1
X
(Y (t + h) j It ) =
h 1
X
and
where
enc
Y =
0
[enc
X
i
u
[enc
Y
i
u
0
nc0
i eX ];
i=0
0
nc0
i eY ]:
i=0
Im2
Thus, we have the following result.
21
0
4
uX
2
uX
2
2
2
2 4
u + XY uY )
qX
((1+ Y Y )2 2u +
)2
2 )2
uY
:
0
;
]
]
Theorem 2 Under assumptions 12, 13; and for h
integer,
C(X
!Y jZ )=ln
h
"
Ph 1
det( i=0 [enc
X
where
nc0
i eX ]) det(
Ph 1
det( i=0 [G i u
Im1
0
G=
Ph
0
u
i
0
Im2
eX =
Im1
eY =
0
1; where h is a positive
0
0
1 nc
i=0 [eY
0
iG
0
i
u
0
nc0
i eY ])
])
#
;
;
0
0
;
Im2
0
:
Remark that the parametric measure of dependence at horizon h can be
deduced from its decomposition given by equation (9).
5.2
Characterization of causality measures in the context of
V M A(q) process
0
0
0
0
Now suppose that the process fW (s) = (X (s); Z (s); Y (s)) ; s
an invertible V M A(q) model:
W (t) =
q
X
j=1
=
q
X
j=1
j u(t
2
4
0
with
In a reduced form:
j) + u(t)
(22)
XXj
XY j
XZj
Y Xj
YYj
Y Zj
ZXj
ZY j
ZZj
uX (t)
1
30
uX (t
5 @ uY (t
uZ (t
+@ uY (t) A
uZ (t)
tg follows
E [u(t)] = 0;
h
i
0
E u(t)u (s) =
u
0
for s = t;
:
for s 6= t
W (t) = (L)u(t);
where
22
1
j)
j) A
j)
2
and
(L)=4
XX (L)
(L)
Y Y (L)
ZY (L)
XZ (L)
XY
Y X (L)
ZX (L)
ii (L)
= Imi +
ik (L)
=
Pq
Pq
ZZ (L)
j
ikj L ;
i; k = X; Z; Y:
From proposition (2) and by taking F=
Im1
0
the constrained process S(t) is an M A(q) with q
S(t) =
(L)"(t) =
5;
j
iij L ,
j=1
j=1
for i 6= k,
Y Z (L)
3
q
X
j "(t
0
0
0
Im2
, the model of
q. We have,
j)
j=0
=
q
X
j=0
XX;j
XZ;j
ZX;j
ZZ;j
"X (t
"Z (t
j)
j)
;
with
E ["(t)] = 0;
h
i
0
E "(t)" (s) =
"
0
for s = t
:
for s 6= t
Using proposition 1 we have the following result.
Theorem 3 Let h1 and h2 two di¤ erent horizons. Under assumption (22)
we have
C(Y !X=Z ) = C(Y !X=Z ); 8 h2 h1 q
h1
h2
This result follows immediately from proposition ?? and by the fact that
for all h2 h1 q; PX (It Yt ; h1 ; h2 ) = PX (It ; h1 ; h2 ):
¯
6
Estimation
From section 5 we have that short run and long run causality measures
depend on parameters of the model describing the process of interest. So
23
the estimation of these measures can be done by replacing the unknown
parameters by their estimates from a …nite sample.
More generally, there are three di¤erent approaches in estimating our
causality measures based on the di¤erent estimation approaches of the model
of the process of interest. The …rst one, called nonparametric approach,
which is the focus of this section supposes that the form of the parametric
model appropriate for the process of interest is unknown and it approximates
this one with a V AR(k) model, where k depends on the sample size [ see
Parzen (1974) [26], Bhansali (1978) [2], Lewis and Reinsel (1985) [23]]. The
second approach supposes that the process follows a …nite order V ARM A
model. The standard methods for the estimation of V ARM A models, such
as maximum likelihood and nonlinear least squares, require nonlinear optimization. This might not be feasible because the number of parameters
can increase quickly. To circumvent this problem several authors [ see Hannan and Rissanen (1982) [17], Hannan and Kavalieris (1984b) [16], Koreisha
and Pukkila (1989) [22], Dufour and Pelletier (2003) [8], and Dufour and
Jouini (2004) [7]] have developed a relatively simple approach based only
on linear regressions. This approach enables estimation of V ARM A models
using a long V AR whose order depends on the sample size. The last and
simplest approach assumes that the process follows a …nite order V AR(p)
model which can be estimated by OLS.
In practice, the precise form of the parametric model appropriate for
a process is unknown. Parzen (1974) [26], Bhansali (1978) [2], Lewis and
Reinsel (1985) [23], and others, have considered a nonparametric approach
of predicting future values using an autoregressive model …tted to a series
of T observations. This approach is based on a very mild assumption of an
in…nite order autoregressive model for the process which includes …nite-order
stationary V ARM A processes as a special case.
In this section, we use the nonparametric approach to estimate the short
and long-run causality measures. First, we discuss the estimation of the
…tted autoregressive constrained and unconstrained models. Next, we point
out some assumptions necessary for the convergence of the estimated parameters. Second, using theorem 6 of Lewis and Reinsel (1985) [23], we de…ne
approximations of covariance matrices of the constrained and unconstrained
forecast errors at horizon h. Last, we use these approximations to construct
an asymptotic estimator of our short and long-run causality measures.
In what follows we focus on the estimation of the unconstrained model.
0
0
0 0
Let us consider a stationary vector process fW (s) = (X(s) ; Y (s) ; Z(s) ) ; s
t)g. Based on Wold’s theorem this process can be written in the form of a
V M A(1) model,
24
W (t) = u(t) +
1
X
'j u(t
j);
j=1
E [u(t)] = 0;
i
0
u for s = t;
E u(t)u (s) =
0
for s 6= t
h
;
P
where u is a positive de…nite matrix. We assume that 1
j=0 k 'j k< 1 and
P
0
j
detf'(z)g 6= 0 forj z j 1, where k 'j k= tr('j 'j ) and '(z) = 1
j=0 'j z ;
with '0 = Im ; an (m m) identity matrix: Under these assumptions W (t)
is invertible and can be written as an in…nite autoregressive process,
W (t) =
1
X
j W (t
j) + u(t);
(23)
j=1
P1
P1
j
1 satis…es
where
j=1 k j k< 1 and (z) = Im
j=1 j z = '(z)
detf (z)g =
6 0 for j z j 1:
Let (k) = ( 1 ; 2 ; :::; k ) denote the …rst k autoregressive coe¢ cients
in the V AR(1) representation. Given a realization fW (1); :::; W (T )g; we
can approximate (23) by a …nite order V AR(k) model, where k depends on
sample size T . The OLS estimator of the autoregressive coe¢ cients of the
…tted V AR(k) model is given by the following equation:
^ (k) = (^ 1k ; ^ 2k ; :::; ^ kk ) = ^ 0 ^ 1 ;
k1 k
PT
0
0
0 0
1
^
where k = (T k)
t=k+1 w(t)w(t) , for w(t) = (W (t) ; :::; W (t k+1) ) ;
P
0
and ^ k1 = (T k) 1 Tt=k+1 w(t)W (t + 1) . An estimator of the covariance
matrix of the error terms is given by:
^k =
u
T
X
0
u
^k (t)^
uk (t) =(T
k);
t=k+1
Pk
j):
where u
^k (t) = W (t)
j=1 ^ jk W (t
Theorem 1 of Lewis and Reinsel (1985) [23] ensures the convergence
of ^ 1k ; ^ 2k ; :::; ^ kk under three assumptions: 1) E j ui (t)uj (t)uk (t)ul (t) j
i; j; k; l
m; 2) k is chosen as a function of T such that
4 < 1; for 1
2 =T ! 0 as k; T ! 1; and 3) k is chosen as a function of T such that k 1=2
k
P1
j=k+1 k j k! 1 as k; T ! 1. Theorem 4 of the same authors derives
the asymptotic distribution for these estimators under 3 assumptions: 1) E j
25
ui (t)uj (t)uk (t)ul (t) j
i; j; k; l m; 2) k is chosen as a function
4 < 1; 1
of T such that k 3 =T ! 0 as k; T ! 1; and 3) it exist fl(k)g a sequence of
0
(km2 1) vectors such that 0 < M1 k l(k) k2 = l(k) l(k) M2 < 1; for
k = 1; 2; ::: We also note that ^ ku converges to u , as k and T ! 1 [ see
Lütkepohl (1993, pages 308-309) [24] ].
Remark 4 The upper bound K of the order k in …tted V AR(k) model depends on the assumptions required to ensure the convergence and the asymptotic distribution of the estimator. For the convergence of the estimator we
2
need to assume that kT ! 0, as k and T ! 1; consequently we can choose
p
as an upper bound K = C T ; where C is a constant: For the asymptotic dis3
tribution of the estimator we need to assume that kT ! 0, as k and T ! 1,
p
and thus we can choose as an upper bound K = C 3 T .
Now we will calculate an asymptotic approximation of the covariance
matrix of the forecast error at horizon h. The forecast error of W (t + h)
based on the V AR(1) model is given by:
enc [W (t + h) j W (t); W (t
1); :::] =
h 1
X
'j u(t + h
j);
j=0
associated to the covariance matrix
[W (t + h) j W (t); W (t
1); :::] =
h 1
X
'j
0
u 'j :
j=0
In the same way, the covariance matrix of the forecast error of W (t + h)
based on V AR(k) model, is given by:
k [W (t
+ h)
j
W (t); W (t
= E[(W (t + h)
1); :::; W (t
k
X
k + 1)]
(h)
^ jk W (t + 1
j))
j=1
(W (t + h)
k
X
j=1
where
26
(h)
^ jk W (t + 1
0
j) ]:
(h+1)
(h+1)
^ jk
(h)
= ^ (j+1)k + ^ 1k ^ jk ; for j 2 f1; ::; g, h
(1)
^ jk
1;
= ^ jk ;
(0)
= Im :
^ jk
[see Dufour and Renault (1998) [9]]. Moreover, let us note that:
k
X
W (t + h)
(h)
^ jk W (t + 1
j=1
1
X
= (W (t + h)
(h)
j W (t
j)
+1
(24)
j))
j=1
k
X
(h)
(
^ jk W (t + 1
1
X
j)
j=1
h 1
X
=
'j u(t + h
(h)
j W (t
+1
j))
j=1
k
X
(h)
(
^ jk W (t + 1
j)
j=0
1
X
j)
j=1
(h)
j W (t
+1
j)):
j=1
where
(h+1)
j
(1)
j
(0)
j
=
(h+1)
j+1
=
j;
+
(h)
j;
1
for j 2 f1; ::; g, h
1;
= Im :
Since the errors terms u(t + h j), for 0 j (h 1), are independent of
0
0
(W (t); W (t 1); :::; ) and (^ 1k ; ^ 2k ; :::; ^ kk ), the two terms on the right-hand
side of the equation (24) are independent. Thus,
k [W (t
+ h)
j
=
W (t); W (t
1); :::; W (t
k + 1)]
[W (t + h) j W (t); W (t
1); :::; ]
k
X
(h)
+E[(
^ jk W (t + 1
j)
j=1
k
X
(
(h)
^ jk W (t + 1
j=1
1
X
(h)
j W (t
+1
j))
j=1
j)
1
X
j=1
27
(25)
(h)
j W (t
+1
0
j)) ]:
An asymptotic approximation, as k and T ! 1, of the last term in the
equation (25) is given by theorem 6 of Lewis and Reinsel (1985) [23]:
k
X
(h)
E[(
^ jk W (t + 1
1
X
j)
j=1
k
X
(h)
(
^ jk W (t + 1
+1
j))
j=1
j)
j=1
1
X
(h)
j W (t
0
+1
j)) ]
j=1
h 1
km X
'j
T
(h)
j W (t
0
u 'j :
j=0
Consequently, an asymptotic approximation of the covariance matrix of the
forecast error is equal to:
k [W (t
+ h)
j
W (t); W (t
(1 +
km
)
T
1); :::; W (t
h 1
X
'j
k + 1)]
(26)
0
u 'j :
j=0
An estimator of this quantity is obtained by replacing the parameters 'j
and u by their estimators '
^ kj and ^ ku ; respectively.
We can also obtain an asymptotic approximation of the covariance matrix of the constrained forecast error at horizon h following the same steps
as before. We denote this covariance matrix by:
k [S(t
+ h)
j
S(t); S(t
1); :::; S(t
k + 1)]
h 1
(1 +
k(m1 + m3 ) X
)
T
j
"
0
j;
j=0
where j ; for j = 1; ::; (h 1), represents the coe¢ cient of V M A representation of the process constrained S and " is the covariance matrix of
0
0 0
"(t) = ("X (t) ; "Z (t) ) : Based on these results an asymptotic approximation
of the causality measure from Y to X is given by:
a
C (Y
"
P
det[ hj=01 ecX j
P
!X=Z ) = ln
h
det[ hj=01 enc
X 'j
28
c0
j eX ]
0 nc0
u 'j eX ]
"
0
#
+ ln 1
km2
;
T + km
where,
enc
X =
Im1
0
0
;
ecX =
Im1
0
:
An estimator of this quantity will be obtained by replacing the unknown
parameters by their estimators,
C^ a (Y
7
"
P
0 #
0
det[ hj=01 ecX ^ kj ^ k" ^ kj ecX ]
+ ln 1
P
!X=Z ) = ln
0
0
h
det[ hj=01 enc
^ kj ^ ku '
^ kj enc
X'
X ]
km2
:
T + km
Evaluation by simulation of causality measures
In this section we propose a simple simulation-based technique to calculate our causality measures at any horizon h, for h 1. To illustrate this
technique we consider the same examples of section 1 limiting ourselves to
horizons 1 and 2.
Since one source of bias in the autoregressive coe¢ cients is …nite sample
size, our technique consists of simulating a large sample from the unconstrained model whose parameters are supposed to be known or estimated
from a real data set. Once the large sample (hereafter large simulation) is
simulated we use it to estimate the parameters of the constrained model.
In what follows we describe an algorithm to calculate the causality measure at given horizon h using large simulation technique:
(1) Given parameters of the unconstrained model and its initial values,
simulate a large sample of T observations, where T can be equal to 1000000,
under assumption that probability distribution of disturbance is completely
speci…ed4 .
(2) Estimate the constrained model using large simulation.
(3) Calculate the constrained and unconstrained covariance matrices of
the forecast errors at horizon h:
(4) Calculate the causality measure at horizon h using constrained and
unconstrained covariance matrices of step (3):
4
The form of the probability distribution of disturbances does not a¤ect the value of
causality measures.
29
Now let reconsider example one of section one:
X(t + 1)
Y (t + 1)
=
=
X(t)
Y (t)
0:5
0:4
+ u(t):
0:7
0:35
(27)
X(t)
Y (t)
uX (t + 1)
uY (t + 1)
+
;
where
E[u(t)] = 0;
E[u(t)u(s)0 ] =
I2 if s = t
:
0 if s 6= t
(28)
Our illustration is done in two steps. First, we calculate the theoretical
values of causality measures at horizons 1 and 2. In case of bivariate V AR(1)
model it is easy to calculate the causality measure at any horizon h using
only the unconstrained parameters [see example (4)]. This is not the case for
models with more than one lag and two variables [see section (5)]. Second,
we evaluate our causality measures using large simulation technique.
The theoretical values of causality measures at horizons 1 and 2 are
recovered as follows:
(1) We calculate variances of the forecast errors of X at horizons 1 and
2 using its own past and past of Y . Note that,
0
[(X(t + h); Y (t + h)) j Xt ; Yt ] =
¯ ¯
From this covariance matrix we have,
h 1
X
i
i0
:
i=0
V ar[X(t + 1) j Xt ; Yt ] = 1;
¯ ¯
and
V ar[X(t + 2) j Xt ; Yt ] =
¯ ¯
0
1
X
e
i
i0
0
e = 1:74;
i=0
where e = (1; 0) :
(2 ) We calculate variances of the forecast errors of X at horizons 1 and
2 using only its own past. In this case we need to determine the structure
30
of the constrained model. This one is given by the following equation [see
example 3]:
X(t + 1) = (
YY
+
XX )X(t)
+(
Y X XY
XX Y Y )X(t
1)
+"X (t + 1) + "X (t):
where Y Y + XX = 0:85 and Y X XY
XX Y Y = 0:105: The parameters
and V ar("X (t)) = 2"X are the solutions to the following system:
8
(1 + 2 ) 2"X = 1:6125
<
:
:
2 = 0:35
"X
The set of possible solutions is f( ; 2"X ) = ( 4:378; 0:08); ( 0:2285; 1:53)g.
To get an invertible solution we must choose the combination which satis…es
the condition j
j< 1; i.e. the combination ( 0:2285; 1:53): Thus, the
variance of the forecast error of X at horizon 1 using only its own past is
given by:
[X(t + 1) j Xt ] = 1:53;
¯
and the variance of the forecast error of X at horizon 2 is:
[X(t + 2) j Xt ] = 2:12:
¯
Thus, we have,
C(Y
!X )
1
= 0:425; C(Y
!X )
= 0:197:
2
In a second step we use the algorithm described in the beginning of this
section to evaluate the causality measures using large simulation technique.
Hereafter, we give the results5 that we obtain for di¤erent lag orders p of
the constrained model.
5
For illustration we take T = 600; 000:
31
Table 1: Evaluation by simulation of C(Y
p
1
2
3
4
5
10
15
20
25
30
35
C(Y
!X )
1
0:519
0:430
0:427
0:425
0:426
0:425
0:426
0:425
0:425
0:426
0:425
C(Y
!X )
1
and C(Y
!X )
2
!X )
2
0:567
0:220
0:200
0:199
0:198
0:197
0:199
0:197
0:199
0:198
0:198
The above results show that large simulation technique works well, and this
can be explained by law of large numbers.
Now consider example two of section one:
0
1 2
X(t + 1)
0:60
@ Y (t + 1) A = 4 0:00
Z(t + 1)
0:00
0:00
0:40
0:60
0:80
0:00
0:10
30
1 0
1
X(t)
"X (t + 1)
5 @ Y (t) A + @ "Y (t + 1) A ;
Z(t)
"Z (t + 1)
(29)
where
E[u(t)] = 0;
I3 if s = t
E[u(t)u(s)0 ] =
:
0 if s 6= t
In this example the analytical calculations of the causality measures at horizons 1 and 2 is not easy. The interest of this example is to illustrate the
indirect causality:
As we have shown in section one, in this example the variable Y does
not cause the variable X at horizon 1, but causes it at horizon 2. So, we
expect that causality measure from Y to X will be equal to zero at horizon
1 and di¤erent from zero at horizon 2. Using large simulation technique for
32
a sample size T = 600; 000 and for di¤erent lag orders p of the constrained
model we get the following results:
Table 2: Evaluation by simulation of C(Y
p
C(Y
1
2
3
4
5
10
15
20
25
30
35
!XjZ )
!X )
C(Y
1
0:000
0:000
0:000
0:000
0:000
0:000
0:000
0:000
0:000
0:000
0:000
and C(Y
1
!X )
(Indirect causality)
2
!XjZ )
2
0:121
0:123
0:122
0:123
0:124
0:122
0:122
0:122
0:124
0:122
0:122
Thus, the above results con…rm our expectations about the presence of an
indirect causality from Y to X.
8
Con…dence intervals
We suppose that process of interest W
t)g6 follows a V AR(p) model,
W (t) =
p
X
j W (t
0
fW (s) = (X(s); Y (s); Z(s)) ; s
j) + u(t);
(30)
j=1
or equivalently,
(I3
p
X
j
j L )W (t)
= u(t);
(31)
j=1
Pp
j
where polynomial (z) = I3
j=1 j z satis…es det[ (z)] 6= 0 for z 2 C
1
with j z j 1; and fu(t)gt=0 is a sequence of i.i.d. random variables with
following properties:
6
Here we suppose that X, Y; and Z are univariate variables. It will be easy to generalize
results of this section to the multivariate case.
33
and
E [u(t)] = 0;
h
i
0
E u(t)u (s) =
u
u
0
if s = t
;
if s 6= t
is a positive de…nite matrix.
Remark 5 If we suppose that process W follows a V AR(1) model, then
we can use Inoue and Kilian’s (2002) [19] approach to get similar results to
those developed in this section.
The parameters of model (30) can be estimated by OLS. For a realization W
fW (1); :::; W (T )g of process (30), an estimate of = ( 1 ; :::; p )
is given by the following equation:
^ = ^0 ^
1
An estimator of
u
1
:
(32)
is given by:
^u =
T
X
0
u
^(t)^
u(t) =(T
p);
t=p+1
P
0
0
0 0
where ^ = (T p) 1 Tt=p+1 w(t)w(t) , for w(t) = (W (t) ; :::; W (t p + 1) ) ;
Pp
0
^ 1 = (T p) 1 PT
^(t) = W (t)
j):
t=p+1 w(t)W (t + 1) ; and u
j=1 ^ j W (t
Now suppose we are interested in measuring causality from Y to X at
horizon h: In this case we need to know the structure of the marginal process
0
fS(s) = (X(s); Z(s)) ; s t)g: This one has a V ARM A(p; q) representation
with p 3p and q 2p,
S(t) =
p
X
j S(t
j) +
j=1
q
X
i "(t
i) + "(t);
i=1
where f"(t)g1
t=0 is a sequence of i.i.d. random variables that satis…es
E ["(t)] = 0;
h
i
0
E "(t)" (s) =
"
0
if s = t
;
if s 6= t
and " is a positive de…nite matrix.
The equation (33) can be written in the following reduced form,
(L)S(t) = (L)"(t);
34
(33)
where (L) = I2
:::
Lp and (L) = I2 + 1 L + ::: + q Lq : We
1L
Pq p j
suppose that (z) = I2 + j=1 j z satis…es det[ (z)] 6= 0 for z 2 C and
j z j 1: Under this assumption the V ARM A(p; q) process is invertible and
has a following pure V AR(1) representation,
1
X
S(t)
c
j S(t
j) = (L)
1
(L)S(t) = "(t):
(34)
j=1
Let c = ( c1 ; c2 ; :::) denote the matrix of all autoregressive coe¢ cients
in model (34) and c (k) = ( c1 ; c2 ; :::; ck ) denote its …rst k autoregressive
coe¢ cients. Suppose that we approximate (34) by a …nite order V AR(k)
model, where k depends on sample size T . In this case an estimator of the
autoregressive coe¢ cients of V AR(k) model is given by:
^ c (k) = (^ c ; ^ c ; :::; ^ c ) = ^ 0 ^ 1 :
k1 k
1k
2k
kk
An estimator of
"
is given by:
^ "k =
T
X
0
^"k (t)^"k (t) =(T
k);
t=k+1
P
0
0
0
where ^ k = (T k) 1 Tt=k+1 Sk (t)Sk (t) , for Sk (t) = (S (t); :::; S (t k +
Pk
P
0
0
c
1)) ; ^ k1 = (T k) 1 Tt=k+1 Sk (t)S(t+1) ; and ^"k (t) = S(t)
j=1 ^ jk S(t
j):
Based on the above notations, the theoretical value of the causality measure from Y to X at horizon h may de…ned as follows:
C(Y
!XjZ ) = ln (
h
G(vec( c ); vech(
H(vec( ); vech(
" ))
u ))
);
where
G(vec(
c
); vech(
" ))
=
h 1
X
0
ec
j=0
0
ec = (1; 1) ;
and
35
c(j)
1
c(j)
" 1 ec ;
H(vec( ); vech(
u ))
h 1
X
=
0
(j)
1
e
(j)
u 1 e;
j=0
0
e = (1; 1; 1) ;
c(j)
c(j 1)
c(j 1)
c(0)
c(1)
c ; for j
with 1 = 2
+ 1
2, 1 = I2 ; and 1 = c1 [see
1
Dufour and Renault (1998) [9]]. vec denotes the column stacking operator
and vech is the column stacking operator that stacks the elements on and
below the diagonal only:
From corollary 6.1.1 of Lütkepohl (1993, page 232) [24] it exists a function f (:) : R9(p+1) ! R4(k+1) which associates the constrained parameters (vec( c ); vech( " )) to the unconstrained parameters (vec( ); vech( u ))
such that (vec( c ); vech( " )) = f (vec( ); vech( u )) [see example (4)].
C(Y
!XjZ )
= ln(
h
G(f (vec( ); vech( u )))
):
H(vec( ); vech( u ))
An estimate of the quantity C(Y
^ Y
C(
!XjZ )
= ln(
h
!XjZ )
is given by:
h
G(vec( ^ c (k)); vech( ^ ";k ))
);
H(vec( ^ ); vech( ^ u ))
(35)
where G(vec( ^ c (k)); vech( ^ ";k )) and H(vec( ^ ); vech( ^ u )) are estimates of
the population quantities.
Now, let consider the following assumptions.
Assumption 1 :
[see Lewis and Reinsel (1985)]
1) E j "h (t)"i (t)"j (t)"l (t) j
h; i; j; l 2;
4 < 1; for 1
2) k is chosen as a function of T such that k 3 =T !
1;
P01as k; T !
1=2
c
3) k is chosen as a function of T such that T
j=k+1 k j k! 0 as
k; T ! 1:
4) Series used to estimate parameters of V AR(k) and series used for
prediction are generated from two independent processes which have the
same stochastic structure.
^ Y
Proposition 3 (Consistency of C(
is a consistent estimator of C(Y
!XjZ ))
h
!XjZ ):
h
36
^ Y
Under assumption 1; C(
!XjZ )
h
^ Y
To establish the asymptotic distribution of C(
!XjZ ),
let start by re-
h
calling the following result [see Lütkepohl (1990, page 118-119) [24] and
Kilian (1998a, page 221) [21]]:
vec( ^ )
vech( ^ u )
p)1=2
(T
vec( )
vech( u )
d
! N (0; );
(36)
where
1
=
0
0
2(D3 D3 )
u
0
1 D0 (
u
3
0
;
1
u )D3 (D3 D3 )
D3 is the duplication matrix, de…ned such that vech(F ) = D3 vech(F ) for
any symmetric 3 3 matrix F .
Assumption 2 :
f (:) is continuous and di¤erentiable function.
According to (36) and using delta method we have following result [see
proof in appendix].
^ Y
Proposition 4 (Asymptotic distribution of C(
!XjZ ))
tions 1 and 2 we have,
(T
^ Y
p)1=2 (C(
!XjZ )
C(Y
h
!XjZ ))
h
Under assump-
h
d
! N (0;
C );
where
@C(Y
C
=(
@C(Y
!XjZ )
h
0
@(vec( ) ; vech(
0
u)
)
) (
!XjZ )
0
h
0
@(vec( ) ; vech(
0
u)
)
);
and
1
=
0
u
0
0
2(D3 D3 )
1 D0 (
u
3
0
u )D3 (D3 D3 )
1
:
See proof in appendix.
The analytical derivative of the causality measure with respect to the
0
vector (vec( ); vech( u )) is not feasible. One way to build con…dence intervals for our causality measures is to use large simulation technique [see
37
section 7] to calculate the derivative numerically. Another way is by building
bootstrap con…dence intervals.
As mentioned by Inoue and Kilian (2002) [19], for the bounded measures
[ as in our case C(Y !XjZ ) 2 [0; 1)], the bootstrap approach is more relih
able then delta-method. The raison is because “the delta-method interval
is not range respecting and may produce con…dence intervals that are logically invalid. In contrast, the bootstrap percentile interval by construction
preserves these constraints”[see Inoue and Kilian (2002, pages 315-318) [19]
and Efron and Tibshirani (1993) [10]].
In the rest of this section we show how one can construct a valid con…dence interval for causality measures using bootstrap technique. In what
follows, we establish the validity of the percentile bootstrap con…dence intervals.
Let consider the following bootstrap approximation to the distribution
of causality measure at given horizon h.
(1) Estimate by OLS a V AR(p) process and save the residuals
u
~(t) = W (t)
p
X
^ j W (t
j); for t = p + 1; :::; T;
j=1
where ^ j , for j = 1; :::; p, are given by equation (32).
(2) Generate (T p) bootstrap residuals u
~ (t) by random sampling with
replacement from the residuals u
~(t); t = p + 1; :::; T:
(3) Generate a random draw for the vector of p initial observations
0
0
0
w(0) = (W (1); :::; W (p)) :
(4) Given ^ [see equation (32) ]; u
~ (t); and w(0); generate bootstrap
data for the dependent variable W (t) from equation:
W (t) =
p
X
^ j W (t
j) + u
~ (t); for t = p + 1; :::; T:
j=1
(5) Calculate the bootstrap OLS regression estimates
^ = (^ 1 ; ^ 2 ; :::; ^ p ) = ^
38
1^
1;
and
T
X
^ =
u
0
u
~ (t)~
u (t) =(T
p);
t=p+1
P
0
0
0
where ^ = (T p) 1 Tt=p+1 w (t)w (t) , for w (t) = (W (t); :::; W (t
P
0
0
p
1)) ; ^ 1 = (T
p) 1 Tt=p+1 w (t)W (t + 1) ; and u
~ (t) = W (t)
P+
p
W
(t
j):
^
j=1 j
(6) Estimate the constrained model of the marginal process (X; Z) using
the bootstrap sample fW (t)gTt=1 :
(7) Calculate the causality measure at horizon h; denoted C^ (j) (Y
!XjZ ),
h
using equation (35).
(8) Choose B such
times.
1
2
(B + 1) is an integer and repeat steps (2)
(7) B
Conditionally on the sample we have,
p)1=2
(T
vec( ^ )
vech( ^ u )
vec( ^ )
vech( ^ u )
d
! N (0; );
(37)
where
1
=
u
0
0
0
2(D3 D3 )
1 D0 (
u
3
0
u )D3 (D3 D3 )
;
1
D3 is the duplication matrix de…ned such that vech(F ) = D3 vech(F ) for
any symmetric 3 3 matrix F .
Proposition 5 (Asymptotic validity of the residual-based bootstrap)
Under assumptions 1 and 2 we have,
p)1=2 (C^ (Y
(T
!XjZ )
h
^ Y
C(
!XjZ ))
h
d
! N (0;
C );
where
@C(Y
C
=(
!XjZ )
@C(Y
h
@(vec( )0 ; vech(
0
u) )
) (
!XjZ )
h
@(vec( )0 ; vech(
0
u) )
0
);
and
1
=
0
u
0
0
2(D3 D3 )
1 D0 (
u
3
39
0
u )D3 (D3 D3 )
1
:
Remark 6 Proposition (5) may be used to justify the application of biascorrected bootstrap con…dence intervals, proposed by Kilian (1998,a) [21], to
improve the performance of the small sample percentile bootstrap intervals
of our causality measures.
Kilian (1998,a) [21] proposes an algorithm to remove the bias in impulse
response function prior to bootstrapping the estimate. As he mentioned, the
small sample bias in impulse response function may arise from bias in slope
coe¢ cient estimates or from the nonlinearity of this function, and this can
translate into changes in interval width and location. However, if we suppose
that ordinary least-squares (OLS) small-sample bias can be responsible for
bias in the estimated impulse response function, replacing the biased slope
coe¢ cient estimates by bias-corrected slope coe¢ cient estimates may help
to reduce the bias in the impulse response function.
Kilian (1998,a) shows that the additional modi…cations proposed in the
bias-corrected bootstrap con…dence intervals method to improve the performance of the percentile bootstrap intervals in small samples do not alter
its asymptotic validity. The reason is that the e¤ect of bias corrections is
negligible asymptotically. For more details on the bias-corrected bootstrap
con…dence intervals method, the reader can refer to Kilian (1998,a) [21].
9
Empirical illustration
In this section we apply our causality measures to quantify the degree of
relationships between macroeconomic and …nancial variables. The data set
considered is the one used by Bernanke and Mihov (1998) [1] and Dufour,
Pelletier and Renault (2003) [6]. This data set consists of monthly observations on nonborrowed reserves (N BR), federal funds rate (r), gross domestic
product de‡ator (P ), and real gross domestic product (GDP ). The monthly
data on GDP and GDP de‡ator were constructed by state space methods
from quarterly observations7 . Our sample goes from January 1965 to December 1996 for a total of 384 observations.
All variables are in logarithmic form [see Figures 1-4 in appendix]. These
variables were also transformed by taking …rst di¤erences [see Figures 5-8
in appendix], so the causality relations will have to be interpreted in term
of the growth rate of variables.
We have performed an Augmented Dickey-Fuller test (hereafter ADF test) for nonstationnarity of the four variables of interest and their …rst
7
For more details, the reader can consults Bernanke and Mihov (1998).
40
di¤erences. The values of the test statistics as well as the critical values
corresponding to 5% signi…cance level are given in tables 4 and 5 [see appendix].
The following table 3 summarizes the results of the stationarity tests for
all variables.
Table 3: Stationarity test results
N BR
r
P
GDP
Variables in logarithmic form
No
No
No
No
First di¤erence
Y es
Y es
No
Y es
As we can read from table 3, all variables in logarithmic form are nonstationary. However, their …rst di¤erences are stationary except for the GDP
de‡ator, P: We also did a nonstationary test for the second di¤erence of variable P: The test statistic values are equal to 11:04826 and 11:07160 for
an ADF -test with only an intercept and both intercept and trend, respectively. The critical values in both cases are equal to 2:8695 and 3:4235;
so the second di¤erence of variable P is stationary.
Once the data is made stationary, we use a nonparametric approach for
the estimation and the Akaike’s information criterion to specify the orders
of the V AR(k) models which describe the constrained and unconstrained
processes. To choose the upper bound K on the admissible lag orders, we
apply the results by Lewis and Reinsel (1985) [23].
Using the Akaike’s criterion for the unconstrained V AR model, which
correspond to four variables, we observe that this one is minimal at k = 16.
We use same criterion to specify the orders of the constrained V AR models,
which correspond to di¤erent combinations of three variables, and we …nd
that orders are all less than or equal to 16. To compare the determinants
of the covariance matrices of the constrained and unconstrained forecast
errors at horizon h, we take the same order k = 16 for the constrained and
unconstrained models.
We calculate di¤erent causality measures for horizons h = 1; :::; 40 [ see
Figures 9-14 in appendix]. Higher values of measures indicate greater causality. We also calculate the corresponding nominal 95% bootstrap con…dence
intervals as it is described in the previous section.
41
Based on results, we have following conclusions. From Figure 9 we remark that nonborrowed reserves (N BR) has an important e¤ect on federal
funds rate (r) at horizon 1 comparatively with other variables [see Figures
10 and 11 ]. Thus, we can conclude that nonborrowed reserves causes the
federal funds rate only in the short-term. This e¤ect is well known in the
literature and can be explain by the theory of supply and demand of the
money. We also note that nonborrowed reserves has a short-term e¤ect on
the GDP [see Figure 10] and it can causes the GDP de‡ator until horizon
4:
Figure 14 shows that the e¤ect of GDP on federal funds rate is signi…cant
for the …rst four horizons. However, the e¤ect of federal funds rate on GDP
de‡ator is signi…cant only at horizon 1 [see Figure 12].
The other more signi…cant result concern the causality from r to GDP:
Figure 13 shows that the e¤ect of the interest rate on GDP is signi…cant
until horizon 16. These results are consistent with conclusions obtained by
Dufour et al. (2005).
Table 6 represents the results of other causality directions until horizon
20. As we can read from this table there is noncausality in these directions.
The above results do not change when we consider the second rather
than …rst di¤erence of variable P .
10
Conclusion
New concepts of causality have been introduced in the literature by Dufour
and Renault (1998, Econometrica) [9]: they de…ne a causality at a given
(arbitrary) horizon h, and causality up to any given horizon h, where h is
a positive integer and can be in…nite (1
h
1). These concepts are
motivated by the fact that, in the presence of an auxiliary variable Z, it is
possible to have situation where the variable Y does not cause variable X
at horizon 1, but causes it at horizon h > 1. In this case, this is an indirect
causality transmitted by the auxiliary variable Z.
Another related problem consists in measuring the importance of causality between two variables. Existing causality measures have been established only for the horizon 1 and fail to capture indirect causal e¤ects. This
paper proposes a generalization of such measures for any horizon h. We
propose parametric and nonparametric measures for feedback and instantaneous e¤ects at any horizon h. Parametric measures are de…ned in terms
of impulse response coe¢ cients of V M A representation. By analogy with
Geweke (1982) [11], we show that it is always possible to de…ne a mea-
42
sure of dependence at horizon h which can be decomposed into a sum of
feedback measures from X to Y; from Y to X; and instantaneous e¤ect at
horizon h. We also show how these causality measures can be related to the
predictability measures developed in Diebold and Kilian (1998) [5].
We propose a new approach to estimate these measures based on simulation of a large sample from the process of interest. We also propose a
valid nonparametric con…dence intervals, using the bootstrap technique.
From an empirical application we …nd that nonborrowed reserves causes
federal funds rate only in short-term, the e¤ect of real gross domestic product on federal funds rate is signi…cant for the …rst four horizons, the e¤ect
of federal funds rate on gross domestic product de‡ator is signi…cant only
at horizon 1; and …nally federal funds rate causes the real gross domestic
product until horizon 16:
43
References
[1] B. S. Bernanke and I. Mihov. Measuring monetary policy. The
Quarterly Journal of Economics 113(3), 869–902 (1998).
[2] R. J. Bhansali. Linear prediction by autoregressive model …tting in
the time domain. Ann. Statist. 6, 224–231 (1978).
[3] H. Boudjellaba, J.-M. Dufour, and R. Roy. Testing causality
between two vectors in multivariate ARMA models. Journal of the
American Statistical Association 87, 1082–1090 (1992).
[4] H. Boudjellaba, J.-M. Dufour, and R. Roy. Simpli…ed conditions
for non-causality between two vectors in multivariate ARMA models.
Journal of Econometrics 63, 271–287 (1994).
[5] F. X. Diebold and L. Kilian. Measuring predictability: Theory
and macroeconomic applications. Journal of Applied Econometrics 16,
657–669 (2001).
[6] D. Dufour, D. Pelletier, and E. Renault. Short run and long run
causality in time series: Inference. Forthcoming in Journal of Econometrics (2003).
[7] J.-M. Dufour and T. Jouini. Asymptotic distribution of a simple linear estimator for VARMA models in echelon form. forthcoming
in Statistical Modeling and Analysis for Complex Data Problems, ed.
by Pierre Duchesne and Bruno Remillard, Kluwer, The Netherlands
(2004).
[8] J.-M. Dufour and D. Pelletier. Linear estimation of weak
VARMA models with a macroeconomic application. Working Paper
(2003).
[9] J.-M. Dufour and E. Renault. Short-run and long-run causality in
time series: Theory. Econometrica 66(5), 1099–1125 (1998).
[10] B. Efron and R. J. Tibshirani. “An Introduction to the Bootstrap”.
New York: Chapman Hall (1993).
[11] J. Geweke. Measurement of linear dependence and feedback between
multiple time series. Journal of the American Statistical Association
77(378), 304–313 (1982).
44
[12] J. Geweke. “Inference and Causality in Economic Time Series”. in
Handbook of Econometrics, Volume 2, ed. by Z. Griliches and M. D.
Intrilligator. Amsterdam: North-Holland, pp. 1102-1144. (1984).
[13] J. Geweke. Measures of conditional linear dependence and feedback
between time series. Journal of the American Statistical Association
79, 907–915 (1984).
[14] C. Gouriéroux, A. Monfort, and E. Renault. Kullback causality
measures. Annales D’Économie et De Statistique 6/7, 369–410 (1987).
[15] C. W. J. Granger. Investigating causal relations by econometric
models and cross-spectral methods. Econometrica 37, 424–459 (1969).
[16] E. J. Hannan and L. Kavalieris. Multivariate linear time series
models. Advances in Applied Probability 16, 492–561 (1984b).
[17] E. J. Hannan and J. Rissanen. Recursive estimation of mixed
autoregressive-moving average order. Biometrika 69, 81–94. Errata
70 (1983), 303 (1982).
[18] C. Hsiao. Autoregressive modeling and causal ordering of economic
variables. Journal of Economic Dynamics and Control 4, 243–259
(1982).
[19] A. Inoue and L. Kilian. Bootstraping smooth functions of slope
parameters and innovation variances in VAR(In…nite) models. International Economic Review 43, 309–332 (2002).
[20] H. Kang. Necessary and su¢ cient conditions for causality testing in
multivariate ARMA model. Journal of Time Series Analysis 2, 95–101
(1981).
[21] L. Kilian. Small-sample con…dence intervals for impulse responses
functions. Review of Economics and Statistics 80, 218–230 (1998a).
[22] S. G. Koreisha and T. M. Pukkila. Fast linear estimation methods
for vector autorgessive moving-average models. Journal of Time Series
Analysis 10(4), 325–339 (1989).
[23] R. Lewis and G. C. Reinsel. Prediction of multivariate time series
by autoregressive model …tting. Journal of Multivariate Analysis 16,
393–411 (1985).
45
[24] H. Lütkepohl. “Introduction to Multiple Time Series Analysis”.
Berlin: Springer-Verlag (1993).
[25] P. Newbold. “Causality Testing in Economics”. in Time Series Analysis: Theory and Practice 1, ed. by O. D. Anderson. Amsterdam: NorthHolland. (1982).
[26] E. Parzen. Some recent advances in time series modelling. IEEE
Trans. Automat. Control AC-19 (1974).
[27] D. A. Pierce and L. D. Haugh. Causality in temporal systems:
Characterizations and survey. Journal of Econometrics 5, 265–293
(1977).
[28] W. Polasek. Temporal causality measures based on AIC. in H. Bozdogan, Proceedings of the Frontier of Statistical Modeling: An Informal
Approach, Kluwer, Netherlands pp. 159–168 (1994).
[29] W. Polasek. Bayesian causality measures for multiple ARCH models
using marginal likelihoods. Working Paper (2000).
[30] C. Sims. Money, income and causality. American Economic Review
pp. 540–552 (1972).
[31] C. Sims. Macroeconomics and reality. Econometrica 48, 1–48 (1980).
[32] N. Wiener. “The Theory of Prediction”. In The Theory of Prediction,
Ed by E. F. Beckenback. New York: McGraw-Hill, Chapter 8. (1956).
46
Proof of proposition 1. From de…nition 3 we have,
C(Y
!XjZ )
h2
det (X(t + h2 ) j It Yt )
¯
det (X(t + h2 ) j It )
det (X(t + h1 ) j It Yt )
= ln
¯
det (X(t + h1 ) j It )
det (X(t + h1 ) j It ) det (X(t + h2 ) j It Yt )
+ ln
¯
det (X(t + h1 ) j It Yt ) det (X(t + h2 ) j It )
¯
= C(Y !XjZ )
= ln
h1
det
det
det
ln
det
(X(t + h1 ) j It )
(X(t + h2 ) j It )
(X(t + h1 ) j It Yt )
:
¯
(X(t + h2 ) j It Yt )
¯
Using the predictability measure of Diebold and Kilian (1998) [5] we have,
+ ln
PX (It
det (X(t + h1 ) j It
det (X(t + h2 ) j It
Yt ; h1 ; h2 ) = 1
¯
and
det (X(t + h1 ) j It )
:
det (X(t + h2 ) j It )
PX (It ; h1 ; h2 ) = 1
Hence the result of proposition 1
C(Y
!XjZ )
h1
C(Y
!XjZ )
= ln [1
PX (It
Yt ; h1 ; h2 )] ln [1
¯
h2
^ Y
Proof (Consistency of C(
Yt )
¯ ;
Yt )
¯
!XjZ )).
PX (It ; h1 ; h2 )] :
Under assumption 1 we have
h
G(vec( ^ c (k)); vech( ^ ";k ))
= (1 +
2k
T )G(vec(
= (1 + O(T
for
2
3
<
c ); vech(
))G(vec(
" ))
c ); vech(
" ));
< 1:
The last equality follows from condition 2 of assumption 1: if we consider
that k = T for > 0, then condition 2 implies that k 3 =T = T 3 1 with
47
1 and T (2T
0 < < 13 : Similarly, we have 2k
T = 2T
Thus,
ln(G(vec( ^ c (k)); vech( ^ ";k )))
= ln(G(vec(
c ); vech(
= ln(G(vec(
c ); vech(
" ))
1)
! 2 for
+ ln(1 + O(T
2
3
<
< 1:
))
(38)
for 23 <
+ O(T
);
< 1 and T ! 1:
In other ways we have ^ !
; ^u !
p
in (vec( ); vech(
" ))
u;
p
and for H(:) a continuous function
u ))
H(vec( ^ ); vech( ^ u )) ! H(vec( ); vech(
u ));
p
or
ln(H(vec( ^ ); vech( ^ u ))) ! ln(H(vec( ); vech(
p
u ))):
(39)
From ( 38) and (39) we get
^ Y
C(
!XjZ ) = ln(
h
for
2
3
<
G(vec( c ); vech(
H(vec( ); vech(
" ))
u ))
) + O(T
) + op (1);
< 1 and T ! 1:
So, under condition 2 of assumption 1, we have,
^ Y
C(
!XjZ )
h
! C(Y
p
!XjZ ):
h
Proof of Proposition 4. We have showed [see proof of consistency]
that
ln(G(vec( ^ c (k)); vech( ^ ";k )))
= ln(G(vec(
c ); vech(
" ))
+ O(T
)
(40)
= ln(G(f (vec( ); vech(
for
2
3
<
u )))
+ O(T
);
< 1:
Under assumption 2 we have,
ln(G(f (vec( ^ ); vech( ^ u )))) ! ln(G(f (vec( ); vech(
p
48
u )))):
(41)
From (40) and (41) we get
ln(G(vec( ^ c (k)); vech( ^ ";k )))
= ln(G(f (vec( ^ ); vech( ^ u )))) + O(T
for
2
3
<
(42)
) + op (1);
< 1:
Thus,
^ Y
C(
!XjZ )
= ln(
h
for
2
3
<
G(f (vec( ^ ); vech( ^ u )))
) + O(T
H(vec( ^ ); vech( ^ u ))
< 1:
^
^
~ Y
We denote ln( G(f (vec(^ );vech(^ u ))) ) by C(
H(vec( );vech(
u ))
!XjZ ):
h
^ Y
Op (1); the asymptotic distribution of C(
~ Y
of C(
) + op (1);
^
^
Since ln( G(f (vec(^ );vech(^ u ))) ) =
H(vec( );vech(
u ))
!XjZ ) will be the same as that
h
!XjZ ).
h
~ Y
Under assumption 2 and by a …rst order Taylor expansion of C(
around the true parameters we get,
~ Y
C(
!XjZ )
h
= C(Y
@C(Y
+(
!XjZ )
h
!XjZ )
h
!XjZ )
h
0
@(vec( ) ; vech(
0
u) )
)
vec( ^ )
vech( ^ u )
vec( )
vech( u )
+ op (T
1
2
);
and
~ Y
p)1=2 (C(
(T
@C(Y
' (
!XjZ )
C(Y
h
!XjZ )
h
0
@(vec( ) ; vech(
0
u) )
)
!XjZ ))
h
p)1=2 vec( ^ )
u)
(T
(T
p)1=2 vech( ^
vec( )
vech( u )
According to (36),
(T
~ Y
p)1=2 (C(
!XjZ )
h
C(Y
!XjZ ))
h
d
! N (0;
C ):
Thus,
(T
p)1=2 (C^Y
!XjZ
h
CY
49
!XjZ )
h
d
! N (0;
C );
:
where
@C(Y
C
=(
!XjZ )
@C(Y
h
0
@(vec( ) ; vech(
0
u)
)
) (
!XjZ )
0
h
0
@(vec( ) ; vech(
0
u)
)
);
and
1
=
0
u
0
0
2(D3 D3 )
1 D0 (
u
3
0
u )D3 (D3 D3 )
1
:
D3 is the duplication matrix, de…ned such that vech(F ) = D3 vech(F ) for
any symmetric 3 3 matrix F .
Proof (Asymptotic validity of the residual-based bootstrap).
Conditionally on the sample one can use a similar argument as in the proof
of the asymptotic distribution to establish the asymptotic validity of the
residual-based bootstrap. It su¢ ces to take vec( ^ ) and vech( ^ u ) as true
parameters.
50
Figure 1: NBR in logarithmic form
11.2
ln(NBR)
11
10.8
10.6
ln(NBR)
10.4
10.2
10
9.8
9.6
9.4
9.2
Jan65
Jan70
Jan75
Jan80
Jan85
Time (t)
Jan90
Jan95
Jan00
Figure 2: r in logarithmic form
3
ln(r)
2.8
2.6
2.4
ln(r)
2.2
2
1.8
1.6
1.4
1.2
1
Jan65
Jan70
Jan75
51
Jan80
Jan85
Time (t)
Jan90
Jan95
Jan00
Figure 3: P in logarithmic form
5
series 1
4.8
4.6
ln(P)
4.4
4.2
4
3.8
3.6
3.4
3.2
Jan65
Jan70
Jan75
Jan80
Jan85
Time (t)
Jan90
Jan95
Jan00
Jan95
Jan00
Figure 4: GDP in logarithmic form
8.9
ln(GDP)
8.8
8.7
8.6
ln(GDP)
8.5
8.4
8.3
8.2
8.1
8
7.9
Jan65
Jan70
Jan75
52
Jan80
Jan85
Time (t)
Jan90
Figure 5: The first dfferentiation of ln(NBR)
0.08
Growth rate of NBR
0.06
Growth rate of NBR
0.04
0.02
0
-0.02
-0.04
-0.06
-0.08
Jan65
Jan70
Jan75
Jan80
Jan85
Time (t)
Jan90
Jan95
Jan00
Figure 6: The first dfferentiation of ln(r)
0.3
Growth rate of r
0.2
Growth rate of r
0.1
0
-0.1
-0.2
-0.3
-0.4
-0.5
Jan65
Jan70
Jan75
53
Jan80
Jan85
Time (t)
Jan90
Jan95
Jan00
15
x 10
-3
Figure 7: The first dfferentiation of ln(P)
Growth rate of P
Growth rate of P
10
5
0
-5
Jan65
Jan70
Jan75
Jan80
Jan85
Time (t)
Jan90
Jan95
Jan00
Figure 8: The first dfferentiation of ln(GDP)
0.025
Growth rate of GDP
0.02
0.015
Growth rate of GDP
0.01
0.005
0
-0.005
-0.01
-0.015
-0.02
-0.025
Jan65
Jan70
Jan75
54
Jan80
Jan85
Time (t)
Jan90
Jan95
Jan00
Table 4: Augmented Dickey-Fuller Test (variables in logarithmic form)
N BR
R
P
GDP
With Intercept
ADF test statistic 5% Critical Value
0:510587
2:8694
2:386082
2:8694
1:829982
2:8694
1:142940
2:8694
With Intercept and Trend
ADF test statistic 5% Critical Value
1:916428
3:4234
2:393276
3:4234
0:071649
3:4234
3:409215
3:4234
Table 5: Augmented Dickey-Fuller Test (First di¤erence)
N BR
r
P
GDP
With Intercept
ADF test statistic 5% Critical Value
5:956394
2:8694
7:782581
2:8694
2:690660
2:8694
5:922453
2:8694
55
With Intercept and Trend
ADF test statistic 5% Critical Value
5:937564
3:9864
7:817214
3:9864
3:217793
3:9864
5:966043
3:9864
Figure 9: Causality measures from Nonborrowed reserves to Federal funds rate
0.25
95% percentile bootstrap interval
OLS point estimate
Causality Measures
0.2
0.15
0.1
0.05
0
0
5
10
15
20
Horizons
25
30
35
40
Figure10: Causality measures from Nonborrowed reserves to GDP Deflator
0.16
95% percentile bootstrap interval
OLS point estimate
0.14
Causality Measures
0.12
0.1
0.08
0.06
0.04
0.02
0
0
5
10
5615
20
Horizons
25
30
35
40
Figure 11: Causality measures from Nonborrowed reserves to Real GDP
0.16
95% percentile bootstrap interval
OLS point estimate
0.14
Causality Measures
0.12
0.1
0.08
0.06
0.04
0.02
0
0
5
10
15
20
Horizons
25
30
35
40
Figure 12: Causality measures from Federal funds rate to GDP Deflator
0.18
95% percentile bootstrap interval
OLS point estimate
0.16
Causality Measures
0.14
0.12
0.1
0.08
0.06
0.04
0.02
0
0
5
10
5715
20
Horizons
25
30
35
40
Figure 13: Causality measures from Federal funds rate to Real GDP
0.16
95% percentile bootstrap interval
OLS point estimate
0.14
Causality Measures
0.12
0.1
0.08
0.06
0.04
0.02
0
0
5
10
15
20
Horizons
25
30
35
40
Figure 14: Causality measures from Real GDP to Federal funds rate
0.35
95% percentile bootstrap interval
OLS point estimate
0.3
Causality Measures
0.25
0.2
0.15
0.1
0.05
0
0
5
10
15
58
20
Horizons
25
30
35
40
Download