Long Memory versus Stochastic Trend

advertisement
Long Memory versus Stochastic Trend
Alexander Aue
1
Lajos Horváth 1
Josef Steinebach
2
Abstract: We study the limiting behaviour of the prominent V /S and R/S test statistics,
aimed at detecting long–range dependence, under the presence of a stochastic trend,
which is given by cumulative random shocks. Depending on the size of these shocks, the
asymptotic distribution is characterized either solely by the stochastic trend, or solely by
the error sequence, or it is a mixture of both random components. Moreover, we also
derive the rate of convergence to the limit.
AMS 2000 Subject Classification: Primary 62M10; Secondary 91B70
Keywords and Phrases: long memory, long–range dependence, stochastic trend, V /S
statistic, R/S statistic, compound Poisson process, convergence rate
1
Department of Mathematics, University of Utah, 155 South 1440 East, Salt Lake City, UT 84112–
0090, USA, emails: aue@math.utah.edu and horvath@math.utah.edu
2
Mathematisches Institut, Universität zu Köln, Weyertal 86–90, D–50931 Köln, Germany, email:
jost@math.uni-koeln.de
Corresponding author: A. Aue, phone: +1–801–581–5231, fax: +1–801–581–4148.
Research partially supported by NATO grant PST.EAP.CLG 980599 and NSF–OTKA grant INT–
0223262
1
Introduction
Many empirical data sets obtained in scientific fields as diverse as geophysics, telecommunications, computer networks and economics [see, for instance, Giraitis et al. (2001) and
the references therein] give rise to determining long–range dependence and to discriminating it from other effects such as [deterministic or stochastic] drifts and level shifts under
weak dependence. The question originally arose in the context of measuring the waterflow of the Nile river. Hurst (1951) reported convincingly that data exihibit the so–called
Hurst effect, which will be explained briefly in the following. We say that a sequence of
random variables is weakly dependent if it is second–order stationary with an absolutely
summable covariance function, whereas the sequence is said to be long–range dependent
if it is second–order stationary with covariances that are not absolutely summable.
Let X1 , . . . , Xn be a collection of random variables. The most prominent test statistic
for detecting long–range dependence is the R/S statistic Qn = Rn /sn , where
Rn = max
1≤k≤n
k
X
(Xj − X̄n ) − min
j=1
1≤k≤n
k
X
(Xj − X̄n )
(1.1)
j=1
is the adjusted range of the partial sums of the Xj ’s,
s2n =
n
1X
(Xj − X̄n )2
n j=1
(1.2)
is the sample variance, and X̄n = n−1 nj=1 Xj is the sample mean. We say that the
Hurst effect is present if, for some H > 1/2, the ratio n−H Qn converges to a limit which
is nonzero with probability one but possibly degenerate. The parameter H is called the
Hurst exponent. There are various controversial theoretical approaches to explain the
phaenomenon. One of them is the presence of long–range dependence in the underlying
sequence, a second one is the presence of a trend function along with weak dependence.
For example, Battacharya et al. (1983) studied the model
P
Xk = fn (k) + εk ,
1 ≤ k ≤ n,
where fn (k), 1 ≤ k ≤ n, is an arbitrary deterministic trend function and {εk } is an
ergodic sequence of short memory random variables. Examining, e.g., the special trend
f (k) = ck β , c 6= 0, they could show that the Hurst effect is present if −1/2 < β < 0 and
β > 0. In the first case it could also be confused with long memory.
2
A second test statistic, the so–called V /S statistic Tn = Vn /s2n , which is based on the
comparison of the sample variances instead of the adjusted range Rn given in (1.1), was
introduced by Giraitis et al. (2001). Here,
Vn =
1
n2


k
n
X
X


(Xj

k=1
j=1
2
− X̄n ) −

k
n X
X
1
(Xj
n k=1 j=1
2 
− X̄n ) 
.
(1.3)
The latter statistic possesses somewhat nicer asymptotic and finite–sample properties.
Moreover, the findings of Bhattacharya et al. (1983)—which involved the R/S statistic—
were transfered also to the V /S statistic [see Giraitis et al. (2001, 2003)].
Leipus and Viano (2003) studied the asymptotics of the V /S statistic in a setting with
stochastic trend
fn (k) =
k
X
j=1
cj,n ηj ,
1 ≤ k ≤ n,
where—amongst other conditions—it is assumed that
{cj,n , 1 ≤ j ≤ n} are independent, identically Bernoulli variables with
P {c1,n = 1} = π/n and P {c1,n = 0} = 1 − π/n for some 0 < π < ∞,
(1.4)
{ηj , 1 ≤ j ≤ n} are independent, identically distributed random variables
(1.5)
such that E|η1 |ν < ∞ for some ν > 0,
(1.6)
{cj,n , 1 ≤ j ≤ n} and {ηj , 1 ≤ j ≤ n} are independent.
(1.7)
and
The setting is referred to as mixture model and has been examined, for instance, by
Diebold and Inoue (2001), and Granger and Hyung (2004). Note that the Bernoulli
parameter π does not depend on n. Therefore, with increasing n, random shocks modelled by fn (k) occur very rarely. Properties of the autocorrelation function and the log
periodogram estimate have been reported in Perron and Qu (2004).
First aim of this paper is to reinvestigate the results of Leipus and Viano (2004) under
a minimal set of assumptions [see Theorem 2.1]. Furthermore, we show that the limit
theorems obtained hold true also in an extended setting in which the stochastic trend
function does not solely determine the limit behaviour anymore [see Theorem 2.2]. To
3
this end, the trend function fn (k) will be modified so that it contains a sequence {an }
converging to zero at a certain rate. Depending on this specific rate, the limit distribution
can be either described in terms of the trend, or in terms of the noise, or it is a mixture of
both. Also, the setting allows for weakly dependent errors {εk } [see Theorem 2.3]. Finally,
we will answer a question of Perron and Qu (2004), by giving the rate of convergence to
the limit distribution in Theorem 2.1 [see Theorem 2.4].
The paper is organized as follows. In Section 2 we collect all results. Proofs shall be
given in Section 3.
2
Results
Throughout, we will work with all or some of the following assumptions. In addition to
(1.4)–(1.7) it is required that
{εk : k ≥ 1}, {cj,n : 1 ≤ j ≤ n} and {ηj : j ≥ 1} are independent,
(2.1)
{εk : k ≥ 1} are independent, identically distributed random variables
(2.2)
E|ε1 |µ < ∞ for some µ > 4.
(2.3)
such that Eε1 = 0 and 0 < σ 2 = Eε21 < ∞,
and
2.1
A limit theorem and some extensions
We start by introducing some more notations. Let {N(t) : t ≥ 0} be a homogeneous
Poisson process with intensity parameter π [that is, EN(t) = πt for all t ≥ 0] which is
independent of {ηj : j ≥ 1}. Set
N (t)
P(t) =
X
ηj ,
j=1
P0 (t) = P(t) −
Z
0
1
P(s)ds,
t ∈ [0, 1].
(2.4)
Note that {P(t) : t ≥ 0} is a compound Poisson process. In the next section, we shall
prove that the process {Un (t) : t ∈ [0, 1]} defined by
⌊nt⌋
Un (t) =
X
i=1
ci,n ηi ,
t ∈ [0, 1],
(2.5)
4
converges in D[0, 1], the Skorohod space [cf. Billingsley (1968)], to the compound Poisson
process. Therein, ⌊·⌋ denotes the integer part. In particular, we obtain the following
convergence in distribution results for the test statistics Qn and Tn .
Theorem 2.1 Let (1.4), (1.5), (2.1) and (2.2) be satisfied. Then, as n → ∞,
Qn D
−→
n
Z
Tn D
−→
n
Z
1
0
P02 (s)ds
+σ
−1/2 "
sup
−1 "Z
Z
2
Z
t
0<t<1 0
P0 (s)ds − inf
Z
t
1
Z
P0 (s)ds
0<t<1 0
#
and
1
P02 (s)ds
0
+σ
2
1
0
t
0
P0 (s)ds
2
dt −
Z
0
0
t
2 #
P0 (s)dsdt
.
It can be seen that the cumulative effect of the random shocks of fn (k) completely
dominates the limiting behaviour of both test statistics Qn and Tn , though their single
impacts are small, while the errors {εk } contribute only via the variance σ 2 .
In the following, we shall study even smaller shock sizes. Introducing a sequence {an }
satisfying
an → 0
(n → ∞),
(2.6)
the stochstic trend is modified to
fn (k) =
k
X
cj,n an ηj ,
j=1
1 ≤ k ≤ n,
(2.7)
that is, the jump size, now given by an ηj , gets arbitrarily small for increasing n. Depending
on the rate of convergence of {an } to zero, we get different limit results, which√
are collected
in the next theorem. More exactly, if an is asymptotically larger than 1/ √n then the
random shocks give the limit distribution. If an is exactly of the order 1/ n then the
asymptotic is determined by a mixture of the random
shocks and and the errors {εk }.
√
Finally, if an is asymptotically smaller than 1/ n then the errors are dominating the
cumulative shocks in the long–run.
Denote by {W0 (t) : t ∈ [0, 1]} a Brownian bridge independent of the compound Poisson
process {P(t) : t ∈ [0, 1]}.
Theorem
√ 2.2 Let (1.4), (1.5), (2.1), (2.2), (2.6) and (2.7) be satisfied.
(i) If nan → ∞, then, as n → ∞,
Qn D 1
−→
nan
σ
"
sup
Z
0<t<1 0
t
P0 (s)ds − inf
Z
t
0<t<1 0
5
#
P0 (s)ds .
(ii) If
√
nan → c 6= 0, then, as n → ∞,
"
t
Qn D 1
√ −→
sup c P0 (s)ds + σW0 (t) − inf c
0<t<1
n
σ 0<t<1
0
√
(iii) If nan → 0, then, as n → ∞,
Z
Z
t
0
P0 (s)ds + σW0 (t)
#
.
Qn D
√ −→ sup W0 (t) − inf W0 (t).
0<t<1
n
0<t<1
√
In part (i) of Theorem 2.2, the test statistic Qn is normalized with nan ≫ n, which
is exactly the normalization in the presence of long memory [cf. the discussion on the
Hurst parameter H in the introduction]. However, the limit distribution is completely
different from the expected fractional Brownian motion under long–range dependence.
Straightforward adaptations establish the same limit theorems also for the statistic Tn .
It turns out that Theorems 2.1 and 2.2 can easily be extended to weakly dependent
errors. In this case assumption (2.2) has to be replaced by the following requirement:
There are a Wiener process {W (t) : t ∈ [0, 1]} and a constant τ > 0 such that, as n → ∞,
⌊nt⌋
1 X D[0,1]
√
εi −→ τ W (t)
n i=1
n
1X
P
ε2i −→ σ 2 > 0
n i=1
and
(2.8)
with D[0, 1] denoting the Skorohod space. Then, we obtain the following modification of
the previous theorems. Note that one can choose W0 (t) = W (t) − tW (1), t ∈ [0, 1], as
Brownian bridge. In particular, {W0 (t) : t ∈ [0, 1]} is independent of {P0 (t) : t ∈ [0, 1]},
since, by (2.1), the random variables {ηj } and {cj,n ) are independent of the {εk }.
Theorem 2.3 Let (1.4), (1.5), (2.1), (2.7) and (2.8) be satisfied.
(i) If an = 1, then, as n → ∞,
#
−1/2 "
Z t
Z t
Qn D Z 1 2
2
P0 (s)ds .
P0 (s)ds − inf
sup
P0 (t)dt + σ
−→
0<t<1 0
n
0
0<t<1 0
√
(ii) If (2.6) is satisfied and nan → ∞, then, as n → ∞,
Qn D 1
−→
n
σ
"
sup
Z
t
0<t<1 0
P0 (s)ds − inf
0<t<1 0
(iii) If (2.6) is satisfied and
Qn D 1
√ −→
n
σ
"
sup
0<t<1
Z
c
Z
0
t
√
t
#
P0 (s)ds .
nan → c 6= 0, then, as n → ∞,
P0 (s)ds + τ W0 (t) − inf
6
0<t<1
Z
c
0
t
P0 (s)ds + τ W0 (t)
#
.
(iv) If (2.6) is satisfied and
Q D τ
√n −→
n
σ
"
√
nan → 0, then, as n → ∞,
#
sup W0 (t) − inf W0 (t) .
0<t<1
0<t<1
As before, Theorem 2.3 can be reformulated in terms of the test statistic Tn . The proofs
of all theorems are given in Section 3.
2.2
Rates of convergence
In this small subsection, we address the question of determining the rate of convergence
to the limit distributions in Theorem 2.1, which are given as functionals of a compound
Poisson process. The question of how fast the asymptotics are reached has, for instance,
been discussed in Perron and Qu (2004), who studied the limiting behaviour of test
statistics under the presence of a random level shift.
The proof of Theorem 2.1 shows that one needs to obtain the rate of convergence of
the process
⌊nt⌋
Un (t) =
X
ci,n ηi
i=1
to the compound Poisson limit P(t). Since both {Un (t) : t ∈ [0, 1]} and {P(t) : t ∈ [0, 1]}
are discontinuous, it is very unlikely that an approximation in the supremum norm can
be achieved. Both processes, however, are in D[0, 1], so we will use the natural Skorohod
metric d there to measure the distance between Un (t) and suitably constructed versions
of its limit.
R
R
On the other hand, the integral processes { 0u Un (s)ds : u ∈ [0, 1]} and { 0u Pn (s)ds : u ∈
[0, 1]}, with {Pn (t) : t ≥ 0} denoting compound Poisson processes defined in Theorem 2.4
below, are both continuous with probability one, so that the uniform distance can be
applied conveniently.
Recall that {N(t) : t ≥ 0} denotes a Poisson process with intensity parameter π.
Theorem 2.4 Let (1.4), (1.5), (2.1), (2.2) and (2.3) be satisfied, and let
E|η1 |ν < ∞ with some ν ≥ 2.
Then there are compound Poisson processes {Pn (t) : t ≥ 0} such that, for all n ≥ 1,
(i)
D
{Pn (t) : t ≥ 0} =

(t)
N
X

i=1


ηi : t ≥ 0 ,
7

1
P Qn − Q∗n > Cn−ν/(3(1+ν)) ≤ Cn−ν/(3(1+ν))
n
with a constant C > 0 and
Q∗n
(ii)
=
Z
1
0
2
Pn,0
(s)ds
+σ
Therein, Pn,0 (t) = Pn (t) −
R1
0
2
−1 "
sup
Z
0<t<1 0
t
Pn,0 (s)ds − inf
Z
0<t<1 0
t
#
Pn,0 (s)ds .
Pn (s)ds for all t ∈ [0, 1].
The proof will be given in the following section.
3
Proofs
The section is organized as follows. In a first part, we introduce some auxiliary results,
while Theorems 2.1–2.4 are proved in the successive subsections.
Our first result provides an approximation for the shock process {Un (t) : t ∈ [0, 1]}
given in (2.5) with the compound Poisson process {P(t) : t ∈ [0, 1]}. All processes are
elements of D[0, 1] which will be endowed with the Skorohod metric d.
For j ≥ 1, let ηj+ and ηj− denote the positive, respectively, negative part of ηj .
Lemma 3.1 Let (1.4), (1.5) and (1.6) be satisfied. Then
d(Un , P) = OP
1
n
(n → ∞),
where Un is defined in (2.5).
Proof. Is is well–known that

n (t)
SX
D
{Un (t) : t ∈ [0, 1]} =

where the partial sums
i=1


ηi : t ∈ [0, 1] ,

⌊nt⌋
Sn (t) =
X
i=1
ci,n ,
t ∈ [0, 1],
(3.1)
count the number of shocks up to time ⌊nt⌋. Applying Major (1990), there exist homogeneous Poisson processes {Nn (t) : t ∈ [0, 1]} having intensity parameter π such that
P
(
sup Sn (t) − Nn
0≤t≤1 ⌊nt⌋
n
!
)
=0 ≥1−
8
c
n
(3.2)
with some constant c, and
{Nn (t) : t ∈ [0, 1], n ≥ 1} and {ηj : j ≥ 1} are independent.
(3.3)
Letting tn = ⌊nt⌋/n, by (3.2) we have
P


Sn (t)
X
sup
ηi
0≤t≤1 i=1
ηi Nn (tn )
X
−
i=1


= 0 ≥ 1 −
c
.
n
(3.4)
Since the distribution of Nn (t) does not depend on n [the intensity parameter remains
the same], and since it is also independent of {ηj : j ≥ 1} for any n, it suffices to estimate
the distance

N (t)
N (tn )
X
X
d
ηi ,
i=1
i=1

ηi  .
First, it holds for the positive and negative parts of the {ηj : j ≥ 1} that
X
i=1
ηi± −
X
i=1
N (t−1/n)
N (t)
N (tn )
N (t)
0≤
ηi± ≤
X
i=1
ηi± −
ηi± .
X
i=1
[Throughout, we will set N(x) = 0 if x < 0.] Hence, it is enough to obtain upper bounds
for

N (t−1/n)
N (t)
d
X
ηi+ ,
X
i=1
i=1
To do so, define
ρ(t) =




















ηi+ 

N (t−1/n)
N (t)
and
2t
d
X
ηi− ,
i=1
1
n
:
1
n−2
n−1
t−
+
2
n
n
:
i=1
1
n
1
n−2
<t≤
n
n
n−2
<t≤1
n
: 0≤t≤
t+
X
9

ηi−  .
and observe that cleary ρ(0) = 0, ρ(1) = 1, and that ρ is strictly monotone increasing
and continuous. Also, for the positive parts {ηj+ : j ≥ 1},
N
(t)
X +
ηi
i=1
N (ρ(t)−1/n)
X
−
i=1
+ ηi ≤
Since
P

(1/n)
NX

ηi+ = 0



i=1

























N (1/n)
2
ηi+
X
i=1
0
:
N (1)
ηi+
X
1
.
n
n−2
1
<t≤
.
n
n
: 0≤t≤
:
i=N (1−2/n)+1
n−2
< t ≤ 1.
n
≥ P {N(1/n) = 0} = e−π/n
and, by stationary increments,
P



N (1)
ηi+ = 0
X


≥ P {N(1) − N(1 − 2/n) = 0} = e−2π/n ,

i=N (1−2/n)+1
the proof is complete on recognizing that similar arguments apply to the negative parts
{ηj− : j ≥ 1}.
2
The second auxiliary result shows that one can define a sequence of compound Poisson
processes, all of which have the same distribution as the original {P(t) : t ∈ [0, 1]}, close
to the processes {Un (t) : t ∈ [0, 1]}. Here, distances can be taken with respect to the
uniform metric.
Lemma 3.2 If (1.4)–(1.7) hold, then compound Poisson processes {Pn (t) : t ∈ [0, 1]}
can be defined such that
D
{Pn (t) : t ∈ [0, 1]} = {P(t) : t ∈ [0, 1]}
for all n,
and
P
(
sup
0≤t≤1
Z
0
t
Un (s)ds −
Z
0
with some constants c1 and c2 .
t
Pn (s)ds
−ν/(1+ν)
> c1 n
10
)
≤ c2 n−ν/(1+ν)
for all n
Proof. Recall that tn = ⌊nt⌋/n. By (3.4) it is enough to consider
Z
0
u

N (t)
N (tn )
X
X

i=1
ηi −
i=1


ηi dt.

We examine positive parts and negative parts separately again. It is easy to see that, for
the {ηj+ : j ≥ 1},
0 ≤
Z
u
≤
≤
Z
1
0
0

ηi+ −
X
i=1
N (t)


N (tn )
N (t)

0
Z

X
i=1
ηi+  dt
N (t−1/n)
X
i=1
(t)
1/n N
X
ηi+ −
X
ηi+ dt
Z
i=1
+
i=1

ηi+  dt
(t)
1 N
X
1/n i=1
ηi+ dt
−
Z
0
1 N (t−1/n)
X
ηi+ dt
i=1
(t)
Z 1−1/n N
(t)
Z 1 N
N (1/n)
X
X
1 X +
+
≤
ηi+ dt
ηi dt −
ηi +
n i=1
0
1/n i=1
i=1
Z 1 N
Z 1−1/n N
N (1/n)
(t)
(t)
X
X
1 X +
+
≤
ηi +
ηi dt −
ηi+ dt
n i=1
1/n i=1
1/n
i=1
≤
N (1/n)
N (t)
Z 1
X
1 X +
ηi +
ηi+ dt
n i=1
1−1/n i=1
≤
(1/n)
(1)
1 NX
1 NX
ηi+ +
η+
n i=1
n i=1 i
≤
N (1)
2 X +
η .
n i=1 i
By Minkowski’s inequality [cf. Hardy,Littlewood and Pólya (1952)] we have that
E
k
X
i=1
ηi+
!ν

k
X


ν


Eηi+


≤





i=1
k
X
i=1
= kEη1+
ν
[Eηi+ ]1/ν
!ν
ν
: 0 < ν ≤ 1.
= k ν Eη1+
11
ν
: 1 < ν < ∞.
The latter result implies that

ν
N (1)
E
X
i=1
ηi+  < ∞,
so that Markov’s inequality yields
P

(1)
 2 NX
n
i=1
ηi+ ≥ n−ν/(1+ν)



≤ n1−ν/(1+ν)
−ν

N (1)
2ν E 
X
i=1
ν
ηi+  .
Using similar arguments in case of the negative parts {ηj− : j ≥ 1} completes the proof of
Lemma 3.2.
2
3.1
Proof of Theorem 2.1
Recall that Qn = Rn /sn , where Rn and sn are defined in (1.1) and (1.2), respectively.
The adjusted range Rn can be decomposed into one term involving {Un (t) : t ∈ [0, 1]}
and a second term containing the errors {εk : k ≥ 1}. Applications of Lemma 3.1 and
Donsker’s invariance principle yield further insight and lead to the desired limit results.
A similar approach works for the sample variance s2n . Details are given next.
Proof of Theorem 2.1. To begin with, we establish the limit distribution of the range
Rn . Note that from Lemma 3.1 we obtain that
D[0,1]
Un (t) −→ P(t)
(n → ∞),
(3.5)
where {P(t) : t ∈ [0, 1]} is the compound Poisson process given by (2.4). On the other
hand, Donsker’s theorem implies
X D[0,1]
1 ⌊nt⌋
√
εi −→ σW (t)
Sn (t) =
n i=1
(n → ∞),
(3.6)
where the processes {Sn (t) : t ∈ [0, 1]} are defined in (3.1) and {W (t) : t ∈ [0, 1]} denotes
a Wiener process. Hence,



k
n
X
Rn
1
kX
max  Un (j/n) −
=
Un (j/n)
n
n 1≤k≤n j=1
n j=1

(3.7)

n

kX
Un (j/n) + OP
− min  Un (j/n) −

1≤k≤n
n j=1
j=1
k
X
12
!
1
√
.
n
It is easy to see that
1
n
max
1≤k≤n
k
X
j=1
Un (j/n) − sup
Z
u
0<u<1 0
and therefore
1
n

max 
1≤k≤n
k
X
j=1
Un (j/n) −
2
≤
sup n 0<u<1
Z
u
Un (s)ds
1
≤
sup n 0<u<1
Z

n
X
k
Un (j/n) − sup
n j=1
0<u<1
Z
u
Z
u
0
u
0
Un (s)ds
Un (s)ds − u
Z
1
Un (s)ds − u
Z
1
0
Un (s)ds .
0
Un (s)ds By the same arguments,
1
n

min 
1≤k≤n
k
X
j=1
Un (j/n) −
2
≤
sup n 0<u<1
Z
u

n
X
k
Un (j/n) − inf
0<u<1
n j=1
0
Un (s)ds .
0
0
Un (s)ds So, using (3.6), it follows from (3.7) that
Rn
=
n
sup
0<u<1
Z
− inf
0<u<1
u
Un (s)ds − u
0
Z
0
u
Z
0
Un (s)ds − u
1
Un (s)ds
Z
1
0
(3.8)
Un (s)ds + oP (1)
(n → ∞).
Next, we deal with the asymptotics of the sample variance s2n . By definition, with ε̄n =
1 Pn
i=1 εi ,
n
s2n
n
X

n
X
2
n
1
1
1X


=
Un (j/n) +
(εj − ε̄n )2
Un (j/n) −
n j=1
n j=1
n j=1
n
n
1X
2X
(εj − ε̄n ) Un (j/n) −
Un (i/n) .
+
n j=1
n i=1
!
The weak law of large number yields
n
1X
P
(εj − ε̄n )2 −→ σ 2
n j=1
(n → ∞).
13
Following the lines of the proof of (3.8), one can verify that
2

n
n
1X
1X
Un (j/n) −
Un (j/n) =
n j=1
n j=1
Z
1
Z
Un (t) −
0
0
1
Un (s)ds
2
dt + oP (1).
Similarly,
n
n
1X
1X
(εj − ε̄n ) Un (j/n) −
Un (i/n)
n j=1
n i=1
=
!
n
n
1X
1X
εj Un (j/n) −
Un (j/n)ε̄n ,
n j=1
n j=1
where we conclude by Lemma 3.1 and (3.6),
n
1X
Un (j/n)ε̄n = OP
n j=1
1
√
n
!
(n → ∞).
Also,
n
1X
1
εj Un (j/n) = √
n j=1
n
Z
0
1
Un (t)dSn (t).
In light of (2.1), it follows from (3.5) and (3.6) that
D[0,1]
(Un (t), Sn (t)) −→ (P(t), σW (t))
(n → ∞),
where {P(t) : t ∈ [0, 1]} and {W (t) : t ∈ [0, 1]} are independent. Hence, an application of
Kurtz and Protter (1991) gives
Z
1
0
D
Un (t)dSn (t) −→ σ
Z
1
0
P(t)dW (t)
(n → ∞).
Thus we have
s2n
=
Z
0
1
Un (t) −
Z
0
1
Un (s)
2
ds + σ 2 + oP (1)
(n → ∞)
(3.9)
and the first part of Theorem 2.1 is now an immediate consequence of Lemma 3.1, (3.8)
and (3.9). Adaptations of these arguments also establish the limit for the V /S statistic
Tn = Vn /s2n , where Vn is given in (1.3).
2
14
3.2
Proofs of Theorems 2.2 and 2.3
We shall only give the proof of Theorem 2.2 here. Replacing assumption (2.2) with the
conditions in (2.8) throughout the proofs of Theorems 2.1 and 2.2 immediately implies
the results for weakly dependent errors collected in Theorem 2.3.
Proof of Theorem 2.2. Recall that the stochastic trend now contains a sequence {an }
converging to zero at a specified rate. At first, we establish the limit of the sample variance
s2n . Write
s2n
2
Z 1
n n
a2n X
1X
=
Un (j/n) −
(εj − ε̄n )2
Un (s)ds +
n j=1
n j=1
0
+
Z 1
n an X
Un (j/n) −
Un (s)ds (εj − ε̄n ).
n j=1
0
Since an → 0, we obtain for the first term, as n → ∞,
2
2
Z 1
Z 1
Z 1
n a2n X
Un (j/n) −
Un (s)ds = a2n
Un (t) −
Un (s)ds dt = o(1)
n j=1
0
0
0
a.s.,
that is, independent of the rate of convergence to zero, the first part will not contribute
to the limit at all. For the second term, it holds by the strong law of large numbers,
n
1X
(εj − ε̄n )2 −→ σ 2
n j=1
a.s.
(n → ∞).
Finally, the third term is negligible, which can be seen by an application of Chebyshev’s
inequality.
On the other hand, the different limits for the suitably normalized Qn follow using the
fact that, for all n ≥ 1,

⌊nt⌋
X

j=1


D
Xj : t ∈ [0, 1] = nan
Z
0
t
P(s)ds +
√
nτ W (t) + rn (t) : t ∈ [0, 1] ,
where the remainder term satisfies the relation
√ sup rn (t) = oP max{nan , n}
(n → ∞).
0≤t≤1
Hence, according to the asymptotic√order of an , a leading term exists, which in turn
determines the limit, or—in case of nan → c—the limit is a mixture of both sources of
randomness. This completes the proof.
2
15
3.3
Proof of Theorem 2.4
In order to prove Theorem 2.4, we first provide a result concerning the order of the
difference between the sample variance s2n from (1.2) and its limit obtained during the
course of the proof of Theorem 2.1.
Lemma 3.3 Let (1.4), (1.5) and (2.1)–(2.3) be satisfied, and assume that E|η1 |ν < ∞
for some ν ≥ 2. Then
P
(
2
sn
"
2
− σ +
Z
1
Pn (t) −
0
Z
1
0
Pn (s)ds
2
#
dt −1/3
> Cn
)
≤ Cn−1/3
for all n,
where {Pn (t) : t ∈ [0, 1]} is the sequence of compound Poisson processes, which is defined
in Lemma 3.2, and C is a constant.
Proof. Write
n
1X
Un (j/n) −
n j=1
Z
1
n
1X
=
Un (j/n) −
n j=1
Z
1
s2n =
0
Un (s)ds
0
n
2X
+
Un (j/n) −
n j=1
Un (s)ds + εj − ε̄n
Z
1
0
2
2
(3.10)
n
1X
(εj − ε̄n )2
+
n j=1
Un (s)ds (εj − ε̄n ).
We shall estimate each term of the sum on the right–hand side of the latter equations.
Starting with the first term, it is easy to see that
n
1X
Un (j/n) −
n j=1
Z
0
1
Un (s)ds
2
=
Z
0
1
Un (u) −
Z
1
0
Un (s)ds
2
du.
Using the approxmiation with a sequence of compound Poisson processes {Pn (t) : t ∈
[0, 1]} established in Lemma 3.2, we can estimate
Z 2
2
Z 1
Z 1
Z 1
1
Un (u) −
Un (s)ds du −
Pn (u) −
Pn (s)ds du
0
0
0
0
≤
Z " 2
Z 1
1
Un (u) −
Un (s)ds
0
0
− Pn (u) −
16
Z
0
1
Pn (s)ds
2 #
du
(3.11)
Z
+2 ≤ 2
"
1
0
Un (u) −
1
0
Un (s)ds − Pn (u) −
× Pn (u) −
sup
0<u<1
Z
|Un (u)| + "
sup
0<u<1
Z
1
0
1
Z
0
1
Pn (s)ds
Un (s)ds +
0
×
+2
Z
du
sup
0<u<1
|Un (u) − Pn (u)| du
Z
|Pn (u)| + 1
0
# Z
Pn (s)ds
1
0
Z
1
0
Pn (s)ds
Z
|Pn (u)| + 1
0
#
Pn (s)ds
|Un (u) − Pn (u)| du.
On applying (3.4) we obtain, for any x ≥ 0,


n (1)

NX
c
P sup |Un (u)| > x ≤ + P
|ηi | > x


n
0<u<1
i=1
(
)
(3.12)
with some positive constant c. Clearly, also
Z
1
0
Un (s)ds
≤ sup |Un (u)|.
0<u<1
Recall that ηi+ and ηi− are the positive part, respectively, negative part of ηi and that
tn = ⌊nt⌋/n. In view of (3.4) we estimate
Z
1
0
|Un (u) − Pn (u)|du
Z
≤
1
0


Nn (tn )
Nn (t)
X
i=1
ηi+
−
X
i=1

ηi+  dt
+
Using the method of Lemma 3.2 we have
Z
0
1
Z
1
0


Nn (tn )
Nn (t)
X
i=1
ηi−
−
X
i=1

ηi−  dt.
N (1)
2 X
|Un (u) − Pn (u)|du ≤
|ηi |,
n i=1
where the νth moment of the latter sum is finite by assumption, i.e.

N (1)
E
X
i=1
ν
|ηi | < ∞
(3.13)
17
[cf. the proof of Lemma 3.2]. Hence, for any y > 0,
P
Z
1
0
|Un (u) − Pn (u)|du > y ≤ C(ny)−ν ,
(3.14)
where C is a positive constant. Choosing y = n−(1+ν)/(2+ν) in (3.14), x = n1/(2+ν) in
(3.12), we conclude from (3.11) that
P
(Z 2
2
Z 1
Z 1
Z 1
1
Un (u) −
Un (s) ds −
Pn (u) −
Pn (s)ds du
0
0
0
0
−ν/(2+ν)
>n
)
≤ Cn−ν/(2+ν) .
Next, we turn our attention to the second term of the right–hand side in (3.10). Here, an
application of Markov’s inequality and Theorem 2.10 of Petrov (1995, p. 62) yields
P
 n
 1 X
(ε2
j
 n j=1
2 − σ )
> n−µ/(4+2µ)

and similarly, for all x > 0,
P
(


≤ Cn−µ/(4+2µ)
n
1 X
−µ −µ/2
ε
n
.
i > x ≤ Cx
n i=1
)
(3.15)
Therein, the parameter µ > 4 is determined by (2.3). Thus we have, on combining the
latter two inequalities,
P
(
n h
i
1 X
(εi − ε̄n )2 − σ 2 > 2n−µ/(4+2µ)
n i=1
)
≤ Cn−µ/(4+2µ) .
Ultimately, rewrite the third term in (3.10) as
n
1X
Un (j/n) −
n j=1
Z
1
0
Un (s)ds (εj − ε̄n ) =
Then it follows from (2.1) that

Var 
n
X
j=1

εj Un (j/n)
= σ
2
n
X
EUn2 (j/n)
j=1
18
n
1X
Un (j/n)εj − ε̄n
n j=1
Z
0
1
Un (s)ds.
n
X
X
= σ2
Eci,n cl,n ηi ηl
j=1 1≤i,l≤j


j
X
1
1

+
≤ C
2
j=1 1≤i,l≤j n
i=1 n
n
X
X
≤ Cn.
By Chebyshev’s inequality, we obtain
P
 n
 1 X
Un (j/n)εj n j=1


>x

for all x > 0, and therefore
P
 n
 1 X
Un (j/n)εj n j=1
≤
≥ n−1/3
Cn
(nx)2



≤ Cn−1/3 .
Using (3.12) and (3.13), we arrive at
Z
P 0
1
Un (s)ds > y ≤
C
C
+ ν
n
y
(3.16)
for all y > 0. Applying (3.15) and (3.16) with
x = n−µ(1+ν)/(2[ν+µ(1+ν)])
and
y = nµ/(2[ν+µ(1+ν)]) ,
respectively, we finally conclude
P
Z
ε̄n
1
0
Un (s)ds
−µν/(2[ν+µ(1+ν)])
>n
≤ Cn−µν/(2[ν+µ(1+ν)]) ,
finishing the proof.
2
Recall that, for t ∈ [0, 1], Pn,0 (t) = Pn (t) − tPn (1), where the compound Poisson
processes {Pn (t) : t ∈ [0, 1]} are defined in Lemma 3.2. We are now able to prove
Theorem 2.4.
Proof of Theorem 2.4. First we note that
1
n
max
1≤k≤n
k
X
j=1
Un (j/n) − sup
Z
0<u<1 0
u
Un (s)ds
19
Z
u
1
≤
sup Un (s)ds
n 0<u<1 0
and therefore
1
n
Clearly,


n
kX
Un (j/n) − sup
max  Un (j/n) −
1≤k≤n
n
0<u<1
j=1
j=1
k
X
2
≤
sup n 0<u<1
u
Z
sup
0<u<1
Z
u
0
0
Z
u
0
Un (s)ds − u
Z
0
Un (s)ds .
Pn (t)dt
1
Un (s)ds Nn (1)
≤
X
i=1
|ηi |,
where, for each n, Nn (1) are Poisson random variables with mean 1. Markov’s inequality
implies that
P

n (1)
 1 NX
n
i=1
|ηi | > n−ν/(1+ν)



≤ n1−ν/(1+ν)
ν

E
ν
Nn (1)
X
i=1
|ηi | ,
where the latter expectation is finite (cf. the proof of Lemma 3.2). So, by Lemma 3.2,
P
(
1
sup n 0<u<1
Z
u
0
Un (s)ds
−ν/(1+ν)
> c1 n
)
≤ c2 n−ν/(1+ν)
(3.17)
with some constants c1 and c2 . Using Lemma 3.2 once more, we obtain
P

 1
 n

max 
1≤k≤n
k
X
j=1
Un (j/n) −
≤ c2 n−ν/(1+ν) .

n
X
k
Un (j/n) − sup
n j=1
0<u<1
Next observe that
k
X
max εi
1≤k≤n i=1
k
n
X
kX
−
εi ≤ 2 max εi .
1≤k≤n n i=1 i=1
Theorem 2.4 of Petrov (1995, p. 52) yields
P
(
k
X
max εi 1≤k≤n i=1
)
> x ≤ 4P
( n
X εi i=1
>x−
20
√
)
2σ 2 n
Z
0
u
P0,n (t)dt
≥ c1 n−ν/(1+ν)



(3.18)
for all x >
P
(
√
2σ 2 n. Hence
k
X
1
max εi > n−µ/(2(1+µ))
n 1≤k≤n i=1 ≤ 4P
( n
X εi )
1−µ/(2(1+µ))
>n
i=1
1−µ/(2(1+µ))
≤ 4 n
−
√
2σ 2 n
(3.19)
−
−µ
≤ cn−µ/(2(1+µ)) ,
√
)
2σ 2 n
µ
n
X
E εi i=1
where the latter inequality is derived from the fact that, by Theorem 2.10 of Petrov (1995,
p. 62), the expected value is of order O(nµ/2 ). It follows from Lemma 3.3 in combination
with Lemma 3.2 and (3.17) that
P
(
s2n
σ2
<
2
)
≤ Cn−1/3 .
(3.20)
Also,
P
(
sup P0,n (t) − inf P0,n (t)
0<u<1
0<u<1
≤ P

n (1)
NX

i=1
)
(
)
> x ≤ P 4 sup |Pn,0 (t)| > x
0<u<1
(3.21)

x
|ηi| >
≤ Cx−ν .
4
Applying (3.18), (3.19) and (3.20), we get that
P
(
1
sn
R
n
n
−
!
sup P0,n (t) − inf P0,n (t) 0<u<1
0<u<1
≤ C n−ν/(1+ν) + n−µ/(2+µ) + n−1/3
≤ Cn−1/3 .
> C n−ν/(1+ν) + n−µ/(2+µ)
Now, for any y > 0,
P
(
1
sup P0,n (t) − inf P0,n (t) sn
0<u<1
0<u<1
= P
1
− > y
s̄n
(
sup P0,n (t) − inf P0,n (t)
0<u<1
0<u<1
)
|s2n − s̄2n |
>y
sn s̄n (sn + s̄n )
21
)
)
where
s̄n = σ 2 +
Z
1
0
Pn (t) −
Z
0
1
Pn (s)ds
2
dt.
Applying (3.21) with x = n1/(3(1+ν)) , (3.20) and Lemma 3.3 to the latter relation with
y = Cn−ν/(3(1+ν)) , we finally arrive at
P
(
sup P0,n (t) − inf P0,n (t)
0<u<1
0<u<1
s2n − s̄2n
> Cn−ν/(3(1+ν))
sn s̄n (sn + s̄n )
)
≤ Cn−ν/(3(1+ν)) ,
which finishes the proof.
2
References
[1] Bhattacharya, R.N., Gupta, V.K., and Waymire, E. (1983). The Hurst effect under
trends. Journal of Applied Probability 20, 649–662.
[2] Billingsley, P. (1968). Convergence of Probability Measures. Wiley, New York.
[3] Diebold, F., and Inoue, A. (2001). Long memory and regime switching. Journal of
Econometrics 105, 131–159.
[4] Giraitis, L., Kokoszka, P., and Leipus, R. (2001). Testing for long memory in the
presence of a general trend. Journal of Applied Probability 38, 1033–1054.
[5] Giraitis, L., Kokoszka, P., Leipus, R., and Teyssiere, G. (2003). Rescaled variance
and related tests for long memory in volatility and levels. Journal of Econometrics
112, 265–294.
[6] Granger, C.W.J., and Hyung, N. (2004). Occasional structural breaks and long memory with an application to the S&P 500 absolute stock returns. Journal of Empirical
Finance 11, 399–421.
[7] Hardy, G. H., Littlewood, J. E., and Pólya, G. (1952). Inequalities (2nd ed.). Cambridge University Press.
[8] Hurst, H. (1951). Long term storage capacity of reservoirs. Transactions of the American Society of Civil Engineers 116, 770–799.
22
[9] Kurtz, T.G., and Protter, P. (1991). Weak limit theorems for stochastic integrals and
stochastic differential equations. Annals of Probability 19, 1035–1070.
[10] Leipus, R., Viano, M.–C. (2003). Long memory and stochastic trend. Statistics and
Probability Letters 61, 177–190.
[11] Major, P. (1990). A note on the approximation of the uniform empirical process.
Annals of Probability 18, 129–139.
[12] Perron, P., and Qu, Z. (2004). An analytical evaluation of the log–periodogram estimate in the presence of level shift and its implications for stock returns volatility.
Preprint, Boston University.
[13] Petrov, V.V. (1995). Limit Theorems of Probability Theory. Clarendon, Oxford.
23
Download