Near–Integrated Random Coefficient Autoregressive Time Series

advertisement
Near–Integrated
Random Coefficient Autoregressive Time Series
Alexander Aue
1
Abstract: We determine the limiting behavior of near–integrated first–order random
coefficient autoregressive RCA(1) time series. It is shown that the asymptotics of the finite
dimensional distributions crucially depends on how the critical value 1 is approached,
which determines whether the process is near–stationary, has a unit–root or is mildly
explosive.
In a second part, we derive the limit distribution of the serial correlation coefficient in
the near–stationary and the mildly explosive setting under very general conditions on the
parameters. The results obtained are in accordance with those available for first–order
autoregressive time series and can hence serve as addition to existing literature in the
area.
AMS 2000 Subject Classification: Primary 62M10, Secondary 91B84.
Keywords and Phrases: near–integrated time series; random coefficient time series;
serial correlation coefficient; near–stationarity; mildly explosive models; unit–root models
1
Department of Mathematics, University of Utah, 155 South 1440 East, Salt Lake City, UT 84112–
0090, USA, email: aue@math.utah.edu
Partially supported by NATO grant PST.EAP.CLG 980599 and NSF–OTKA grant INT–0223262
1
Introduction
Let {Xt } be a first–order autoregressive AR(1) process, i.e. the solution of the recurrence
equations
Xt = ϕXt−1 + εt ,
−∞ < t < ∞.
(1.1)
In case the noise {εt } is an independent, identically distributed (iid) sequence, it is well–
known that the process is strictly stationary if ϕ < 1 and explosive if ϕ > 1 (cf. for
instance [6]). {Xt } is said to have a unit–root if ϕ = 1. Then, the process resembles the
behavior of a random walk. If ϕ ≈ 1 and ϕ → 1 with increasing sample size, {Xt } is
referred to as near–integrated.
A number of papers has been devoted to examining the asymptotic structure of {Xt }
if the coefficient ϕ is reparameterized to
ϕn = 1 −
β
,
nα
α ∈ [0, 1],
n ≥ 1,
(1.2)
with some fixed real number β, and if model (1.1) is transformed to
Xt (n) = ϕn Xt−1 (n) + εt ,
t = 1, . . . , n,
(1.3)
with initial value X0 (n) = 0 for all n. For instance, Chan and Wei (cf. [7], [8]) studied
the near–stationary AR(1) process (i.e. β > 0) as well as the mildly explosive case (i.e.
β < 0) if α = 1 and the noise {εt } constitutes a sequence of martingale differences.
An extensive analysis for near–integrated time series was carried out by Phillips [16].
His more general multivariate approach, which includes multiple times series in which
some component series have unit–roots, some are stationary, while others are explosive,
is in one dimension asymptotically equivalent to the setting introduced by Chan and
Wei in [7] and [8]. Related work in the above framework was done by Ling and Li [13],
who considered unstable autoregressive moving average ARMA processes with general
autoregressive heteroscedastic GARCH errors. In recent work, Phillips and Magdalinos
[17] established limit theorems in the general parameter case ϕn → 1 for what they call
moderate deviations from a unit–root.
The asymptotic theory obtained in all these papers has consequences for testing procedures for the presence of unit–roots. Usually these tests have a low power because the
behavior of the test statistic under consideration does not crucially differ between the
(null) hypothesis of a unit–root and the alternatives of near–stationary or mildly explosive processes. Unit–root tests play a fundamental role in economometrics, cf. for example
[16] and [14] and the references therein.
2
Motivated by results for GARCH(1,1) sequences obtained by Berkes, Horváth and
Kokoszka [3], we will study the finite dimensional distributions of so–called first–order
random coefficient autoregressive RCA(1) time series, which are defined by the stochastic
recurrence equations in (2.1). Originally, these processes have been applied to model
certain dynamical systems (cf. [15]), but are also used in fincance (see [18]). Surveys
on basic properties of this model, such as stationarity and existence of moments, can be
found, for instance, in [15] and [2].
Our results differ from the above approach in two directions. On one hand, the parameter ϕn in (1.3) is allowed to be random, which is enabled by changing it to the
corresponding quantity ϕn + bt (n) in (2.1) involving an independent sequence {bt (n)} of
random variables having finite second moments. On the other hand, the parameterization (1.2) can be changed. The limit results of Theorems 2.1 (near–stationary case), 2.2
(unit–root) and 2.3 (mildly explosive case) are obtained under the asymptotic rates given
in (2.6). The proofs of these theorems are based on an additive representation of the
RCA(1) process.
A great variety of the afore mentioned unit–root tests are based on the least squares
estimator
!
ϕ̂n =
n
X
t=1
2
(n)
Xt−1
−1 n
X
Xt−1 (n)Xt (n),
t=1
n ≥ 2,
of ϕn . Its limiting behavior has been examined thoroughly. However, most of the existing
literature is devoted to the special case α = 1 in (1.2). See, e.g., Chan and Wei [7],
[8]. For the general convergence
ϕn → 1, it turns out that a Gaussian limit exists under
q
a normalization of order n/|ϕn − 1| if the AR(1) process is near–stationary. On the
other hand, if Xt (n) is in the explosive regime, i.e. ϕn > 1, the limit will no longer be
Gaussian but a Cauchy random variable, while the normalization sequence has now the
exponential order en(ϕn −1) /(ϕn − 1). These results can be found in [17]. We will show
that similar statements hold true also in the random coefficient setting provided in (2.1).
Theorem 2.4 below deals with mild explosion, Theorem 2.5 gives the corresponding result
for near–stationarity.
Ultimately, it is worthwhile mentioning that, since all calculations can be accomplished
also if bt (n) ≡ 0, all results remain valid for AR(1) time series. [This applies primarily
to the finite dimensional distribution results.] So, the paper can serve as complementary
work to the existing results cited above.
The paper is organized as follows. In Section 2, we present the RCA(1) model and state
and discuss the main results. Proofs are given in Section 3.
3
2
Main Results
Consider the sequence of random coefficient autoregressive RCA(1) models
Xt (n) = [ϕn + bt (n)]Xt−1 (n) + εt ,
t = 1, . . . , n,
(2.1)
with initial value X0 (n) = 0 for all n. Therein, the random vectors
{(γt , εt )} are independent, identically distributed with
E(γ1 , ε1 ) = (0, 0) and Eγ12 = 1, Eε21 = σ 2
(2.2)
and, for all n,
bt (n) = ωn2 γt ,
t = 1, . . . , n.
(2.3)
We shall assume that
ϕn → 1 and ωn2 → 0
(n → ∞),
(2.4)
which implies immediately that here ϕ2n + ωn2 → 1, i.e. the parameter sum converges
to the critical value determining whether or not the process allows for (nonanticipative)
second–order stationary solutions. We will find out that the rate of convergence as well
as the limit distribution of the sequence of RCA(1) processes crucially depends on how
the parameter ϕn approaches 1. Hence, we define
ϕ̃n = ϕn − 1,
n ≥ 1,
(2.5)
and differ between the three cases ϕ̃n > 0, ϕ̃n = 0 and ϕ̃n < 0 for all n. Moreover, we
need the following rates
√
nϕ̃n → 0
(n → ∞)
(2.6)
n|ϕ̃n | → ∞ but
and
nωn → ∞ but
√
nωn → 0
(n → ∞).
(2.7)
Since we are actually dealing with a sequence of models, the point of view is the one of
finite dimensional distributions and we study the asymptotic behavior of the vectors
h
i
X⌊nκm ⌋ : m = 1, . . . , N ,
where N is a fixed integer, 0 < κ1 < . . . < κN ≤ 1 and ⌊·⌋ the integer part.
At first, we investigate the case ϕ̃n > 0 for all n. Hence, we study a sequence of
nonstationary models. The asymptotics is given in the following theorem.
4
Theorem 2.1 (mildly explosive case) Let {Xt (n)} be an RCA(1) time series satisfying model (2.1) with ϕ̃n > 0 for all n. If (2.2)–(2.7) hold and if
E|ε1|2+δ < ∞
the random variables
for some δ > 2,
"√
2ϕ̃n
X⌊nκm ⌋ : m = 1, . . . , N
enϕ̃n σ
(2.8)
#
are asymptotically (n → ∞) independent standard normal.
If ϕ̃n = 0 for all positive integers n, a different limiting distribution under a different
normalization is obtained. Since the random part bt (n) of the coefficients dies out for
n → ∞ by assumption (2.4), it is not surprising to see a resemblance of the random walk
here.
Theorem 2.2 (unit–root case) Let {Xt (n)} be an RCA(1) time series satisfying model
(2.1) with ϕ̃n = 0 for all n. If (2.2)–(2.5), (2.7) hold and if
E|ε1|2+δ < ∞
for some δ > 0,
(2.9)
then,
"
#
1
D
√ X⌊nκm ⌋ : m = 1, . . . , N −→ [W (κm ) : m = 1, . . . , N]
nσ
(n → ∞),
where {W (s) : s ∈ [0, 1]} denotes a standard Wiener process.
Finally, we examine the stationary
case,
√
√ i.e. ϕ̃n < 0 for all n. While the order of the
and unit–root setting,
oscillations is exponential enϕ̃n / ϕ̃n and n in the near–explosive
q
respectively, fluctuations increase with the smaller rate |ϕ̃| in case of near–stationary
models.
Theorem 2.3 (near–stationary case) Let {Xt (n)} be an RCA(1) time series satisfying model (2.1) with ϕ̃n < 0 for all n. If (2.2)–(2.7) hold and if
E|ε1|2+δ < ∞
the random variables
q
for some δ > 0,
−1
2|ϕ̃n |σ X⌊nκm ⌋ : m = 1, . . . , N
are asymptotically (n → ∞) independent standard normal.
5
(2.10)
The proofs of all three theorems are based on a common asymptotic additive representation of {Xt (n)}, which is given in Theorem 3.1 below. Using this result as a starting
point, the respective asymptotic distributions are obtained by applying and exploiting
special characteristics of the individual cases to filter out the dominating part of the
decomposition. Details are deferred to the next chapter.
We conclude the paragraph with the remark that Theorems 2.1–2.3 hold true also if
the noise sequence {εt } is allowed to depend on n, i.e. εt = εt (n) = σn2 δt , where {δt }
are independent, identically distributed and centered random variables with Eδ12 = 1 and
σn2 → σ 2 > 0 as n → ∞.
The remainder of the section is devoted to examining the serial correlation coefficient.
From now on, we assume that the initial values X0 (n) = ε0 are constant. Denote by
ϕ̂n =
n
X
t=1
2
Xt−1
(n)
!−1
n
X
Xt−1 (n)Xt (n),
t=1
n ≥ 1,
(2.11)
the (conditional) least squares estimator for ϕn . We say, that a random variable ξ is
standard Cauchy if its density is given by
1
f (x) =
,
−∞ < x < ∞.
π(1 + x2 )
We start with giving the asymptotics for the serial autocorrelation coefficient ϕ̂n − ϕn
if ϕn is larger than 1, that is, approaches 1 from the explosive region. Under a suitable
normalization this limit is standard Cauchy.
Theorem 2.4 (mild explosion) Let {Xt (n)} be an RCA(1) time series satisfying model
(2.1) with ϕ̃n > 0 for all n. If (2.2)–(2.7) hold, then
enϕ̃n
D
(ϕ̂n − ϕn ) −→ ξ
2ϕ̃n
(n → ∞),
where ϕ̂n is defined in (2.11) and ξ denotes a standard Cauchy random variable.
In contrast to the Cauchy limit for moderate deviations in the explosive direction, the
near–stationary case provides a Gaussian limit theorem.
Theorem 2.5 (near–stationarity) Let {Xt (n)} be an RCA(1) time series satisfying
model (2.1) with ϕ̃n < 0 for all n. If (2.2)–(2.7) hold, then
s
n
D
(ϕ̂n − ϕn ) −→ ζ
2|ϕ̃n |
(n → ∞),
where ϕ̂n is defined in (2.11) and ζ denotes a standard normal random variable.
6
The results are in accordance with results existing for near–integrated first–order autoregressive time series. The most recent work has been done in Phillips and Magdalinos
[17] for an equivalent general parameter case. Previous results in the literature were
mainly concerned with the special cases ϕn = 1 + β/nα with α ∈ [0, 1] and the parameter
β determininig whether the setting is mildly explosive (i.e. β > 0) or near–stationary (i.e.
β < 0). Using this particular form of ϕn , one can gain deeper insight on how to bridge
gaps between limit results for near–integration and unit–roots on one hand, and between
near–integration and stationarity, respectively, explosion on the other. For a detailed
discussion, we refer to [17].
More general limit theorems hold true also if the noise {εt } is in the domain of attraction
of a stable law. Corresponding results are due to Aue and Horváth [1] for first–order
autoregression. Adaptations of the proofs will as well work in the present model with
random coefficients.
The proof of Theorem 2.4 falls back on the representation of Xt (n) given in Theorem
3.1, while Theorem 2.5 can be established exploiting a result in [17]. Details are given in
Sections 3.5 and 3.6 below.
3
Proofs
Chapter 3 is divided into six sections. In the first, we derive the representation of {Xt (n)}
which is the basic tool for the proofs of Theorems 2.1–2.5 in the subsequent sections.
To simplify notation, we suppress the dependence on n in the following. Hence, we
abbreviate
Xt = Xt (n)
and
bt = bt (n)
as well as
ϕ = ϕn , ϕ̃ = ϕ̃n
and
ω 2 = ωn2 .
But nonetheless, we should always keep in mind, that we are actually dealing with a
triangular array of random variables. Throughout the chapter, we will use
t = ⌊κn⌋
3.1
with some 0 < κ < 1.
An asymptotic representation of RCA(1) time series
We shall need the following small lemma.
7
(3.1)
Lemma 3.1 Under the assumptions (2.2)–(2.7) and (3.1), we have
max |ϕ − 1 + bt−i | = oP (1)
(n → ∞).
0≤i<t
Proof. We can estimate
max |ϕ − 1 + bt−i | ≤ |ϕ̃| + ω max |γt−i |.
0≤i<t
(3.2)
0≤i<t
Using Corollary 3 of [9], p. 90, we see that
max |γt−i | = max |γj | = O
0≤i<t
1≤j≤t
√ t
a.s.
as n → ∞. So the claim follows from (3.2) after an application of (2.6) and (2.7).
2
A repeated application of the defining equations (2.1) yields
Xt = X0
Y
(ϕ + bt−i ) +
0≤i<t
X
Y
(ϕ + bt−i )εt−j .
0≤j<t 0≤i<j
The next theorem gives the promised additive decomposition by examining the products
in the previous equation.
Theorem 3.1 Let {Xt } follow model (2.1). If (2.2)–(2.7) and (3.1) are satisfied, then

Xt = X0 etϕ̃ 1 +
+
X
0≤j<t
where, as n → ∞,
X
0≤j<t

ej ϕ̃ 1 +

(1)
bt−j + Rt 
X
0≤i<j
(1)
(2)

(3)
|Rt | = OP t(ϕ̃2 + ω 2 ) ,
(2)
max |Rt,j | = OP (tω 2 ),
0≤j<t
(2)
|Rt,j |
= OP (ω 2 ),
max
0≤j<t j log log j
1 (3)
max |Rt,j | = OP (ϕ̃2 + ω 2).
0≤j<t j
8
bt−i + Rt,j  1 + Rt,j εt−j ,
Proof. The proof is given in three steps. In the first step, we obtain an additive repreQj−1
sentation for the product i=0
(ϕ + bt−i ), which is refined in a second and third step.
(i) Define the events
1
,
An = max |ϕ̃ + ωγt−i | ≤
0≤i<t
2
n ≥ 1.
From Lemma 3.1, we obtain [recall that t = ⌊κn⌋ by (3.1)]
lim P (An ) = 1.
n→∞
A Taylor expansion yields | log(1 + x) − x| ≤ 2x2 for |x| ≤ 21 . Hence for all j < t, on An ,
X
X
log(ϕ + bt−i ) −
(ϕ̃ + ωγt−i )
0≤i<j
0≤i<j
X
≤ 2
(ϕ̃ + ωγt−i)2
0≤i<j
≤ 4j ϕ̃2 + 4ω 2
X
2
γt−i
.
0≤i<j
Using the strong law of large numbers and the fact that by definition Eb21 = ω 2, we can
estimate the sum of the γi2 ’s as
1
1
D
2
| = max |γ12 + . . . + γj2 | = OP (1)
max |γt2 + . . . + γt−i
1≤i≤t i
1≤i≤t i
(n → ∞).
So, we get for the product of interest
Y
0≤i<j

(ϕ + bt−i ) = exp(j ϕ̃) exp ω

= exp(j ϕ̃) exp 
X
0≤i<j
X
0≤i<j
as n → ∞, where



γt−i  exp(St,j )

(3)
bt−i  1 + Rt,j

X
1
1
max |St,j | = OP  max 4j ϕ̃2 + 4ω 2
γt−i 
0≤j<t j
0≤j<t j
0≤i<j
= OP ϕ̃2 + ω 2
9
(n → ∞).
(3.3)
(3)
Hence, we can choose Rt,j such that the corresponding rate given in Theorem 3.1 is
satisfied, namely
1 (3)
(n → ∞).
max |Rt,j | = OP (ϕ̃2 + ω 2 )
0≤j<t j
(ii) Since {γi} is a sequence of independent, identically distributed random variables with
Eγ12 = 1, it follows from the weak convergence of partial sums to Brownian motion that
X
γt−i max 0≤j<t 0≤i<j
= OP
√ t
(n → ∞).
Hence, from (2.7) and (3.1),
X
max bt−i 0≤j<t 0≤i<j
=
Define the events
X
max ω
γt−i 0≤j<t 0≤i<j
Bn =
Then,
√
= OP ( tω) = oP (1)
X
bt−i max 0≤j<t 0≤i<j



1
≤
,
2
(n → ∞).
(3.4)
n ≥ 1.
lim P (Bn ) = 1.
n→∞
Using again a Taylor expansion argument, | exp(x) − (1 + x)| ≤
for all j < t, on Bn ,
with

 

X
X
exp 



b
−
1
+
b
t−i
t−i 0≤i<j
0≤i<j

max 
0≤j<t
X
0≤i<j
2
bt−i  = max ω 2 
0≤j<t
as n → ∞ using (3.4). Furthermore,


2
X
0≤i<j
√
e 2
x
2
for |x| ≤ 21 . Hence
2
√ 
e X
bt−i 
≤
2 0≤i<j
2
γt−i  = OP (tω 2 )

2
X
X
ω2
1


bt−i  = max
γt−i  = OP (ω 2 )
max
0≤j<t j log log j
0≤j<t j log log j
0≤i<j
0≤i<j
10
as n → ∞, where the law of the iterated logarithm has been applied to the sequence {γi }
to obtain the desired rate. Inserting these results into the right–hand side of (3.3), we
arrive at


Y
0≤i<j
(2)
(ϕ + bt−i ) = exp(j ϕ̃) 1 +
X
0≤i<j
(3)
(2)
(3)
bt−i + Rt,j  1 + Rt,j ,
(3.5)
where Rt,j and Rt,j satisfy the corresponding rates given in Theorem 3.1.
(iii) From the first part of the proof,
max
0≤j<t
(3)
|Rt,j |
= OP
max |St,j | = OP t(ϕ̃2 + ω 2 ) = oP (1)
0≤j<t
(n → ∞).
Thus, we obtain from (3.5)
Y
0≤i<t

(ϕ + bt−i ) = exp(tϕ̃) 1 +
X
0≤i<t

bt−i + OP (tω 2 ) 1 + OP t(ϕ̃2 + ω 2 )
.
Applying the central limit theorem to the partial sums of {γi } and taking assumption
(2.7) into account,

1 +
X
0≤i<t


= 1 + ω
= 1+
bt−i + OP (tω 2 ) 1 + OP t(ϕ̃2 + ω 2 )
X
0≤i<t
X
0≤i<t

γt−i + OP (tω 2 ) 1 + OP t(ϕ̃2 + ω 2 )
bt−i + OP t(ϕ̃2 + ω 2 )
as n → ∞. So, finally
Y
0≤i<t
(1)
where Rt

(ϕ + bt−i ) = exp(tϕ̃) 1 +
X
bt−i +
0≤i<t
satisfies the rate stated in Theorem 3.1.
11

(1)
Rt 
(n → ∞),
2
3.2
Proof of Theorem 2.1
The following lemma will be frequently used throughout the proofs of Theorems 2.1–2.3.
It has been established in [3] within the context of near–integrated GARCH(1,1) sequences
and we state it without proof. Let {an } and {bn } be two sequences of real numbers. We
say that an ∼ bn if limn→∞ an b−1
n = 1.
Lemma 3.2 For any ν ≥ 0 it holds,
X
0≤j<t
where Γ(t) =
R∞
0
j ν ej ϕ̃ ∼
Γ(ν + 1)
,
|ϕ̃|ν+1
e−x xt−1 dx denotes the Gamma function.
Proof: See [3], Lemma 4.1.
2
Theorem 2.1 is proved by an application of Theorem 3.1. Rewriting the decomposition
yields
tϕ̃

Xt = X0 e 1 +
+
X
0≤j<t
+
X
X
bt−j +
0≤j<t

ej ϕ̃ 1 +
j ϕ̃
e

(1)
Rt 
(2)
X
0≤i<j
(2)
Rt,j εt−j

(3)
bt−i + Rt,j  Rt,j εt−j
0≤j<t
+
X
0≤j<t
(1)

ej ϕ̃ 1 +
(2)
X
0≤i<j
(3)

bt−i  εt−j
(4)
= Xt + Xt + Xt + Xt
(4)
(3.6)
It turns out, that the dominating part of Xt is Xt . Its limiting behavior is derived in
(1)
(2)
Lemma 3.3. In the subsequent Lemmas 3.4–3.6 it is proved that the terms Xt , Xt and
(3)
Xt are negligible. That is, Theorem 2.1 follows readily from these auxiliary results.
Lemma 3.3 Under the assumptions of Theorem 2.1,
q
−nϕ̃
lim ϕ̃e
n→∞
(4) D
Xt = n→∞
lim
q
2ϕ̃σ
Z
0
t
e(x−n) d[W (t) − W (x)],
where {W (t) : t ≥ 0} denotes a standard Wiener process.
12
√
Proof: From the central limit theorem, γt + γt−1 + . . . + γ1 = OP ( t) as n → ∞. Hence,
X
X
bt−j = ω
0≤j<t
γt−j = oP (1)
(n → ∞),
0≤j<t
(3.7)
by (2.7). This yiels
(4)
Xt
X
=
ej ϕ̃ εt−j (1 + oP (1))
(n → ∞)
0≤j<t
and we have to study the sum on the right–hand side of the latter equation. Write
X
0≤j<t
ej ϕ̃ εt−j =
Z
0
t
exϕ̃ d[S(t) − S(x)] = S(t) −
Z
0
t
[S(t) − S(x)]ϕ̃exϕ̃ dx
with
S(0) = 0,
S(x) =
X
εj ,
x > 0.
0≤j<x
By (2.8) and the Komlós, Major and Tusnády approximations (see [11] and [12]), there
is a Wiener process {W (t) : t ≥ 0} such that
S(x) − σW (x) = o x1/(2+δ)
a.s.
as x → ∞, where δ > 2. Hence,
√
Z κn
ϕ̃
xϕ̃
sup
([S(x)
−
S(κn)]
−
[W
(x)
−
W
(κn)])
ϕ̃e
dx
n
ϕ̃
e 0≤κ≤1 0
√ 1/(2+δ)
Z κn
ϕ̃n
xϕ̃
= OP (1)
ϕ̃e dx
sup n
ϕ̃
e
0≤κ≤1 0
q
= OP (1) ϕ̃n1/(2+δ)
= oP (1)
as n → ∞ by (2.6), since δ > 2. Also,
√ 1/(2+δ)
√
ϕ̃
ϕ̃n
sup
|S(κn)
−
σW
(κn)|
=
O
(1)
= oP (1)
P
enϕ̃ 0≤κ≤1
enϕ̃
(n → ∞).
Using partial integration in the opposite direction for {W (t) : t ≥ 0}, we see that
q
2ϕ̃σ
has the same limit distribution as
Z
t
0
√
e(x−n)ϕ̃ d[W (t) − W (x)]
(4)
ϕ̃e−nϕ̃ Xt .
13
2
Lemma 3.4 Under the assumptions of Theorem 2.1,
√ −tϕ̃
ϕ̃e
(1)
Xt = oP (1)
(n → ∞),
σ
(1)
where Rt
satisfies the corresponding rate in Theorem 3.1.
Proof: Using Theorem 3.1, (2.6) and (2.7), we obtain that
(1)
Rt = OP t(ϕ̃2 + ω 2) = oP (1)
(n → ∞).
Moreover, bt + bt−1 + . . . + b1 = oP (1) by (3.7) and clearly also
√
√ −tϕ̃
ϕ̃e
ϕ̃X0
tϕ̃
X0 e =
= oP (1)
(n → ∞)
σ
σ
by the fact that ϕ̃ → 0. Hence, the proof is complete.
2
Lemma 3.5 Under the assumptions of Theorem 2.1,
√ −tϕ̃
ϕ̃e
(2)
Xt = oP (1)
(n → ∞),
σ
(2)
(3)
where Rt,j and Rt,j satisfy the corresponding rates in Theorem 3.1.
Proof: Note that by Theorem 3.1 and (2.7)
(2)
max |Rt,j | = OP (tω 2 ) = oP (1)
(n → ∞).
0≤j<t
Thus, by (3.7) and Theorem 3.1 it holds,
√ −tϕ̃
√ −tϕ̃
X
ϕ̃e
ϕ̃e
(3)
(2)
ej ϕ̃ Rt,j εt−j
Xt
= OP (1)
σ
σ
0≤j<t
q
= OP (1) ϕ̃e−tϕ̃ (ϕ̃2 + ω 2)
= OP (1)
−tϕ̃
e
2
2
s X
j 2 e2j ϕ̃
0≤j<t
(ϕ̃ + ω )
√ 3
2 ϕ̃
= oP (1)
as n → ∞, where Lemma 3.2 has been used to obtain the third equality sign, while the
final one follows from (2.6) and (2.7).
2
14
Lemma 3.6 Under the assumptions of Theorem 2.1,
√ −tϕ̃
ϕ̃e
(3)
Xt = oP (1)
(n → ∞),
σ
(2)
where Rt,j satisfies the corresponding rate in Theorem 3.1.
Proof: It holds,
√
ϕ̃e−tϕ̃ (3)
Xt
=
σ
√
ϕ̃e−tϕ̃ X j ϕ̃ (2)
e Rt,j εt−j
σ
0≤j<t
√ −tϕ̃
X
ϕ̃e
ej ϕ̃ εt−j
tω 2
= OP (1) √
tσ
0≤j<t
q
= OP (1) ϕ̃e−tϕ̃ tω 2
q
= OP (1) ϕ̃e−tϕ̃ tω 2
X
e2j ϕ̃
0≤j<t
1
2ϕ̃
e−tϕ̃ tω 2
= OP (1) √
2 ϕ̃
= oP (1)
as n → ∞, where the respective equality sign have been obtained by (1) definition, (2)
Theorem 3.1, (3) the central limit theorem for the {εi}, (4) Lemma 3.2, (5) simple algebra,
and (6) by (2.6) and (2.7).
2
Proof of Theorem 2.1: Putting together Lemmas 3.3–3.6, it remains only to check the
asymptotic independence of the finite dimensional distributions. Choose ti = ⌊κi n⌋ >
⌊κj n⌋ = tj . Then,
E
Z
ti
xϕ̃
e
0
=
Z
tj
d[W (t) − W (x)]
Z
0
tj
xϕ̃
e
d[W (t) − W (x)]
e2xϕ̃ dx
0
2tj ϕ̃
e
(1 + o(1)).
2ϕ̃
Thus, using the normalization provided in Theorem 2.1, the finite dimensional distributions of the limiting process have exponentially decreasing covariances
=
e2(κj −1)nϕ̃ → 0
(n → ∞),
since κj − 1 < 0 and nϕ̃ → ∞ by (2.6), completing the proof.
15
2
3.3
Proof of Theorem 2.2
We follow the arguments of the previous section using the decomposition (3.6). It will
(4)
be shown in a series of lemmas that Xt is determining the limit distribution, while
the other terms are asymptotically negligible. The main auxiliary result is the following
lemma. Recall that t = ⌊κn⌋ with some κ ∈ (0, 1) by (3.1).
Lemma 3.7 Under the assumptions of Theorem 2.2,
1
D
√ Xt(4) −→ W (κ)
nσ
(n → ∞),
where {W (s) : s ∈ [0, 1]} denotes a standard Wiener process.
Proof. As derived in (3.7),
X
0≤j<t
so
bt−j = ω
X
γt−j = oP (1)
(n → ∞),
0≤j<t
1
1 X
√ Xt(4) = √
εt−j (1 + oP (1))
nσ
nσ 0≤j<t
(n → ∞).
By assumption (2.9) there is a Wiener process {W (1) (s) : s ≥ 0} such that
X
1≤j≤s
εj − σW (1) (s) = o s1/(2+δ)
a.s.
(s → ∞)
for some δ > 0. Using this strong approximation and the scale transformation for Wiener
processes, we obtain on recalling once more (3.1) that
1 X
D
√
εt−j −→ W (κ)
nσ 0≤j<t
(n → ∞),
where {W (t) : t ∈ [0, 1]} denotes a standard Wiener process. This completes the proof.
2
It remains to show that all other terms in the representation (3.6) are negligible. This
will be done in the following lemmas.
Lemma 3.8 Under the assumptions of Theorem 2.2,
1
√ Xt(1) = oP (1)
nσ
(1)
where Rt
(n → ∞),
satisfies the corresponding rate in Theorem 3.1.
16
√
(1)
Proof. By (2.7) and Theorem 3.1, Rt = oP (1). Moreover, ( nσ)−1 X0 = oP (1) and,
as in (3.7), bt + . . . + b1 = oP (1) after an application of the central limit theorem to the
partial sums of the γi ’s and the claim follows.
2
Lemma 3.9 Under the assumptions of Theorem 2.2,
1
√ Xt(2) = oP (1)
nσ
(2)
(n → ∞),
(3)
where Rt,j and Rt,j satisfy the corresponding rates in Theorem 3.1.
Proof. Theorem 3.1, (2.7) and (3.7) imply that
X
γt−j
max ω
0≤j<t 0≤i<t
Hence,
(2) + Rt,j = oP (1)
(n → ∞).
1 X (3)
1
√ Xt(2) = OP (1) √
R εt−j
nσ
nσ 0≤i<t t,j
ω2 X
jεt−j
= OP (1) √
nσ 0≤i<t
ω2 s X 2
j
= OP (1) √
n 0≤i<t
ω 2t3/2
= OP (1) √
n
= oP (1)
(3)
as n → ∞, where we have applied (1) the rate established for Rt,j in Theorem 3.1, (2)
the central limit theorem for the sequence {iεi }, (3) that the order of the sum of squares
of the first t positive integers is t3 , and finally (4) assumptions (2.7) and (3.1).
2
Lemma 3.10 Under the assumptions of Theorem 2.2,
1
√ Xt(3) = oP (1)
nσ
(2)
(n → ∞),
where Rt,j satisfies the corresponding rate in Theorem 3.1.
17
Proof. By similar arguments as in the previous proofs,
1
1 X (2)
1
√ Xt(3) = √
Rt,j εt−j = OP (1) √ t3/2 ω 2 = oP (1)
nσ
nσ 0≤j<t
n
(n → ∞),
using Theorem 3.1, (2.7) and (3.1).
2
Proof of Theorem 2.2: The assertion follows directly from Lemmas 3.7–3.10.
2
3.4
Proof of Theorem 2.3
Finally, we repeat the steps accomplished in the previous two subsections. First, we will
(4)
derive the limiting behavior of the term Xt and then show that all other terms of the
decomposition do not contribute to the asymptotics. Let
tm = ⌊κm n⌋
with some 0 < κm < 1,
m = 1, . . . , N.
(3.8)
Our first auxiliary result is the following lemma.
Lemma 3.11 Let ϕ̃ < 0. If the second part of assumption (2.6) is satisfied, then
q
D
2|ϕ̃|[τ1 , . . . , τN ] −→ σ[ξ1 , . . . , ξN ]
(n → ∞),
where, for m = 1, . . . , N,
τm =
X
0≤i<tm
eiϕ̃ εtm −i
and ξ1 , . . . , ξN are idependent, identically distributed standard normal random variables.
Proof. The proof is given in two steps. First, the Cramér–Wold device (see e.g. [4],
Theorem 29.4) is applied, i.e. we are investigating linear combinations of τ1 , . . . , τN . Then,
we show that the central limit theorem is satisfied by verifying Liapunov’s condition (see
e.g. [4], Theorem 27.3).
(i) Choose real numbers λ1 , . . . , λN and write
λ1 τ1 + . . . +h λN τN
i
X
=
λ1 e(t1 −i)ϕ̃ + λ2 e(t2 −i)ϕ̃ + . . . + λN e(tN −i)ϕ̃ εi
1≤i≤t1
+
X
λ2 e(t2 −i)ϕ̃ + . . . + λN e(tN −i)ϕ̃ εi
t1 <i≤t2
+...+
i
h
X
λN e(tN −i)ϕ̃ εi
tN−1 <i≤tN
= S1 + . . . + SN .
(3.9)
18
Clearly, S1 , . . . , SN have expectation zero. It holds for the variance of S1 ,
X h
ES12 = σ 2
λ1 e(t1 −i)ϕ̃ + . . . + λN e(tN −i)ϕ̃
1≤i≤t1
= σ
2
X
λ2j
X
e2(tj −i)ϕ̃ + σ 2
1≤i≤t1
1≤j≤N
e2(t1 −i)ϕ̃ =
X
0≤i<t1
1≤i≤t1
X
λj λl
e2iϕ̃ ∼
1
2|ϕ̃|
X
e(tj +tl −2i)ϕ̃ .
1≤i≤t1
1≤j6=l≤N
By Lemma 3.2,
X
i2
(n → ∞).
On the other hand, for 2 ≤ j ≤ N, using Lemma 3.2 also
X
2(tj −i)ϕ̃
e
2(tj −t1 )ϕ̃
=e
1≤i≤t1
X
2(t1 −i)ϕ̃
e
1≤i≤t1
1
=o
|ϕ̃|
!
(n → ∞),
since (tj − t1 )ϕ̃ < 0. Similar arguments apply to the second part of ES12 . In detail,
X
e(tj +tl −2i)ϕ̃ = e(tj −t1 )ϕ̃ e(tl −t1 )ϕ̃
1≤i≤t1
X
e2(t1 −i)ϕ̃
1≤i≤t1
1 tj −t1 )ϕ̃ (tl −t1 )ϕ̃
∼
e
e
2|ϕ̃|
!
1
= o
|ϕ̃|
as n → ∞. Hence, putting together the previous calculations, we obtain
ES12
1
λ2 σ 2
= 1 +o
2|ϕ̃|
|ϕ̃|
!
(n → ∞).
Repeating the above arguments for any of the remaining ESj2 (j = 2, . . . , N), we arrive
at
!
σ2 2
1
2
2
E [λ1 τ1 + . . . + λN τN ] =
(n → ∞),
(λ + . . . + λN ) + o
2|ϕ̃| 1
|ϕ̃|
finishing the first part of the proof.
(ii) It remains to verify Liapunov’s condition. Observe that by an rearrangement of (3.9),
we can write
λ1 τ1 + . . . + λN τN = µ1 ε1 + . . . + µtN εtN
19
with appropriately chosen coefficients µi , (i = 1, . . . , tN ). In the following it will be shown
that
h
|µ1 |2+δ E|ε1 |2+δ + . . . + |µtN |2+δ E|εtN |2+δ
h
µ21 Eε21 + . . . + µ2tN Eε2tN
i1/2
i1/(2+δ)
= o(1)
(n → ∞)
(3.10)
for some δ > 0. First, we determine the asymptotic rate of the denominator. It holds,
µ21 Eε21 + . . . + µ2tN Eε2tN
1
= σ 2 (µ21 + . . . + µ2tN ) = O
|ϕ̃|
!
(n → ∞).
To check the numerator, select the summation range i = 1, . . . , t1 and observe that by
Jensen’s inequality
2+δ
|µi|2+δ = λ1 e(t1 −i)ϕ̃ + . . . + λN e(tN −i)ϕ̃ h
i
≤ C1 (N) |λ1 |2+δ e(t1 −i)ϕ̃(2+δ) + . . . + λN |2+δ e(tN −i)ϕ̃(2+δ) .
Hence,
|µ1|
2+δ
+ . . . + |µt1 |
C1 (N)|λ1 |2+δ
1
∼
+o
(2 + δ)|ϕ̃|
|ϕ̃|
2+δ
!
≤
C2 (N)
|ϕ̃|
(n → ∞).
Using similar calculations for the remaining summation ranges, we find that the asymptotic rate in (3.10) can be estimated by
"
1
O
|ϕ̃|
#1/(2+δ)−1/2 

This completes the proof.
= o(1)
(n → ∞).
2
Lemma 3.11 immediately yields the limit distribution on noticing that
q
(4)
2|ϕ̃|Xt
∼
q
2|ϕ̃|
X
ej ϕ̃ εt−j
0≤j<t
(n → ∞)
along the lines of Lemmas 3.3 and 3.7.
Lemma 3.12 Under the assumptions of Theorem 2.3,
q
(1)
2|ϕ̃|σ −1 Xt
(1)
where Rt
= oP (1)
(n → ∞),
satisfies the corresponding rate in Theorem 3.1.
20
(1)
Proof. By (2.6), (2.7) and Theorem 3.1, Rt = oP (1). Moreover,
q
2|ϕ̃|σ −1 X0 etϕ̃ = oP (1)
(n → ∞),
since tϕ̃ → −∞ by assumption (2.6). Finally, (3.7) yields the assertion.
2
Lemma 3.13 Under the assumptions of Theorem 2.3,
q
(2)
2|ϕ̃|σ −1 Xt
(2)
= oP (1)
(n → ∞),
(3)
where Rt,j and Rt,j satisfy the corresponding rate in Theorem 3.1.
Proof. Following the lines of the proofs of Lemmas 3.5 and 3.9, we get
q
2|ϕ̃|σ
−1
q
(2)
Xt
= OP (1) 2|ϕ̃|σ −1
q
X
(3)
ej ϕ̃ Rt,j εt−j
0≤j<t
−1
= OP (1) 2|ϕ̃|σ (ϕ̃2 + ω 2)
q
= OP (1) 2|ϕ̃|(ϕ̃2 + ω 2 )
X
ej ϕ̃ jεt−j
0≤j<t
s X
e2j ϕ̃ j 2
0≤j<t
1
(ϕ̃2 + ω 2 )
= OP (1) q
2|ϕ̃|
= oP (1)
as n → ∞, where we have used Lemma 3.2 to obtain the fourth equality sign. The final
equality is verified on observing that by (2.6) and (2.7), ω 2 |ϕ̃|−1/2 = (ω 2 n1/2 )(|ϕ̃|n)−1/2 →
0 as n → ∞.
2
Lemma 3.14 Under the assumptions of Theorem 2.3,
q
(3)
2|ϕ̃|σ −1 Xt
= oP (1)
(n → ∞),
(2)
where Rt,j satisfies the corresponding rate in Theorem 3.1.
Proof. Applying Lemma 3.2 yields
q
(3)
2|ϕ̃|σ −1 Xt
=
q
2|ϕ̃|σ −1
X
0≤j<t
21
(2)
ej ϕ̃ Rt,j εt−j
q
= OP (1) 2|ϕ̃|σ −1 tω 2
q
= OP (1) 2|ϕ̃|tω 2
= OP (1)tω
= oP (1)
2
X
ej ϕ̃ εt−j
0≤j<t
s X
e2j ϕ̃
0≤j<t
as n → ∞ by (2.7) and (3.1), finishing the proof.
2
Proof of Theorem 2.3: It follows from Lemmas 3.11–3.14.
3.5
2
Proof of Theorem 2.4
The proof is based on the decomposition (3.1). So, we will study the limiting behavior of
(4)
(i)
the leading term Xt first, and thereafter show that terms involving Xt , i = 1, 2, 3, do
not contribute to the asymptotics.
(4) 2
In a first part, we shall study the partial sums of Xk
(4)
Xt
=
t−1
X
j ϕ̃
e εt−j (1 + oP (1)) =
t
X
. Since by assumption ε0 = X0 ,
e(t−j)ϕ̃ εj (1 + oP (1)),
t = 1, . . . , n,
j=1
j=0
n ≥ 1.
From the preceding, we obtain
n
X
(4) 2
Xt
=

n
X

t=1
t=1
t
X
j=1
2
e(t−j)ϕ̃ εj  (1 + oP (1))
and it suffices to investigate the sum on the right–hand side of the latter equation. Let
T > 0. Then
n
X
t=1


t
X
j=1
2
e(t−j)ϕ̃ εj 
=
n
X
t=1
=
e2tϕ̃ 
⌊T /ϕ̃⌋
X
t=1
+

t
X
j=1

e2tϕ̃ 
n
X
e−j ϕ̃ εj 
t
X
j=1
t=⌊T /ϕ̃⌋+1
2
2
e−j ϕ̃ εj  +

e2tϕ̃ 
⌊T /ϕ̃⌋
X
j=1
22
n
X
t=⌊T /ϕ̃⌋+1
2
e−j ϕ̃ εj 

e2tϕ̃ 
t
X
j=⌊T /ϕ̃⌋+1
2
e−j ϕ̃ εj 
= ST,1 + ST,2 + ST,3 .
(3.11)
The next lemma identifies ST,3 as leading term by showing that ST,1 and ST,2 are asymptotically small.
Lemma 3.15 Let the assumptions of Theorem 2.4 be satisfied.
(i) For every T > 0 and ε > 0,
lim P
n→∞
(
ϕ̃2
ST,1 ≥ ε = 0
e2nϕ̃
)
where ST,1 is defined in (3.11).
(ii) For every ε > 0,
lim lim sup P
T →∞
n→∞
(
ϕ̃2
ST,2 ≥ ε = 0,
e2nϕ̃
)
where ST,2 is defined in (3.11).
Proof. The proof is based on an application of Markov’s inequality. Throughout the
proof C denotes a universal constant, which may vary from line to line.
(i) Note that

⌊T /ϕ̃⌋
 X
EST,1 = E 
≤ C
t=1

e2tϕ̃ 
t
X
j=1
⌊T /ϕ̃⌋
σ 2 X 2tϕ̃
e
ϕ̃ t=1
2 

ρn−j ϕ̃ εj  
≤ C
σ2
exp(2ϕ̃⌊T /ϕ̃⌋)
ϕ̃2
≤ C
σ2
exp(2T )
ϕ̃2
for any fixed T > 0. Now Markov’s inequality yields
P
(
ϕ̃2
ϕ̃2
S
≥
ε
≤
EST,1
T,1
e2nϕ̃
εe2nϕ̃
)
23
for any ε > 0. The quantity on the right–hand side of the latter inequality becomes
arbitrarily small as n → ∞, hence part (i) is proved.
(ii) By similar arguments

EST,2 = E 

n
X
t=⌊T /ϕ̃⌋+1
≤ Cσ 2
σ2
≤ C
ϕ̃
≤ C

e2tϕ̃ 
n
X
t
X
j=⌊T /ϕ̃⌋
t
X
e2tϕ̃
t=⌊T /ϕ̃⌋+1
n
X
2 

e−j ϕ̃ εj  
e−2j ϕ̃
j=⌊T /ϕ̃⌋
exp(2tϕ̃) exp(−2ϕ̃⌊T /ϕ̃⌋)
t=⌊T /ϕ̃⌋+1
σ2
exp(−2T ) exp(2nϕ̃).
ϕ̃2
So, Markov’s inequality gives part (ii) of the lemma and the proof is complete.
2
Lemma 3.15 tells us that it suffices to derive the limit distribution of the remaining
(4)
term ST,3 . The second part of the proof deals with the partial sums of Xt−1 εt . Let again
be T > 0. Then
n
X
t=1
(4)
Xt−1 εt
=
⌊T /ϕ̃⌋
X
t=1
+
n
X
(4)
Xt−1 εt +
e(t−1)ϕ̃ εt
t=⌊T /ϕ̃⌋+1
n
X
e(t−1)ϕ̃ εt
⌊T /ϕ̃⌋
X
t
X
e−j ϕ̃ εj
j=⌊T /cn ⌋
e−j ϕ̃ εj
j=1
k=⌊T /ϕ̃⌋+1
= RT,1 + RT,2 + RT,3 .
(3.12)
It will be shown next that RT,3 is the leading term, while RT,1 and RT,2 do not contribute
to the limit distribution.
Lemma 3.16 Let the assumptions of Theorem 2.4 be satisfied.
(i) For every T > 0 and ε > 0,
lim P
n→∞
ϕ̃
|RT,1 | ≥ ε = 0,
enϕ̃
24
where RT,1 is defined in (3.12).
(ii) For every ε > 0,
lim lim sup P
T →∞
n→∞
ϕ̃
|RT,2 | ≥ ε = 0,
enϕ̃
where RT,2 is defined in (3.12).
Proof. The proof is based on an application of Chebyshev’s inequality. Since ERT,i = 0,
2
we have VarRT,i = ERT,i
for i = 1, 2, 3 and it is enough to give estimates for the second
moments. Let C denote a universal but varying constant.
(i) Observe that

2
ERT,1
= E
≤ C
≤ C
⌊T /ϕ̃⌋
X
t=1
2
(4)
Xt−1 εk 
⌊T /ϕ̃⌋
σ 4 X 2tϕ̃
e
ϕ̃ t=1
σ4
exp(2T ),
ϕ̃2
(4)
where the second line is obtained using the independence of εk and Xt−1 . On recognizing
that by Chebyshev’s inequality
P
ϕ̃2
ϕ̃
2
|R
|
≥
ε
≤
ERT,1
,
T,1
n
ϕ̃
2
e
ε exp(2nϕ̃)
the probability on the left–hand side gets arbitrarily small for any fixed T > 0 if n → ∞.
So, part (i) is readily proved.
(ii) Note that,

2
= E
ERT,2
≤ C
n
X
e(t−1)ϕ̃ εt
k=⌊T /ϕ̃⌋+1
t−1
X
j=⌊T /ϕ̃⌋
σ4
exp(−2T ) exp(2nϕ̃).
ϕ̃2
Again, Chebyshev’s inequality gives claim (ii).
25
2
e−j ϕ̃ εj 
2
So, it suffices again to investigate the term RT,3 to obtain the limit distribution. Checking the remaining summation ranges in the corresponding double sum shows us that RT,3
is a product of two independent factors, whose limit can be calculated separately.
Lemma 3.17 Let the assumptions of Theorem 2.4 be satisfied. Then
2ϕ̃2
ϕ̃
D
ST,3 , 2 nϕ̃ RT,3 −→ (N12 , N1 N2 )
2
2n
ϕ̃
σ e
σ e
!
(n → ∞)
for any fixed T > 0, where N1 and N2 denote independent zero mean normal random
variables with EN12 = 1 − e−T and EN22 = 1.
Proof. The proof is given in two steps, each of them investigating one of the components
of the vector of interest. The joint behavior is obtained on combining both results.
(i) Observe that
ST,3 =

⌊T /ϕ̃⌋
n
X
t=⌊T /ϕ̃⌋

e2tϕ̃ 
X
j=0
2
e−j ϕ̃ εj 
2
⌊T /ϕ̃⌋
e2nϕ̃ q X −j ϕ̃ 
=
e εj (1 + o(1))
ϕ̃
2ϕ̃2
j=0
as n → ∞. Let S(x) = ε0 + . . . + ε⌊x⌋ . We need to determine the limit of
UT,1 =
=
q
ϕ̃
⌊T /ϕ̃⌋
X
e−j ϕ̃ εj
(3.13)
j=0
q Z
ϕ̃
T /ϕ̃
0
exp(−xϕ̃)dS(x)
!
x
ϕ̃
exp(−x)dS
=
ϕ̃
0
!
! #
"
Z T
q
x
T
=
exp(−x)S
− ε0 +
dx
ϕ̃ exp(−T )S
ϕ̃
ϕ̃
0
√
using integration by parts. The weak convergence of ϕ̃S(t/ϕ̃) to σW (t) on [0, T ], with
{W (t) : t ∈ [0, T ]} denoting a standard Wiener process, implies
q Z
q
ϕ̃
⌊T /ϕ̃⌋
X
j=0
T
D
exp(−j ϕ̃)εj −→ σ
Z
T
0
exp(−x)dW (x)
26
(n → ∞),
where we have also used integration by parts for the Wiener process.
(ii) For the second component we have,
ϕ̃
RT,3 = UT,1 UT,2 ,
enϕ̃
where
UT,2 =
√
n
X
ϕ̃
enϕ̃
e(t−1)ϕ̃ εt .
t=⌊T /ϕ̃⌋+1
The second sum contains noise terms starting with index t = ⌊T /ϕ̃⌋ + 1, while these
terms stop at j = ⌊T /ϕ̃⌋ in the first one. Hence, UT,1 and UT,2 are independent, yielding
the independence of N1 and N2 . The limit of UT,1 is already established in (3.13). So, it
remains to determine the limit distribution of
√
n
n
q
X
ϕ̃ X
(t−1)ϕ̃
ϕ̃
exp (−(n − t + 1)ϕ̃) εt
e
ε
=
t
enϕ̃ t=⌊T /ϕ̃⌋+1
t=⌊T /ϕ̃⌋+1
D
=
q
n−⌊T /ϕ̃⌋
ϕ̃
X
exp(−j ϕ̃)εj + oP (1).
j=0
Let {W̃ : t ≥ 0} be a standard Wiener process independent of {W (t) : t ∈ [0, T ]}. Then,
arguing as in part (i) of the proof,
√
Z ∞
n
ϕ̃ X
D
(t−1)ϕ̃
e
εt −→ σ
exp(−x)dW̃ (x)
(n → ∞).
enϕ̃ t=⌊T /ϕ̃⌋+1
0
The proof is complete after combining the results of parts (i) and (ii).
It is enough to show that the squares of the remaining terms in (3.1) are negligible.
Lemma 3.18 Let the assumptions of Theorem 2.4 be satisfied. Then,
n
ϕ̃2 X
(i) 2
(n) = oP (1)
X
t
e2nϕ̃ k=1
(n → ∞)
for i = 1, 2, 3.
Proof. (i) Let i = 1. Then
n
n
ϕ̃2 X
ϕ̃2 X
(1) 2
2 2tϕ̃
2
= oP (1)
X
=
X
e
=
O
n
ϕ̃
P
t
0
e2nϕ̃ t=1
e2nϕ̃ t=1
27
(n → ∞).
2
(ii) Let i = 2. Then,
n
ϕ̃2 X
(2) 2
Xt = OP
2n
ϕ̃
e
t=1
n(ϕ̃2 + ω 2 )2
ϕ̃e2nϕ̃
!
= oP (1)
(n → ∞).
(iii) Let i = 3. Then,
n
ϕ̃2 X
(3) 2
= OP
X
t
e2nϕ̃ t=1
n2 ω 2
e2nϕ̃
!
= oP (1)
(n → ∞).
Throughout, we have implicitly used Lemmas 3.4–3.6. The proof is complete.
2
Lemma 3.19 Let the assumptions of Theorem 2.4 be satisfied. Then,
n
ϕ̃ X
(i)
Xt−1 εt = oP (1)
n
ϕ̃
e t=1
(n → ∞)
for i = 1, 2, 3.
Proof. Similar arguments as in Lemma 3.18 lead to the conclusions paraphrasing the
proofs of Lemmas 3.4–3.6.
2
Proof of Theorem 2.4. Collecting the results of the previous lemmas, we arrive at
lim lim sup P
T →∞
n→∞
(
enϕ̃n
2ϕ̃n
σ 2 e2nϕ̃ 1
ϕ̃
(ϕ̂n − ϕn ) −
R
T,3 > ε = 0
2ϕ̃2 ST,3 σ 2 enϕ̃
)
for all ε > 0. By Lemma 3.17, for any T > 0 we have that
ϕ̃
σ 2 e2nϕ̃ 1
D N1 N2
RT,3 −→
2
2
n
ϕ̃
2ϕ̃ ST,3 σ e
N12
We note
D
(N1 , N2 ) =
q
1−
e−T M1 , M2
(n → ∞).
,
where M1 and M2 are independent standard normal random variables. So, we get
1
M2
N1 N2 D
=√
,
2
N1
1 − e−T |M1|
which converges to a standard Cauchy random variable as T → ∞.
28
2
3.6
Proof of Theorem 2.5
We follow the pattern provided in the previous subsection.
Lemma 3.20 Let the assumptions of Theorem 2.5 be satisfied. Then
(i)
(ii)
n
2|ϕ̃| X
(4) 2 P
Xt−1 −→ 1
2
σ n t=1
s
(n → ∞).
n
2|ϕ̃| X
D
(4)
Xt−1 εt −→ ζ
4
σ n t=1
(n → ∞),
where ζ denotes a standard normal random variable.
Proof. (i) An application of Markov’s inequality yields
P
(
n
2|ϕ̃| X
(4) 2
Xt−1
2
σ n
t=1
− 1
n
1 2|ϕ̃| X
(4) 2
Xt−1 − 1 .
≥ε ≤ E 2
ε σ n t=1
)
It is therefore enough to consider the expectation on the right–hand side. It holds,
n
2|ϕ̃| X
(4) 2
Xt−1
E 2
σ n t=1
"
#
=
=
=
t−1
n X
2|ϕ̃| X
e2j ϕ̃ σ 2
σ 2 n t=1 j=0
n
2|ϕ̃| X
1 − e2tϕ̃
n t=1 1 − e2ϕ̃
1 − e2nϕ̃ (1 + o(1))
as n → ∞. So, the claim follows since nϕ̃ → −∞.
(ii) See Theorem 3.2(b) in Phillips and Magdalinos (2005).
2
Lemma 3.21 Let the assumptions of Theorem 2.5 be satisfied. Then,
n
ϕ̃2 X
(i) 2
Xt = oP (1)
2n
ϕ̃
e
k=1
(n → ∞)
for i = 1, 2, 3.
Proof. (i) Let i = 1. Then
n
n
2|ϕ̃| X
2|ϕ̃| X
(1) 2
Xt =
X02 e2tϕ̃ = OP |ϕ̃|e2nϕ̃ = oP (1)
n t=1
n t=1
29
(n → ∞).
(ii) Let i = 2. Then,
n
2|ϕ̃| X
(2) 2
Xt = OP
n t=1
(ϕ̃2 + ω 2 )2
|ϕ̃|2
!
= oP (1)
(n → ∞).
(iii) Let i = 3. Then,
n
2|ϕ̃| X
(3) 2
Xt = OP n2 ω 4 = oP (1)
n t=1
(n → ∞).
The proof is complete.
2
Lemma 3.22 Let the assumptions of Theorem 2.5 be satisfied. Then,
n
2|ϕ̃| X
(i)
Xt−1 εt = oP (1)
n t=1
(n → ∞)
for i = 1, 2, 3.
Proof. Similar arguments as in Lemma 3.18 lead to the conclusions.
2
Proof of Theorem 2.5. Collecting the results of the previous lemmas, we arrive at
s
n
n
2|ϕ̃| X
(4) 2
Xt−1
(ϕ̂n − ϕn ) =
2
2|ϕ̃n |
σ n t=1
"
#−1 s
n
2|ϕ̃| X
(4)
Xt−1 εt + oP (1) → ζ
4
σ n t=1
as n → ∞, where ζ denotes a standard normal random variable.
2
References
[1] Aue, A., and Horváth, L. (2005). A limit theorem for mildly explosive autoregression
with stable errors. Preprint, University of Utah.
[2] Aue, A., Horváth, L., and Steinebach, J. (2004). Estimation in random coefficient
autoregressive models. Preprint, University of Utah.
[3] Berkes, I., Horváth, L., and Kokoszka, P. (2005). Near–integrated GARCH sequences.
Ann. Appl. Probab. 15, 890–913.
[4] Billingsley, P. (1995). Probability and Measure (3rd ed.). Wiley, New York.
30
[5] Bollerslev, T. (1986). Generalized autoregressive conditional hetereoscedasticity. J.
Econ. 31, 307–327.
[6] Brockwell P.J., and Davis R.A. (1991). Time Series: Theory and Methods (2nd ed.).
Springer–Verlag, Berlin.
[7] Chan, N.H., and Wei, C.Z. (1987). Asymptotic inference for nearly nonstationary
AR(1) processes. Ann. Statist. 15, 1050–1063.
[8] Chan, N.H., and Wei, C.Z. (1988). Limiting distributions of least squares estimates
of unstable autoregressive processes. Ann. Statist. 16, 367–401.
[9] Chow, Y.S., and Teicher, H. (1988). Probability Theory: Independence, Interchangeability, Martingales (2nd ed.). Springer–Verlag, New York.
[10] Feigin, P.D., and Tweedie, R.L. (1985). Random coefficient autoregressive processes:
a Markov chain analysis of stationarity and finiteness of moments. J. Time Ser. Anal.
6, 1–14.
[11] Komlós, J., Major, P., and Tusnády, G. (1975). An approximation of partial sums of
independent r.v.’s and the sample d.f. I. Z. Wahrsch. Verw. Gebiete 32, 111–131.
[12] Komlós, J., Major, P., and Tusnády, G. (1976). An approximation of partial sums of
independent r.v.’s and the sample d.f. II. Z. Wahrsch. Verw. Gebiete 34, 33–58.
[13] Ling, S., and Li, W.K. (1998). Limiting distributions of maximum likelihood estimators for unstable autoregressive moving–average time series with general autoregressive heteroscedastic errors. Ann. Statist. 26, 84–125.
[14] Nelson, C.R., and Plosser, C. (1982). Trends and random walks in macroeconomic
time series: some evidence and implications. Journal of Monetary Economics 10,
139–162.
[15] Nicholls, D.F., and Quinn, B.G. (1982). Random Coefficient Autoregressive Models:
an Introduction. Springer–Verlag, New York.
[16] Phillips, P.C.B. (1988). Regression theory for near–integrated time series. Econometrica 56, 1021–1043.
[17] Phillips, P.C.B., and Magdalinos, T. (2005). Limit theory for moderate deviations
from a unit root. Preprint, Yale University.
31
[18] Tong, H. (1990). Non–Linear Time Series: a Dynamical System Approach. Clarendon, Oxford.
32
Download