Selection from a stable box

advertisement
Selection from a stable box
Alexander Aue
1
István Berkes
2
Lajos Horváth
3
Abstract: Let {Xi } be independent, identically distributed random variables. It is well–
known that the functional CUSUM statistics and its randomly permuted version both converge
weakly to a Brownian bridge if second moments exist.
Surprisingly, an infinite variance counterpart does not hold true. In the present paper, we
let {Xi } be in the domain of attraction of a strictly α–stable law, α ∈ (0, 2). While the
functional CUSUM statistics itself converges to an α–stable bridge, and so does the permuted
version as long as both the {Xi } and the permutation are random, the situation turns out
to be more delicate if a realization of the {Xi} is fixed and randomness is restricted to the
permutation. Here, the conditional characteristic function of the permuted CUSUM statistics
converges in probability to a random and nondegenerate limit.
AMS 2000 Subject Classification: Primary 60F17, 60G52; Secondary 60E07.
Keywords and Phrases: Stable distributions; CUSUM; Order statistics, Functional limit
theorems, Permutation principle
1
Department of Mathematics, University of Utah, 155 South 1440 East, Salt Lake City, UT 84112–0090,
USA, email: aue@math.utah.edu
2
Department of Statistics, Graz University of Technology, Steyrergasse 17/IV, A–8010 Graz, Austria, email:
berkes@tugraz.at
3
Department of Mathematics, University of Utah, 155 South 1440 East, Salt Lake City, UT 84112–0090,
USA, email: horvath@math.utah.edu
Corresponding author: A. Aue, phone: +1–801–581–5231, fax: +1–801–581–4148.
Research partially supported by NATO grant PST.EAP.CLG 980599, NSF–OTKA grant INT–0223262 and
Hungarian National Foundation for Scientific Research Grants T 043037 and T 037886
1
Introduction
Let X, X1 , . . . , Xn be independent, identically distributed random variables. CUSUM–based
procedures, frequently used to test for time homogeneity of an underlying random phenomenon, are functionals of the process
⌊nt⌋
1 X
Zn (t) =
(Xi − X̄n ),
sn i=1
t ∈ [0, 1],
(1.1)
where ⌊·⌋ denotes integer part and
n
1X
Xi
X̄n =
n i=1
n
and
s2n
1 X
=
(Xi − X̄n )2
n − 1 i=1
are the sample mean and the sample variance, respectively. The asymptotic properties of
{Zn (t)} are immediate consequences of Donsker’s invariance principle [see Billingsley (1968)].
D[0,1]
Denote by −→ convergence in the Skorohod space D[0, 1]. We state the following fundamental
theorem.
Theorem 1.1 If EX 2 < ∞, then
D[0,1]
Zn (t) −→ B(t)
(n → ∞),
where {B(t) : t ∈ [0, 1]} is a Brownian bridge.
One of the main problems in testing for a homogeneous data generating process is to obtain
approximations for the critical values. It is well–known, however, that the convergence of functionals of Zn (t) to their limit distributions is often quite slow [sometimes even the convergence
speed is unknown], raising the question if Theorem 1.1 is reasonably applicable in case of small
sample sizes. To avoid these complications, along with bootstrap methods, the permutation
principle has been suggested to simulate critical values. For a comprehensive recent survey on
this topic we refer to Hušková (2004). Let π = (π(1), . . . , π(n)) be a random permutation of
(1, . . . , n) which is independent of X = (X1 , X2 , . . .). Then, the permuted version of Zn (t) is
defined by
Zn,π (t) =
1 X
(Xπ(i) − X̄n ),
sn i≤nt
t ∈ [0, 1].
(1.2)
One can think of Zn,π (t) as a CUSUM process whose summands are chosen from X1 , . . . , Xn
by drawing from a box without replacement. The following result establishes the limiting
behavior of Zn,π (t) and, hence, shows that the permutation method works to simulate critical
values. Note also that, by empirical evidence, the convergence to the limit is faster than for
Zn (t).
2
Theorem 1.2 If EX 2 < ∞, then, for almost all realizations of X,
D[0,1]
Zn,π (t) −→ B(t)
(n → ∞),
where {B(t) : t ∈ [0, 1]} is a Brownian bridge.
The proof of Theorem 1.2 can be found in Billingsley (1968, p. 209–214).
In this paper we are interested in the asymptotics of the analogues of Zn (t) and Zn,π (t) in
the infinite variance case EX 2 = ∞. To this end, let α ∈ (0, 2] and recall that a random
variable ξα is strictly α–stable if its characteristic function has the form

exp(−t2 /2)
if α = 2,


exp(−|t|α [1 − iβ sgn(t) tan( 21 α)])
if α ∈ (0, 1) ∪ (1, 2), β ∈ [−1, 1],
φα (t) =


exp(−|t|)
if α = 1.
The definition includes the normal law [α = 2] and the Cauchy law [α = 1], where the latter is
moreover assumed to be symmetric. Note that E[ξα ] exists only for α ∈ (1, 2] and E[ξα2 ] only
for α = 2. The class of stable distributions has become more and more important in theory as
well as in applications. Stable laws appear in a natural way in areas such as radio engineering,
electronics, biology and economics [see Zolotarev (1986, Chapter 1)]. For other expositions on
α–stable random variables and processes we refer to Samorodnitsky and Taqqu (1994), and
Bingham et al. (1987).
In what follows, let α ∈ (0, 2) and consider independent, identically distributed random
variables {Xi } which are in the domain of attraction of a strictly α–stable law. This means
that there exists a function L(x), slowly varying at infinity, such that
n
X
1
D
(Xi − µα ) −→ ξα
1/α
n L(n) i=1
(n → ∞),
(1.3)
where ξα is a strictly α–stable random variable and µα = 0 if α ∈ (0, 1]. If (1.3) holds, then
there is a p ∈ [0, 1] such that the tail probabilities of X and |X| satisfy the relations
P {X > y}
=p
y→∞ P {|X| > y}
lim
and
P {X < −y}
= q,
y→∞ P {|X| > y}
lim
(1.4)
where q = 1 − p. See Feller (1966) and Petrov (1995) for details.
To define the counterparts of Zn (t) and Zn,π (t) in the α–stable case, introduce the process
An (t) =
1 X
(Xi − X̄n ),
Tn i≤nt
t ∈ [0, 1],
(1.5)
and its permuted version
An,π (t) =
1 X
(Xπ(i) − X̄n ),
Tn i≤nt
t ∈ [0, 1],
3
(1.6)
where Tn = max1≤i≤n |Xi |, and π = (π(1), . . . , π(n)) is a random permutation of (1, . . . , n),
independent of X. InP
contrast to the finite variance case, Tn is a natural norming factor in
(1.5) and (1.6), since 1≤i≤n (Xi − µα )/Tn has a nondegenerate limit distribution.
One might now expect that {An (t)} and {An,π (t)} follow an asymptotic pattern similar to
that of {Zn (t)} and {Zn,π (t)}. Consequently, given a realization of X, one might want to use
the permutation principle to simulate critical values that have a higher accuracy in case of small
sample sizes. Surprisingly, this turns out to be impossible. While a version of Theorem 1.1
holds true for {An (t)}, an adaptation of Theorem 1.2 cannot be given for almost all realizations
of X. On the contrary, we will show that, conditioned on X, the characteristic function of the
permuted process {An,π (t)} converges in distribution to a random and nondegenerate limit for
every fixed t. On the other hand, there is an averaging effect: if both X and π are assumed
to be random, An,π (t) and An (t) converge weakly to the same limit, namely to an α–stable
bridge.
The paper is organized as follows. In Section 2 we will formulate precisely our results. Their
proofs will be given in Section 3.
2
Results
First we study the limiting behavior for the sequence {An (t) : t ∈ [0, 1]}. Its weak convergence
can easily be derived from the joint convergence of


⌊nt⌋


X
1
Tn
.
(X
−
µ
)
:
t
∈
[0,
1]
and
i
α
1/α
1/α
 n L(n)

n
L(n)
i=1
To deal with these two sequences, we are going to apply the method developed in LePage
et al. (1981), who proved that any stable distribution can be represented as an infinite sum
of random variables constructed from partial sums of exponential random variables. More
specifically, let {Ei } be a sequence of independent, identically distributed exponential random
variables having expected value 1, and define the sequence {Zi } by
Zi = (E1 + . . . + Ei )−1/α ,
i ≥ 1.
(2.1)
Additionally, let {δi } be independent, identically distributed random variables, independent
of {Ei }, satisfying P {δi = 1} = p and P {δi = −1} = q with p and q specified in (1.4). Then,
define
 ∞
X



δk Zk
0 < α ≤ 1,


k=1
(2.2)
η=
∞
X



(δk Zk − (p − q)E[Zk I{0 < Zk ≤ 1}])
1 < α < 2,


k=1
and Z = Z1 . LePage et al. (1981, Theorem 1) showed that the sums in (2.2) converge with
D
probability one and η = ξα . Consider a random vector (Wα (t), Z) such that Wα (t) is a strictly
4
D
α–stable process, Z = Z1 and the joint distribution of (Wα (t), Z) is defined by the requirement
that for any 0 = t1 < . . . < tK = 1, we have
(Wα (t2 ) − Wα (t1 ), Wα (t3 ) − Wα (t2 ), . . . , Wα (tK ) − Wα (tK−1 ), Z)
D
1/α
1/α
1/α
1/α
=
(t2 − t1 ) η1 , (t3 − t2 ) η2 , . . . , (tK − tK−1 ) ηK−1 , max (ti+1 − ti ) Zi ,
1≤i≤K−1
where (η1 , Z1 ), . . . , (ηK−1, ZK−1) are independent, identically distributed random vectors, distributed as (η, Z). Using the above, we will prove the following theorem.
Theorem 2.1 If (1.3) holds, then


⌊nt⌋
X
2 [0,1]
Tn
1
 D−→

(Wα (t), Z)
(X
−
µ
),
i
α
n1/α L(n) i=1
n1/α L(n)
(n → ∞),
where the distribution of the vector (Wα (t), Z) is defined above.
The asymptotic behavior of {An (t)} is now an immediate consequence of Theorem 2.1.
Corollary 2.1 If (1.3) holds, then
D[0,1]
An (t) −→
1
Bα (t)
Z
(n → ∞),
where, for t ∈ [0, 1], Bα (t) = Wα (t) − tWα (1).
The process {Bα (t) : t ∈ [0, 1]} is sometimes called an α–stable bridge. Very little is known
about the distributions of functionals of Bα (t)/Z and therefore it is hard to determine critical
values needed to construct asymptotic test procedures. So it would be nice to apply the permutation method to obtain critical values for functionals of An (t). However, as the subsequent
series of theorems shows, Theorem 1.2 does not have an infinite variance counterpart.
Consider the process {An,π (t)} defined in (1.6) and let both X and π be random. Then, we
obtain the following result.
Theorem 2.2 If (1.3) holds, then
D[0,1]
An,π (t) −→
1
Bα (t)
Z
(n → ∞),
(2.3)
where, for t ∈ [0, 1], Bα (t) = Wα (t) − tWα (1).
Obviously, Theorem 2.2 is much weaker than Theorem 1.2, since it is valid only for the average
of the realizations. However, a stronger result, holding true for almost all realizations of X,
cannot be proved. To this end, let PX and EX denote conditional probability and expected
value given X.
5
Theorem 2.3 Let t ∈ (0, 1). If (1.3) holds, then, as n → ∞,
1
PX {An,π (t) ≤ x} −→ P
Bα (t) ≤ x
a.s.
Z
cannot hold true.
Our last result states that the conditional characteristic function of An,π (t) given X converges in distribution, for any fixed t ∈ (0, 1), to a nondegenerate random variable which can
be explicitly given via a power series whose coefficients involve the {Zi } defined in (2.1). For
any m ≥ 2, set
Rm =
∞
X
(δi Zi )m
and
i=1
Qm =
Rm
.
Zm
(2.4)
Let further
cm (t) =
m
X
k X
k
(−1)k−ℓ ℓm ,
k ℓ=0 ℓ
k
k+1 t
(−1)
k=1
t ∈ (0, 1), m ≥ 2.
Theorem 2.4 Let t ∈ (0, 1). If (1.3) holds, then
)
( ∞
m
X
Qm (is)
D
cm (t)
EX eisAn,π (t) −→ exp
m!
m=2
(n → ∞)
for all real |s| < 0.05.
The previous results concern the asymptotic properties of {An (t)} and {An,π (t)} but it is
possible to consider modifications of these processes replacing the normalization Tn by a more
general sequence
Tn(ν) =
n
X
i=1
|Xi − X̄n |ν
!1/ν
with some ν > α.
The corresponding CUSUM processes are then defined by
A(ν)
n (t) =
⌊nt⌋
1 X
(ν)
Tn
i=1
(Xi − X̄n ),
A(ν)
n,π (t) =
⌊nt⌋
1 X
(ν)
Tn
i=1
(Xπ(i) − X̄n ),
t ∈ [0, 1].
Analogues of Theorems 2.1 and
2.2 can be easily
by exploiting the joint convergence
P⌊nt⌋established
P⌊nt⌋
|Xi|ν . Theorems 2.3 and 2.4 also remain true
Xi and i=1
of the partial sum processes i=1
[with some modifications in the corresponding limit processes], so that, conditionally on X,
(ν)
the permuted sequence {An,π (t)} cannot converge weakly in D[0, 1].
6
3
Proofs
The proof of Theorem 2.1 is based on the following result due to LePage et al. (1981). Their
method relies heavily on a representation of the order statistics X1,n , . . . , Xn,n in terms of
partial sums of exponential random variables as given in the previous section, and makes use
of the fact that the upper extreme values of |X1 |, . . . , |Xn | are asymptotically independent of
the signs corresponding to these extremes.
Theorem 3.1 If (1.3) holds, then
1
1/α
N L(N)
N
X
i=1
(Xi − µα ), TN
!
D
−→ (η, Z)
as N → ∞.
Proof of Theorem 2.1. First, we establish the convergence of the finite dimensional distributions. Let 0 = t1 < t2 < . . . < tK = 1 and observe that
Tn =
max
max
1≤j≤K−1 ntj <i≤ntj+1
Then, clearly

⌊ntj+1 ⌋
X

(Xi − µα ),
i=⌊ntj ⌋+1
|Xi |.
max
ntj <i≤ntj+1

|Xi | ,
j = 1, . . . , K − 1,
(3.1)
are independent vectors. Since, by the specification of {Xi }, the distribution of the vector in
(3.1) depends only on the number ∆j,n = ⌊ntj+1 ⌋ − ⌊ntj ⌋ of terms Xi − µα but not on the
specific indices i, we get for any j = 1, . . . , K − 1, as n → ∞,


∆j,n
X
1
 (Xi − µα ), T∆j,n 
1/α
n L(n) i=1
=
1/α
∆j,n L(∆j,n )
n1/α L(n)


∆j,n
1
1/α
∆j,n L(∆j,n )

X
i=1
D
(Xi − µα ), T∆j,n  −→ (tj+1 − tj )1/α ηj , (tj+1 − tj )1/α Zj ,
using Theorem 3.1 and that L is slowly varying at infinity. Therein, {ηj } and {Zj } are independent copies of η and Z, respectively. We conclude that the finite dimensional distributions
converge.
P⌊nt⌋
(Xi − µα )/(n1/α L(n)) is proven in Gikhman and Skorohod (1969, p.
The tightness of i=1
480).
2
Proof of Theorem 2.2. Note that, by the assumptions on {Xi },
D
{An (t), t ∈ [0, 1]} = {An,π (t), t ∈ [0, 1]} ,
7
so the result follows from Theorem 2.1.
2
From now on, we fix a realization of X. Permuting the X’s is the same as selecting n elements
from X1 , X2 , . . . , Xn without replacement. The proof of Theorem 2.3 will be based on the order
statistics of X1 , . . . , Xn . Our first goal is to express An,π (t) in terms of X1,n , . . . , Xn,n . To do
so, let
(
1
if Xi,n is among the first ⌊nt⌋ elements chosen,
(n)
εi (t) =
0
otherwise,
(n)
that is, εi (t) identifies those of the order statistics which are selected by the permutation π
in the first ⌊nt⌋ positions. Then, it is easy to see that
X
(Xπ(i) − X̄n ) =
i≤nt
(n)
(n)
n
X
j=1
(n)
(Xj,n − X̄n )εj (t) =
(n)
n
X
j=1
(n)
(Xj,n − X̄n )ε̄j (t),
(n)
(n)
where ε̄j (t) = εj (t) − Eεj (t) = εj (t) − ⌊nt⌋/n is the centered version of εi (t).
Let VarX denote conditional variance with respect to X.
Proof of Theorem 2.3. In a first step, we compute the conditional expected value and
variance of An,π (t). Observe that
2
2 ⌊nt⌋
⌊nt⌋
E ε̄k (t) =
−
(3.2)
n
n
and
h
i ⌊nt⌋(⌊nt⌋ − 1) ⌊nt⌋ 2
⌊nt⌋(n − ⌊nt⌋)
(n)
(n)
=−
−
.
E ε̄j (t)ε̄k (t) =
n(n − 1)
n
n2 (n − 1)
Then, it is straightforward that
n
h
i
1 X
(n)
(Xj,n − X̄n )E ε̄j (t) = 0
EX [An,π (t)] =
Tn j=1
and, on applying (3.2) and (3.3),
VarX (An,π (t)) = VarX
=
n
1 X
(n)
(Xj,n − X̄n )ε̄j (t)
Tn j=1
!
n
h
i
1 X
(n)
(n)
(X
−
X̄
)(X
−
X̄
)E
ε̄
(t)ε̄
(t)
j,n
n
k,n
n
j
k
Tn2 j,k=1
n
1 X
(Xj,n − X̄n )2 + Dn ,
= (t(1 − t) + o(1)) 2
Tn j=1
8
(3.3)
where
X
1
|Dn | ≤
(Xj,n − X̄n )(Xk,n − X̄n )
2
(n − 1)Tn j6=k
!2
X
n
n
X
1
2
=
(X
−
X̄
)
(X
−
X̄
)
−
k,n
n
k,n
n
(n − 1)Tn2 k=1
k=1
=
n
X
1
(Xk,n − X̄n )2 ,
(n − 1)Tn2
k=1
since
that
Pn
k=1 (Xk,n
1
2/α
n L2 (n)
− X̄n ) = 0. It follows from LePage et al. (1981) [cf. also Theorem 3.1 above]
n
X
i=1
(Xi − µα )2 , Tn2
D
D
−2/α
where η ∗ = ξα/2 and Z ∗ = E1
D
VarX (An,π (t)) →
!
D
−→ (η ∗ , Z ∗ )
(n → ∞),
(3.4)
. Consequently, the preceding calculations imply that
η∗
Z∗
(3.5)
as n → ∞. It is clear that η ∗ /Z ∗ is a nondegenerate random variable. Similarly to (3.5) one
can show that
D
EX A4n,π (t) → ζ,
(3.6)
as n → ∞ with some random variable ζ. Fix now t ∈ (0, 1). Relation (3.6) implies that
the sequence {EX A4n,π (t), 1 ≤ n < ∞} is tight and thus for any given ǫ > 0 there exists a
K = K(ǫ) > 0 and for each n ≥ 1 an event Ωn with P {Ωn } ≥ 1 − ǫ such that
EX A4n,π (t) ≤ K
on Ωn .
(3.7)
The inequality in (3.7) implies
EX A2n,π (t)I{|An,π (t) ≥ y}
1 2
2
A (t)y I{|An,π (t) ≥ y}
= EX
y 2 n,π
1
≤ 2 EX A4n,π (t)
y
K
≤ 2
y
9
(3.8)
for any y > 0 on Ωn . Let us assume indirectly that, for all points x of continuity of some
distribution function F , the following convergence result holds true:
PX {An,π (t) ≤ x} −→ F (x)
a.s.
(n → ∞).
This implies that for any continuity point y of F we have
Z y
2
EX An,π (t)I{|An,π (t)| < y} →
x2 dF (x)
a.s.
(3.9)
(3.10)
−y
We choose y ≥ (K/ǫ)1/2 and conclude by Egorov’s theorem [cf. Hewitt and Stromberg (1969,
p. 158)] that there exists an event Ω∗ with P {Ω∗ } ≥ 1 − ǫ on which (3.10) is uniform and
consequently there exists a nonrandom n0 such that
Z y
2
EX A2 (t)I{|An,π (t)| < y} −
(3.11)
x dF (x) ≤ ǫ
n,π
−y
on Ω∗ if n ≥ n0 . By (3.8) and (3.11) we have
Z y
2
EX A2 (t) −
x dF (x) ≤ 2ǫ
on Ωn ∩ Ω∗ ,
n,π
(3.12)
−y
if n ≥ n0 . Since we can choose y as large as we want in (3.12), we conclude
Z ∞
P
2
EX An,π (t) →
x2 dF (x)
(3.13)
−∞
regardless if the integral is finite or infinite. Since η ∗ /Z ∗ is nondegenerate, (3.13) contradicts
(3.5).
2
The proof of Theorem 2.4 will be based on a series of lemmas. Since {Xi } is a sequence of
strictly α–stable random variables with α ∈ (0, 2), only the very small and very large order
statistics will contribute asymptotically to An,π (t), which is shown in Lemmas 3.1 and 3.3.
Lemma 3.1 If (1.3) holds, then
( (
)
)
n−K
X
1
(n)
lim lim sup P PX
(Xi,n − X̄n )ε̄i (t) ≥ ε ≥ δ = 0
1/α
K→∞ n→∞
n L(n) i=K
for all ε > 0 and δ > 0.
Proof. Without loss of generality, we can assume that µα = 0 for all α ∈ (0, 2). By (3.2) and
(3.3) we have, similarly as in the proof of Theorem 2.3,
"n−K
#
X
(n)
EX
(Xi,n − X̄n )ε̄i (t) = 0
i=K
10
and
VarX
n−K
X
i=K
(Xi,n −
(n)
X̄n )ε̄i (t)
!
≤C
n−K
X
2
Xi,n
+
nX̄n2
i=K
with some constant C > 0. Using (1.3) we get
!2
n
2
X
nX̄n
1
1
Xi
=
= oP (1)
2/α
2
1/α
n L (n)
n n L(n) i=1
!
(n → ∞).
Let U1,n ≤ U2,n ≤ . . . ≤ Un,n be the order statistics of n independent, uniformly distributed
random variables on [0, 1]. The [generalized] inverse of the distribution function of X is denoted
by Q. It is well–known that
n−K
X
2 D
Xi,n
=
i=K
n−K
X
Q2 (Ui,n ).
i=K
Using the method of Csörgő et al. (1986), one can show that, for any K ≥ 1,
n−K
X
1
D
Q2 (Ui,n ) −→ A(K)
2/α
2
n L (n) i=K
(n → ∞),
where A(K) is a random variable satisfying A(K) → 0 a.s. as K → ∞. The methodology
developed in Csörgő et al. (1986) is based on the fact that the upper and lower extreme values
of X1 , . . . , Xn are asymptotically independent.
The proof of Lemma 3.1 is complete after an application of Markov’s inequality.
2
Without loss of generality, we assume that all processes used in this paper are defined on
the same probability space. Next, we utilize a result of Berkes and Philipp (1979) to show that
(n)
the dependent random variables {εi } can be approximated with a sequence of independent
Bernoulli variables.
Lemma 3.2 If (1.3) holds, then, for each n and each t ∈ (0, 1), there are independent, identically distributed random variables δi,n (t), i = 1, . . . , n, independent of {Xi }, with
P {δi,n (t) = 1} =
⌊nt⌋
n
and
P {δi,n (t) = 0} =
n − ⌊nt⌋
n
(3.14)
such that
PX
(
PX
(
and
K
)
X
1
P
(n)
(Xi,n − X̄n )(εi (t) − δi,n (t)) > x −→ 0
n1/α L(n) i=1
n
)
X
1
P
(n)
(Xi,n − X̄n )(εi (t) − δi,n (t)) > x −→ 0
1/α
n L(n) i=n−K+1
for all x > 0 and K ≥ 1.
11
(n → ∞)
(n → ∞)
(3.15)
(3.16)
(n)
(n)
Proof. For i ≥ 1, let γ2i−1 = εi (t) and γ2i = εn−i+1 . It is easy to see that there is a constant
C > 0 such that, for ai = 0 or 1, i ≥ 1, we have
P γk = ak γk−1 = ak−1 , . . . , γ1 = a1 − P {γk = ak } ≤ C k
n
for all k = 2, . . . , n. Hence, applying Theorem 2 of Berkes and Philipp (1979), there exist
independent Bernoulli random variables δi,n (t) satisfying (3.14) such that
k
k
(n)
≤ 6C
P εk (t) − δk,n (t) ≥ 6C
n
n
and
k
k
(n)
P εn−k+1(t) − δn−k+1,n (t) ≥ 6C
≤ 6C
n
n
for k = 1, . . . , ⌊n/2⌋. Hence, we can estimate
)
(
K
X
1
(n)
PX
(Xi,n − X̄n )(εi (t) − δi,n (t)) ≥ x
n1/α L(n) i=1
K
≤C +I
n
)
K
X
K
1
|Xi,n − X̄n | ≥ x .
C
n n1/α L(n) i=1
(
We need to further estimate the right–hand side of the latter inequality. Note that, from (1.3),
we readily obtain
X̄n
1+1/α
n
L(n)
P
−→ 0
(n → ∞).
Along the lines of Corollary 3.1 in Csörgő et al. (1986), it can be verified that
!
K
X
1
D
|Xi,n | − bn,α −→ τ
(n → ∞),
1/α
n L(n) i=1
where τ is a nondegenerate random variable and {bn,α } is a sequence satisfying the condition
bn,α (n1+1/α L(n))−1 → 0 as n → ∞. Thus, we arrive at
" (
)#
K
X
K
1
E I C
|Xi,n − X̄n | ≥ x
−→ 0
(n → ∞),
n n1/α L(n) i=1
completing the proof of (3.15). The same argument proves (3.16).
2
(n)
The following lemma is the counterpart of Lemma 3.1, in which the ε̄i are replaced by
mean–corrected variables δ̄i,n (t) obtained from the independent sequence {δi,n (t)}.
12
Lemma 3.3 If (1.3) holds, then
( (
)
)
n−K
X
1
lim lim sup P PX
(Xi,n − X̄n )δ̄i,n (t) ≥ ε ≥ δ = 0
1/α
K→∞ n→∞
n L(n) i=K
for all ε > 0 and δ > 0, where δ̄i,n (t) = δi,n (t) − ⌊nt⌋/n.
Proof. The assertion follows along the lines of the proof of Lemma 3.1.
2
Lemma 3.4 Let t ∈ (0, 1). If (1.3) holds, then for all ε > 0,
P
PX {|An,π (t) − Bn (t)| > ε} −→ 0
(n → ∞),
(3.17)
where
n
⌊nt⌋
1 X
Xi,n − X̄n δi,n (t) −
Bn (t) =
Tn i=1
n
and δi,n (t), i = 1, . . . , n, t ∈ [0, 1], are defined in Lemma 3.2.
Proof. It follows readily from Lemmas 3.1–3.3.
2
To complete the proof of Theorem 2.4, it suffices to show, in view of (3.17), that
)
( ∞
X Qm (is)m
D
cm (t)
(n → ∞).
EX eisBn (t) −→ exp
m!
m=2
√
Let Si,n = Tn−1 (Xi,n − X̄n ). Clearly, |Si,n | ≤ 2. To lighten the notation we write z = −1z,
where z is real. Since {δi,n (t)} is a sequence of independent Bernoulli variables, we get
zBn (t)
EX e
n
Y
⌊nt⌋ zSi,n
=
exp(−zSi,n ⌊nt⌋/n) 1 +
e
−1
n
i=1
and therefore
ψn (t, z) = log EX ezBn (t)
n
n
X
⌊nt⌋ zSi,n
⌊nt⌋ X
e
−1
z
Si,n +
log 1 +
=−
n
n
i=1
i=1
=
n
X
i=1
⌊nt⌋ zSi,n
log 1 +
e
−1 .
n
The next lemma is an immediate consequence of the mean–value theorem applied to the
function log(1 + x).
13
Lemma 3.5 Let |z| <
ψn (t, z) −
n
X
i=1
1
2
and t ∈ (0, 1). If (1.3) holds, then
log 1 + t ezSi,n − 1
P
−→ 0
(n → ∞).
Proof. By the mean–value theorem for the real and imaginary parts,
⌊nt⌋
e |z|
zS
zS
i,n
i,n
log 1 + t e
− 1 − log 1 +
e
− 1 ≤
|Si,n |.
n
1−t n
Arguments already applied in the proof of Lemma 3.2 yield that
n
n
1X
1 X
P
|Si,n | =
|Xi,n − X̄n | −→ 0
n i=1
nTn i=1
(n → ∞),
completing the proof.
2
Lemma 3.6 For any t ∈ (0, 1) and |z| < 0.1 it holds
log (1 + t(e − 1)) =
z
∞
X
zj
j=1
j!
cj (t),
and the series is absolutely convergent, where
j
k k X
X
k
k+1 t
(−1)k−ℓ ℓj .
(−1)
cj (t) =
k ℓ=0 ℓ
k=1
Proof. We note that
∞
k
X
k+1 u
log(1 + u) =
(−1)
k
k=1
(3.18)
is absolutely convergent for any complex u with |u| < 1. Also |ex − 1| < 1 for all complex
|x| < 0.1. First using (3.18), then the binomial theorem and, finally, Taylor expansion for the
exponential function, we conclude
x
log (1 + t(e − 1)) =
=
∞
X
(−1)k+1
k=1
∞
X
(−1)
k=1
=
∞
X
k=1
(−1)
[t(ex − 1)]k
k
k
k+1 t
k
k
k+1 t
k
k X
k
ℓ=0
ℓ
k X
k
ℓ=0
ℓ
(−1)k−ℓ eℓx
k−ℓ
(−1)
∞
X
(xℓ)j
j=0
j!
∞
k ∞
X
xj X (−1)k+1 k X k
(−1)k−ℓ ℓj .
t
=
ℓ
j!
k
j=0
k=1
ℓ=0
14
On observing that
k X
k
(−1)k−ℓ ℓj = 0
l
ℓ=0
if k > j,
Lemma 3.6 is proved.
2
Lemma 3.7 Let t ∈ (0, 1) and |z| < 0.1. If (1.3) holds, then
!
n
∞
n
j
X
X
X
z
j
log 1 + t ezSi,n − 1 =
Si,n
cj (t).
j!
i=1
j=2
i=1
Proof. Since
n
X
Si,n = 0,
i=1
the assertion follows directly from Lemma 3.6.
Lemma 3.8 If (1.3) holds, then
!
n
n
X
X
D
3
2
Si,n
,
Si,n
, . . . −→ (Q2 , Q3 , . . .)
i=1
i=1
2
(n → ∞),
where Q2 , Q3 , . . . are defined in (2.4) and the convergence is meant in the sense of convergence
of the finite dimensional distributions.
Proof. If m = 2 then, by (1.3),
2 X
n
n
X
nX̄n2
1
1
2
2
=
(X
−
X̄
)
−
X
=
i
n
i
1/α
2/α
2
n L (n)
n L(n) i=1
n
i=1
If m > 2, the mean–value theorem implies
n
n
n
X
X
X
m
m
2m−1 |Xi |m−1 + |X̄n |m−1
Xi ≤ m|X̄n |
(Xi − X̄n ) −
i=1
i=1
i=1
!
n
X
|Xi |m−1 + n|X̄n |m .
= m2m−1 |X̄n |
i=1
On applying (1.3) once again, we obtain
n
|X̄n |
1/α
n L(n)
m
=
1
nm−1
n
X
1
Xi
n1/α L(n) i=1
15
!m
P
−→ 0
n
X
1
Xi
n1/α L(n) i=1
(n → ∞)
!2
P
−→ 0.
and
1
n1/α L(n)
m
|X̄n |
n
X
i=1
|Xi |m−1
as m → ∞, since (n1/α L(n))−(m−1)
to show that, as n → ∞,
1
n1/α L(n)
Tn ,
1
n1/α L(n)
m−1 X
n
n
X
1
1
1
P
Xi |Xi|m−1 −→ 0,
= 1/α
1/α
n n L(n)
n L(n)
i=1
Pn
i=1
2 X
n
|Xi |m−1 converges in distribution. Hence it is enough
Xi2 ,
i=1
i=1
1
n1/α L(n)
3 X
n
!
Xi3 , . . .
i=1
D
−→ (Z1 , R2 , R3 , . . .) ,
where R2 , R3 , . . . are defined in (2.4). But this follows from the proof of Theorem 1 in LePage
et al. (1981) and the proof is complete.
2
Proof of Theorem 2.4. Applying Lemmas 3.5 and 3.7, it is enough to prove that
!
∞
n
∞
X
X
zj X j
zj
D
Qj cj (t)
(n → ∞).
Si,n cj (t) −→
j!
j!
j=2
i=1
j=2
For any M ≥ 2, Lemma 3.8 implies
!
M
n
M
X
X
zj X j
zj
D
Si,n cj (t) −→
Qj cj (t)
j! i=1
j!
j=2
j=2
(n → ∞).
Next we show that
∞
X
zj
P
Qj cj (t) −→ 0
j!
j=M +1
(M → ∞).
(3.19)
According to the strong law of large numbers, there is a random variable k0 such that
E2 + . . . + Ek ≥
k
2
if k ≥ k0 .
Hence,
|Qj | ≤
=
j
∞ X
Zi
i=1
∞ X
E1
E1 + . . . + Ei
j/α
k0 X
E1
E1 + . . . + Ei
j/α
i=1
=
(3.20)
Z1
i=1
+
∞
X
i=k0 +1
E1
E1 + . . . + Ei
16
j/α
≤ k0 +
j/α
E1
∞
X
i=k0 +1
1
E1 + i/2
j/α
≤ k0 + C(2E1 )j/α (2E1 + 1)−j/α+1
≤ k0 + C(2E1 + 1)
with some constant C > 0, where we replaced the last sum with an integral. The series in
Lemma 3.6 is absolutely convergent and thus we obtain
∞
∞
X
j
X
|z|j
z
P
Q
c
(t)
≤
(k
+
C(2E
+
1))
|cj (t)| −→ 0,
(3.21)
j j
0
1
j!
j!
j=M +1
j=M +1
establishing (3.19).
The final step contains the proof of
( ∞
)
!
n
X zj X
j
lim lim sup P S
cj (t) > ε = 0
M →∞ n→∞
j! i=1 i,n
j=M +1
(3.22)
for all ε > 0, t ∈ (0, 1) and |z| < 0.05. To this end, let K be the quantile function of |Xi |. By
(1.3) we have that
K(x) = (1 − x)−1/α U(1 − x)
(3.23)
with some function U slowly varying at zero. Also,
E1 + . . . + Ei
D
{|X|i,n, 1 ≤ i ≤ n} = K
,1 ≤ i ≤ n ,
E1 + . . . + En+1
where |X|1,n ≤ |X|2,n ≤ . . . ≤ |X|n,n denote the order statistics of |X1 |, . . . , |Xn |. For i =
1, . . . , n + 1, let
Si = E1 + . . . + Ei
and
Si∗ = Ei + . . . + En+1 .
Using (3.23), we obtain
1/α
∗ 1/α
∗
∗
∗
U Si+1
/Sn+1
Sn+1
/Sn+1
U Si+1
/Sn+1
Sn+1
K (Si /Sn+1 )
=
.
=
1/α
∗
∗
∗
∗
K (Sn /Sn+1 )
Si+1
U Sn+1
/Sn+1
U Sn+1
/Sn+1
Si+1
/Sn+1
It is easy to see that
)
(
) ( 1/α
1/α
∗
∗
U Si+1
/Sn+1
Sn+1
U
(S
/S
)
S
D
i
n+1
1
, 1 ≤ i ≤ n =
,1≤i≤n .
∗
∗
Si+1
S
U
(S
/S
U Sn+1 /Sn+1
i
1
n+1 )
17
Let δ > 0 and γ > 0. By the Karamata representation theorem [cf. Bingham et al. (1987)],
there are functions c(t) and u(t) such that c(t) → c0 , u(t) → 0 as t → 0 and
!
Z Si /Sn+1
c(Si /Sn+1 )
U(Si /Sn+1 )
u(t)
=
exp
dt .
U(S1 /Sn+1 )
c(S1 /Sn+1 )
t
S1 /Sn+1
There is ǫ > 0 such that
sup |u(t)| ≤ γ
0≤t≤2ǫ
and
c(t)
≤ 1 + δ.
0≤s,t≤2ǫ c(s)
sup
So by the law of the large numbers we have
γ
U(Si /Sn+1 )
Si
lim P
≤ (1 + δ)
for all 1 ≤ i ≤ ⌊nǫ⌋ = 1.
n→∞
U(S1 /Sn+1 )
S1
Let C = ((2/ǫ)γ supǫ/2≤x≤1 U(x))−1 . Then there is x0 > 0 such that
1
≤ Cx−γ
U(x)
if 0 ≤ x ≤ x0 .
On the event {S1 /Sn+1 ≤ x0 , S⌊nǫ⌋ /Sn+1 ≥ ǫ/2} we have, for all ⌊nǫ⌋ ≤ i ≤ n,
γ
Sn+1
U(Si /Sn+1 )
≤ sup U(x)C
U(S1 /Sn+1 ) ǫ/2≤x≤1
S1
γ
γ
Si
Sn+1
=
sup U(x)C
S1 ǫ/2≤x≤1
Si
γ
γ
Sn+1
Si
sup U(x)C
≤
S1
S⌊nǫ⌋
ǫ/2≤x≤1
γ
Si
≤
.
S1
Thus we conclude that for all δ > 0 and γ > 0
γ
U(Si /Sn+1 )
Si
lim P
≤ (1 + δ)
for all 1 ≤ i ≤ n = 1.
n→∞
U(S1 /Sn+1 )
S1
It follows from the estimates in (3.20) and (3.21) that
1/α−γ !j
∞
∞ X
X
zj
S1
P
→ 0,
(1 + δ)
j!
Si
i=1
j=M +1
if 0 < δ < 1/2 and 0 < γ < (2 − α)/(2α).
2
18
References
[1] Berkes, I., and Philipp, W. (1979). Approximation theorems for independent and weakly
dependent random vectors. The Annals of Probability 7, 29–54.
[2] Billingsley, P. (1968). Convergence of Probability Measures. Wiley, New York.
[3] Bingham, N.H., Goldie, C.M., and Teugels, J.L. (1987). Regular Variation. Cambridge
University Press, Cambridge, U.K.
[4] Csörgő, M., Csörgő, S., Horváth, L., and Mason, D.M. (1986). Normal and stable convergence of integral functions of the empirical distribution function. The Annals of Probability
14, 86–118.
[5] Feller, W. (1966). An Introduction to Probability Theory and its Applications, Vol. 2.
Wiley, New York.
[6] Gikhman, I.I., and Skorohod, A.V. (1969). Introduction to the Theory of Random Processes. Saunders, Philadelphia.
[7] Hewitt, E. and Stromberg, K. (1969). Real and Abstract Analysis. Springer, New York.
[8] Hušková, M. (2004). Permutation principle and bootstrap in change point analysis. In:
Asymptotic Methods in Stochastics, eds. L. Horváth and B. Szyszkowicz. American Mathematical Society, Providence, RI, pp. 273–291.
[9] LePage, R., Woodroofe, M., and Zinn, J. (1981). Convergence to a stable distribution via
order statistics. The Annals of Probability 9, 624–632.
[10] Petrov, V.V. (1995). Limit Theorems of Probability Theory. Clarendon, Oxford, U.K.
[11] Samorodnitsky, G., and Taqqu, M. (1994). Stable Non–Gaussian Random Processes.
Chapman & Hall, New York.
[12] Zolotarev, V.M. (1986). One–Dimensional Stable Distributions. Translations of Mathematical Monographs, Vol. 65. The American Mathematical Society, Providence, RI.
19
Download