PARAMETER ESTIMATION FOR PERIODICALLY STATIONARY TIME SERIES By Paul L. Anderson and Mark M. Meerschaert Albion College and University of Otago First Version received December 2002 Abstract. The innovations algorithm can be used to obtain parameter estimates for periodically stationary time series models. In this paper, we compute the asymptotic distribution for these estimates in the case, where the innovations have a finite fourth moment. These asymptotic results are useful to determine which model parameters are significant. In the process, we also develop asymptotics for the Yule–Walker estimates. Keywords. Time series; periodically stationary; innovations algorithm. AMS 1991 subject classifications. Primary 62M10, 62E20; secondary 60E07, 60F05. 1. INTRODUCTION A stochastic process Xt is called periodically stationary (in the wide sense) if lt ¼ EXt and ct(h) ¼ EXtXtþh for h ¼ 0, ±1, ±2, . . . are all periodic functions of time t with the same period m 1. If m ¼ 1, then the process is stationary. Periodically stationary processes manifest themselves in such fields as economics, hydrology and geophysics, where the observed time series are characterized by seasonal variations in both the mean and covariance structure. An important class of stochastic models for describing periodically stationary time series are the periodic ARMA models, in which the model parameters are allowed to vary with the season. Periodic ARMA models are developed in Adams and Goodwin (1995), Anderson and Vecchia (1993), Anderson and Meerschaert (1997, 1998), Anderson et al. (1999), Jones and Brelsford (1967), Lund and Basawa (2000), Pagano (1978), Salas et al. (1985), Tjostheim and Paulsen (1982), Troutman (1979), Vecchia and Ballerini (1991) and Ula (1993). Anderson et al. (1999) develop the innovations algorithm for periodic ARMA model parameters. In this paper, we provide the asymptotic estimates necessary to determine which of these estimates are statistically different from zero, under the classical assumption that the innovations have finite fourth moment. Brockwell and Davis (1988), discuss asymptotics of the innovations algorithm for stationary time series, using results of Berk (1974) and Bhansali (1978). Our results reduce to theirs, when the period m ¼ 1. Since our technical approach extends that of Brockwell and Davis (1988), we also need to develop periodically stationary 0143-9782/05/04 489–518 JOURNAL OF TIME SERIES ANALYSIS Vol. 26, No. 4 Ó 2005 Blackwell Publishing Ltd., 9600 Garsington Road, Oxford OX4 2DQ, UK and 350 Main Street, Malden, MA 02148, USA. 490 P. L. ANDERSON AND M. M. MEERSCHAERT analogues of results in Berk (1974) and Bhansali (1978). In particular, we obtain asymptotics for the Yule–Walker estimates of a periodically stationary process. Although the innovations estimates are more useful in practice, the asymptotics of the Yule–Walker estimates are also of some independent interest. 2. THE INNOVATIONS ALGORITHM FOR PARMA PROCESSES The periodic ARMA process fX~t g with period m [denoted by PARMAm(p, q)] has representation Xt p X /t ðjÞXtj ¼ et j¼1 q X ht ðjÞetj ; ð1Þ j¼1 where Xt ¼ X~t lt and fetg is a sequence of random variables with mean zero and SD rt such that fr1 t et g is i.i.d. The autoregressive (AR) parameters /t(j), the moving average parameters ht(j), and the residual SDs rt are all periodic functions of t with the same period m 1. In this paper, we will make the classical assumption Ee4t < 1, which leads to normal asymptotics for the parameter estimates. We also assume that the model admits a causal representation Xt ¼ 1 X wt ðjÞetj ; ð2Þ j¼0 where wt(0) ¼ 1 and condition P1 j¼0 jwt ðjÞj < 1 for all t, and satisfies an invertibility et ¼ 1 X pt ðjÞXtj ; ð3Þ j¼0 P where pt(0) ¼ 1 and 1 j¼0 jpt ðjÞj < 1 for all t. ðiÞ Let X^iþk ¼ PHk;i Xiþk denote the one-step predictors, where Hk,i ¼ spfXi, . . ., Xiþk1g, k 1, and PHk,i is the orthogonal projection onto this space, which minimizes the mean squared error (MSE) ðiÞ ðiÞ vk;i ¼ kXiþk X^iþk k2 ¼ EðXiþk X^iþk Þ2 : Then ðiÞ ðiÞ ðiÞ X^iþk ¼ /k;1 Xiþk1 þ þ /k;k Xi ; ðiÞ ðiÞ k 1; ð4Þ ðiÞ where the vector of coefficients /k ¼ ð/k;1 ; . . . ; /k;k Þ0 solves the prediction equation ðiÞ ðiÞ Ck;i /k ¼ ck Ó Blackwell Publishing Ltd 2005 ð5Þ 491 PARAMETER ESTIMATION ðiÞ with ck ¼ ðciþk1 ð1Þ; ciþk2 ð2Þ; . . . ; ci ðkÞÞ0 and Ck;i ¼ ciþk‘ ð‘ mÞ ð6Þ ‘;m¼1;...;k is the covariance matrix of (Xiþk1, . . . , Xi)0 for each i ¼ 0, . . . , m 1. Let ^ci ð‘Þ ¼ N 1 N 1 X ð7Þ Xjmþi Xjmþiþ‘ j¼0 denote the (uncentered) sample autocovariance, where Xt ¼ X~t lt . If we replace the autocovariances in the prediction equation (5) with their corresponding ðiÞ ðiÞ sample autocovariances, we obtain the estimator /^k;j of /k;j . Because the process is nonstationary, the Durbin–Watson algorithm for ðiÞ computing /^k;j does not apply. However, the innovations algorithm still applies to a nonstationary process. Writing ðiÞ X^iþk ¼ k X ðiÞ ðiÞ hk;j ðXiþkj X^iþkj Þ ð8Þ j¼1 ðiÞ yields the one-step predictors in terms of the innovations Xiþkj X^iþkj . Proposition 4.1 of Lund and Basawa (2000), shows that if r2i > 0 for i ¼ 0, . . . , m 1, then for a causal PARMAm(p, q) process the covariance matrix Ck,i is nonsingular for every k 1 and each i. Anderson et al. (1999) shows that if EXt ¼ 0 and Ck,i is nonsingular for each k 1, then the one-step predictors X^iþk , k 0, and their mean-square errors (MSEs)vk,i, k 1, are given by v0;i ¼ ci ð0Þ; ‘1 X ðiÞ ðiÞ ðiÞ h‘;‘j hk;kj vj;i ; hk;k‘ ¼ ðv‘;i Þ1 ciþ‘ ðk ‘Þ ð9Þ j¼0 vk;i ¼ ciþk ð0Þ k1 X ðiÞ ðhk;kj Þ2 vj;i ; j¼0 ðiÞ ðiÞ ðiÞ ðiÞ ðiÞ ðiÞ where (9) is solved in the order v0,i, h1;1 , v1,i, h2;2 , h2;1 , v2,i, h3;3 , h3;2 , h3;1 , v3,i, . . . The results in Anderson et al. (1999) show that ðhikiÞ hk;j ! wi ðjÞ; vk;hiki ! r2i ; ðhikiÞ /k;j ð10Þ ! pi ðjÞ as k ! 1 for all i, j, where j m½j=m if j ¼ 0; 1; . . ., hji ¼ m þ j m½j=m þ 1 if j ¼ 1; 2; . . . and [Æ] is the greatest integer function. Ó Blackwell Publishing Ltd 2005 492 P. L. ANDERSON AND M. M. MEERSCHAERT If we replace the autocovariances in (9) with the corresponding sample ðiÞ autocovariances (7), we obtain the innovations estimates h^k;‘ and ^vk;i . Similarly, replacing the autocovariances in (5) with the corresponding sample ðiÞ autocovariances yields the Yule–Walker estimators /^k;‘ . The consistency of these estimators was also established in Anderson et al. (1999). Suppose that fXtg is the mean zero PARMA process with period m given by (1) and that Eðe4t Þ < 1. Assume that the spectral density matrix f(k) of the equivalent vector ARMA process is such that mz0 z z0 f(k)z Mz0 z, p k p, for some m and M such that 0 < m M < 1 and for all z in Rm. If k is chosen as a function of the sample size N, so that k2/N ! 0 as N ! 1 and k ! 1, then the results in Anderson et al. (1999) also show that ðhikiÞ P ^ hk;j ! wi ðjÞ; P ^vk;hiki ! r2i ; ð11Þ P ^ ðhikiÞ ! pi ðjÞ / k;j for all i, j. This yields a practical method for estimating the model parameters, in the classical case of finite fourth moments. The results of Section 3, can then be used to determine which of these model parameters are statistically significantly different from zero. 3. ASYMPTOTIC RESULTS We compute the asymptotic distribution for the innovations estimates of the parameters in a periodically stationary time series (2) with period m 1. In the process, we also obtain the asymptotic distribution of the Yule–Walker estimates. For any periodically stationary time series, we can construct an equivalent (stationary) vector moving average process in the following way: Let Zt ¼ (etm, . . . , e(tþ1)m1)0 and Yt ¼ (Xtm, . . . , X(tþ1)m1)0 , so that Yt ¼ 1 X Wj Ztj ; ð12Þ j¼1 where Wj is the m m matrix with i‘ entry wi(tm þ i ‘), and we number the rows and columns 0, 1, . . . , m 1 for ease of notation. Also, let N(m, C) denote a Gaussian random vector with mean m and covariance matrix C, and let ) indicate convergence in distribution. Our first result gives the asymptotics of the Yule–Walker estimates. A similar result was obtained by Lewis and Reinsel (1985) for vector AR models, however the prediction problem here is different. For example, suppose that (2) represents monthly data with m ¼ 12. For a periodically stationary model, the prediction equations (4) use observations for earlier months in the same year. For the equivalent vector moving average model, the prediction equations use only observations from past years. Ó Blackwell Publishing Ltd 2005 493 PARAMETER ESTIMATION Theorem 1. Suppose that the periodically stationary moving average (2) is causal, invertible, Ee4t < 1, and that for some 0 < g G < 1 we have gz0 z z0 f(k)z Gz0 z for all p k p, and all z in Rm, where f(k) is the spectral density matrix of the equivalent vector moving average process (12). If k ¼ k(N) ! 1 as N ! 1 with k3/N ! 0 and 1 X N 1=2 jp‘ ðk þ jÞj ! 0 for ‘ ¼ 0; 1; . . . ; m 1 ð13Þ j¼1 then for any fixed positive integer D hiki N 1=2 pi ðuÞ þ /^k;u : 1 u D; i ¼ 0; . . . ; m 1 ) N ð0; KÞ; ð14Þ where K ¼ diagðr20 Kð0Þ ; r21 Kð1Þ ; . . . ; r2m1 Kðm1Þ Þ ð15Þ with ðKðiÞ Þu;v ¼ m1 X pimþs ðsÞpimþs ðs þ jv ujÞr2 imþs ð16Þ s¼0 and m ¼ min(u, v), 1 u, v D. In Theorem 1, note that K(i) is a D D matrix and the Dm-dimensional vector given in (14) is ordered h0ki h0ki hm1ki N 1=2 ðp0 ð1Þþ /^k;1 ;...;p0 ðDÞþ /^k;D ;...;pm1 ð1Þ þ /^k;1 hm1ki 0 ;...;pm1 ðDÞ þ /^k;D Þ: Next we present our main result, giving asymptotics for innovations estimates of a periodically stationary time series. Theorem 2. Suppose that the periodically stationary moving average (2) is causal, invertible, Ee4t < 1, and that for some 0 < g G < 1 we have gz0 z z0 f(k)z Gz0 z for all p k p, and all z in Rm, where f(k) is the spectral density matrix of the equivalent vector moving average process (12). If k ¼ k(N) ! 1 as N ! 1 with k3/N ! 0 and N 1=2 1 X jp‘ ðk þ jÞj ! 0 for ‘ ¼ 0; 1; . . . ; m 1 ð17Þ j¼1 then ðhikiÞ N 1=2 ðh^k;u wi ðuÞ : u ¼ 1; . . . ; D; i ¼ 0; . . . ; m 1Þ ) N ð0; V Þ; ð18Þ V ¼ A diagðr20 Dð0Þ ; . . . ; r2m1 Dðm1Þ ÞA0 ; ð19Þ where Ó Blackwell Publishing Ltd 2005 494 P. L. ANDERSON AND M. M. MEERSCHAERT A¼ D1 X En P½DmnðDþ1Þ ; ð20Þ n¼0 Dn n n zfflfflfflffl}|fflfflfflffl{ zfflfflfflfflfflfflfflfflfflfflfflfflffl}|fflfflfflfflfflfflfflfflfflfflfflfflffl{ zfflfflfflffl}|fflfflfflffl{ En ¼ diagð0; . . . ; 0; w0 ðnÞ; . . . ; w0 ðnÞ ; 0; . . . ; 0; w1 ðnÞ; . . . ; w1 ðnÞ ; . . . ; 0; . . . ; 0; wm1 ðnÞ; . . . ; wm1 ðnÞ Þ; |fflfflfflffl{zfflfflfflffl} |fflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflffl{zfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflffl} |fflfflfflfflfflfflfflfflfflfflfflfflffl{zfflfflfflfflfflfflfflfflfflfflfflfflffl} n Dn Dn 2 2 DðiÞ ¼ diagðr2 i1 ; ri2 ; . . . ; riD Þ and P an orthogonal Dm Dm cyclic permutation matrix, 1 0 0 1 0 0 0 B0 0 1 0 0C C B .. C B. . . . P ¼ B .. .. .. .. C: . C B @0 0 0 0 1A 1 0 0 ð21Þ 0 ð22Þ ð23Þ 0 Note that P0 is the Dm Dm identity matrix and P‘ (P0 )‘. Matrix multiplication yields Corollary 1. Corollary 1. Regarding Theorem 2, in particular, we have that ðhikiÞ wi ðuÞÞ ) N ð0; r2 N 1=2 ðh^k;u iu u1 X r2in w2i ðnÞÞ: ð24Þ n¼0 Remark. Corollary 1 also holds the asymptotic result for the second-order stationary process, where the period is just m ¼ 1. In this case r2i ¼ r2 so (24) becomes N 1=2 ðh^k;u wðuÞÞ ) N ð0; u1 X w2 ðnÞÞ; n¼0 which agrees with Theorem 2.1 in Brockwell and Davis (1988). 4. PROOFS The proof of Theorem 1 requires some preliminaries. First we show that (K(i))u,v is the limit of the u, v entry in the inverse covariance matrix (6). Lemma 1. Suppose that the periodically stationary moving average (2) is causal, invertible, and that Ee4t < 1. If Ck,i is given by (6) and (A)u,v denotes the u, v entry of the matrix A, then for any i ¼ 0, 1, . . ., v 1 we have Ó Blackwell Publishing Ltd 2005 495 PARAMETER ESTIMATION ðiÞ ðC1 k;hiki Þu;v ! ðK Þu;v ¼ m1 X pimþs ðsÞpimþs ðs þ jv ujÞr2 imþs ð25Þ s¼0 as k ! 1, where m ¼ min(u, v). Proof. The prediction equation (5) yield ðiÞ ðiÞ ciþ‘ ðk ‘Þ /k;1 ciþ‘ ðk 1 ‘Þ /k;k ciþ‘ ð‘Þ ¼ 0 ð26Þ ðiÞ for 0 ‘ k1 and since vk;i ¼ hXiþk ; Xiþk X^iþk i we also have ðiÞ ðiÞ vk;i ¼ ciþk ð0Þ /k;1 ciþk ð1Þ /k;k ciþk ðkÞ: ð27Þ Define 0 1 B B0 B F ¼B B ... B @ 0 /k1;1 ðiÞ /k1;2 ðiÞ 1 /k2;1 .. . ðiÞ 0 0 ðiÞ /k1;k1 1 C ðiÞ /k2;k2 C C C .. C . C ðiÞ /1;1 A 1 and use the fact that cr(s) ¼ crþs(s), so that (Ck,i)j,‘ ¼ ciþk‘(‘j), to compute that for 1 j < ‘ and 2 ‘ k the (j, ‘) element of the matrix FCk,i is ðiÞ ðiÞ ciþk‘ ð‘ jÞ /kj;1 ciþk‘ ð‘ j 1Þ /kj;kj ciþk‘ ð‘ kÞ: Substitute n0 ¼ k j and ‘0 ¼ k ‘ to obtain ðiÞ ðiÞ ðF Ck;i Þj;‘ ¼ ciþ‘0 ðn0 ‘0 Þ /n0 ;1 ciþ‘0 ðn0 ‘0 1Þ /n0 ;n0 ciþ‘0 ð‘0 Þ ¼ 0 in view of (26), so that FCk,i is lower triangular. Also (27) yields (FCk,i)j,j ¼ vkj,i for 1 j k. Since the transpose F0 is also lower triangular, the matrix FCk,iF 0 is still lower triangular, with (j, j) element vkj,i. But FCk,iF 0 is a symmetric matrix, 1 1 so it is diagonal. In fact, FCk,iF 0 ¼ H1, where H ¼ diagðv1 k1;i ; vk2;i ; . . . ; v0;i Þ. 1 0 Then Ck;i ¼ F HF , and a computation shows that ðC1 k;i Þu;v ¼ minðu;vÞ X ðiÞ ðiÞ /k‘;u‘ /k‘;v‘ v1 k‘;i ; ð28Þ ‘¼1 ðiÞ where the number of summands is fixed and finite for all k, and we set /k;0 1 to simplify the notation. Substitute hi ki for i in (28) and apply (10) using hi ki ¼ h(i ‘)(k ‘)i to obtain vk‘;hiki ! r2i‘ ; ðhikiÞ /k‘;u‘ ! pi‘ ðu ‘Þ; ðhikiÞ /k‘;v‘ ð29Þ ! pi‘ ðv ‘Þ Ó Blackwell Publishing Ltd 2005 496 P. L. ANDERSON AND M. M. MEERSCHAERT as k ! 1, so that ðC1 k;hiki Þu;v ! minðu;vÞ X pi‘ ðu ‘Þpi‘ ðv ‘Þr2 i‘ ‘¼1 as k ! 1. Substitute m ¼ min(u, v) and commute the first two terms if m ¼ v to finish the proof of Lemma 1. u In order to prove Theorem 1, it will be convenient to use a slightly different ðiÞ ðiÞ estimator fk;j for the coefficients /k;j of the one-step predictor (4), obtained by minimizing wk;i ¼ ðN kÞ1 NX k1 ðiÞ ðiÞ ðXjmþiþk fk;1 Xjmþiþk1 fk;k Xjmþi Þ2 ð30Þ j¼0 ðiÞ for i ¼ 0, . . . , v 1. For i fixed, take partials in (30) with respect to fk;‘ for each ‘ ¼ 1, . . . , k to obtain NX k1 ðiÞ ðiÞ ðXjmþiþk fk;1 Xjmþiþk1 fk;k Xjmþi ÞXjmþiþk‘ ¼ 0 j¼0 and rearrange to get ðiÞ ðiÞ ðiÞ ðiÞ ðiÞ fk;1^sk1;k‘ þ þ fk;k ^s0;k‘ ¼ ^sk;k‘ ð31Þ for each ‘ ¼ 1, . . . , k, where 1 ^sðiÞ m;n ¼ ðN kÞ NX k1 Xjmþiþm Xjmþiþn : j¼0 ðiÞ ðiÞ ðiÞ ðiÞ ðiÞ ðiÞ Now define ^rk ¼ ð^sk;k1 ; . . . ; ^sk;0 Þ0 and fk ¼ ðfk;1 ; . . . ; fk;k Þ0 and let 0 1 ðiÞ ðiÞ ^sk1;k1 ^s0;k1 B C .. C .. ^ ðiÞ ¼ B R k @ . A . ðiÞ ðiÞ ^sk1;0 ^s0;0 so that (31) becomes ^ ðiÞ f ðiÞ ¼ ^rðiÞ R k k k ð32Þ analogous to the prediction equations (5). Theorem 1 and Lemma 1 depend on modulo m arithmetic, which requires our hi ki-notation. Since Lemmas 2–8 do not have this dependence, we proceed with the less cumbersome i-notation. ðiÞ ðiÞ Lemma 2. Let pk ¼ ðpiþk ð1Þ; . . . ; piþk ðkÞÞ0 and Xj ðkÞ ¼ ðXðjkÞmþiþk1 ; . . . ; XðjkÞmþi Þ0 . Then for all i ¼ 0, . . ., m 1 and k 1 we have Ó Blackwell Publishing Ltd 2005 497 PARAMETER ESTIMATION ðiÞ ðiÞ ^ ðiÞ Þ1 pk þ fk ¼ ðR k N 1 1 X ðiÞ X ðkÞeðjkÞmþiþk;k ; N k j¼k j ð33Þ where et,k ¼ Xt þ pt(1)Xt1 þ þ pt(k)Xtk. Note that Proof. ^ ðiÞ ¼ R k N 1 1 X ðiÞ ðiÞ X ðkÞXj ðkÞ0 N k j¼k j ðiÞ and ^rk ¼ N1 1 X ðiÞ X ðkÞXðjkÞmþiþk N k j¼k j and apply (32) to obtain ðiÞ ðiÞ ðiÞ ðiÞ ðiÞ ^ Þ1^r pk þ fk ¼ pk þ ðR k hk i ðiÞ 1 ^ ðiÞ ðiÞ ðiÞ ^ ¼ ðRk Þ Rk pk þ ^rk ^ ðiÞ Þ1 ¼ ðR k ðiÞ ^ Þ1 ¼ ðR k N 1 1 X ðiÞ ðiÞ ðiÞ ðiÞ X ðkÞXj ðkÞ0 pk þ Xj ðkÞXðjkÞmþiþk N k j¼k j N 1 h i 1 X ðiÞ ðiÞ ðiÞ Xj ðkÞ Xj ðkÞ0 pk þ XðjkÞmþiþk ; N k j¼k u which is equivalent to (33). Lemma 3. For all i ¼ 0, . . . , m 1 and k 1 we have ðiÞ 0 ðiÞ wk;i ¼ ^riþk ð0Þ fk ^rk ; where ^ri ð0Þ ¼ ðN kÞ Proof. j¼k ð34Þ 2 XðjkÞmþi . The right-hand side of (34) equals ^riþk ð0Þ ¼ 1 PN 1 N 1 1 X ðiÞ 0 ðiÞ fk Xj ðkÞXðjkÞmþiþk N k j¼k N 1h i 1 X ðiÞ ðiÞ XðjkÞmþiþk fk;1 XðjkÞmþiþk1 fk;k XðjkÞmþi XðjkÞmþiþk ; N k j¼k ^ 2 , where we let Y ¼ (Xiþk, Xmþiþk,. . ., while wk;i ¼ ðN kÞ1 kY Yk 0 X(Nk1)mþiþk) , Xt ¼ (Xiþkt, Xmþiþkt, . . . , X(Nk1)mþiþkt)0 for t ¼ 1, . . . , k ^ ¼ f ðiÞ X1 þ þ f ðiÞ Xk . Since hY; ^ Y Yi ^ ¼ 0, we also have and Y k;1 k;k ^ Y Yi ^ ¼ hY Y ^ þ Y; ^ Y Yi ^ ¼ hY; Y Yi ^ ðN kÞwk;i ¼ hY Y; N 1h i X ðiÞ ðiÞ XðjkÞmþiþk fk;1 XðjkÞmþiþk1 fk;k XðjkÞmþi XðjkÞmþiþk ¼ j¼k and (34) follows easily. u Ó Blackwell Publishing Ltd 2005 498 P. L. ANDERSON AND M. M. MEERSCHAERT Lemma 4. For all i ¼ 0, . . . , m 1 and k 1 we have ðiÞ ðiÞ ðiÞ wk;i r2iþk ¼ ð^riþk ð0Þ ciþk ð0ÞÞ ðfk þ pk Þ0 ck ðiÞ 0 ðiÞ ðiÞ ðiÞ ðiÞ ðiÞ ðiÞ þ pk ð^rk ck Þ ðfk þ pk Þ0 ð^rk ck Þ 1 X piþk ðjÞciþkj ðjÞ; ð35Þ j¼kþ1 ðiÞ where ck ¼ ðciþk1 ð1Þ; ciþk2 ð2Þ; . . . ; ci ðkÞÞ0 . Proof. From (34) we have ðiÞ 0 ðiÞ wk;i r2iþk ¼ ^riþk ð0Þ fk ^rk r2iþk ðiÞ ðiÞ ðiÞ ¼ ð^riþk ð0Þ ciþk ð0ÞÞ ðfk þ pk Þ0 ck ðiÞ 0 ðiÞ ðiÞ ðiÞ ðiÞ ðiÞ ðiÞ þ pk ð^rk ck Þ ðfk þ pk Þ0 ð^rk ck Þ ðiÞ 0 ðiÞ þ ciþk ð0Þ þ pk ck r2iþk since the remaining terms cancel. Define X^t ¼ Xt et so that X^iþk ¼ Xiþk eiþk ¼ piþk ð1ÞXiþk1 piþk ð2ÞXiþk2 and r2iþk ¼ Eðe2iþk Þ ¼ E½ðXiþk X^iþk Þ2 . Since E½X^iþk ðXiþk X^iþk Þ ¼ 0 it follows that r2iþk ¼ E½Xiþk ðXiþk X^iþk Þ ¼ ciþk ð0Þþ piþk ð1Þciþk1 ð1Þ þ piþk ð2Þciþk2 ð2Þ þ and (35) follows easily. u Lemma 5. Let ci(‘), di(‘) for ‘ ¼ 0, 1, 2, . . . be arbitrary sequences of real numbers, let 1 1 X X ci ðkÞetmþik and vtmþj ¼ dj ðmÞetmþjm utmþi ¼ and set k¼0 Ci2 ¼ m¼0 1 X ci ðkÞ2 k¼0 for 0 i, j < m. Then 2 E4 M X and D2j ¼ 1 X dj ðmÞ2 m¼0 !2 3 utmþi vtmþj 5 4M 2 C 2 D2 g; ð36Þ t¼1 where C 2 ¼ maxðCi2 Þ, D2 ¼ maxðD2j Þ, g ¼ max(gt), and gt ¼ Eðe4t Þ. Proof. Since r1 t et are i.i.d., g ¼ max(gt: 1 < t < 1) ¼ max(gt: 0 i < m) < 1. The left-hand side of (36) consists of M2 terms of the form Eðui vj u‘mþi v‘mþj Þ ¼ 1 X 1 X 1 X 1 X ci ðkÞdj ðmÞci ðrÞdj ðsÞEðeik ejm e‘mþir e‘mþjs Þ: ð37Þ k¼0 m¼0 r¼0 s¼0 The terms in (37) with i k ¼ j m ¼ ‘m þ i r ¼ ‘m þ j s sum to Ó Blackwell Publishing Ltd 2005 499 PARAMETER ESTIMATION 1 X ci ðkÞdj ðk þ j iÞci ðk þ ‘mÞdj ðk þ ‘m þ j iÞEðe4ik Þ k¼0 g g 1 X k¼0 1 X ci ðkÞdj ðk þ j iÞci ðk þ ‘mÞdj ðk þ ‘m þ j iÞ jci ðkÞdj ðkÞj2 k¼0 g 1 X ci ðkÞ2 k¼0 1 X dj ðmÞ2 m¼0 gC 2 D2 P using the factP that for an, bn > 0 we always have ð an bn Þ2 P P maxðan Þ2 b2n a2n b2n . The terms in (37) with i k ¼ j m 6¼ ‘m þ i r ¼ ‘m þ j s sum to 1 X 1 X ci ðkÞdj ðk þ j iÞci ðrÞdj ðr þ j iÞEðe2ik e2‘mþir Þ k¼0 r¼0 g g g 1 X 1 X jci ðkÞdj ðk þ j iÞci ðrÞdj ðr þ j iÞj k¼0 r¼0 1 X jci ðkÞdj ðkÞj k¼0 1 X 1 X jci ðrÞdj ðrÞj r¼0 ci ðkÞ2 k¼0 1 X dj ðmÞ2 m¼0 gC 2 D2 qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi since Eðe2ik e2‘mþir Þ ¼ Eðe2ik ÞEðe2ir Þ Eðe4ik ÞEðe4ir Þ ¼ gik gir g by the Schwarz inequality. Similarly, the terms in (37) with i k ¼ ‘m þ i r 6¼ j m ¼ ‘m þ j s sum to 1 X 1 X ci ðkÞdj ðmÞci ð‘m þ kÞdj ð‘m þ mÞEðe2ik e2jm Þ gC 2 D2 k¼0 m¼0 and the terms in (37) with i k ¼ ‘m þ j s 6¼ j m ¼ ‘m þ i r sum to 1 X 1 X ci ðkÞdj ðmÞci ð‘m þ m þ i jÞdj ð‘m þ k þ j iÞEðe2ik e2jm Þ gC 2 D2 ; k¼0 m¼0 while the remaining terms are zero. Then E(uivju‘mþiv‘mþj) < 4gC2D2 and (36) follows immediately. u Ó Blackwell Publishing Ltd 2005 500 P. L. ANDERSON AND M. M. MEERSCHAERT Lemma 6. For et,k as in Lemma 2 and utmþi as in Lemma 5 we have 2 !2 3 N 1 1 X X E4 uðtkÞmþi ðeðtkÞmþ‘;k eðtkÞmþ‘ Þ 5 4gðN kÞ2 C 2 B max p‘ ðk þ jÞ2 ; ‘ t¼k where C2, g are as in Lemma 5 and B ¼ Proof. P P m1 1 i¼0 ‘¼0 jwi ð‘Þj 2 j¼1 ð38Þ . Write 1 X etmþ‘ etmþ‘;k ¼ ¼ ¼ p‘ ðmÞXtmþ‘m m¼kþ1 1 X p‘ ðmÞ m¼kþ1 1 X 1 X w‘m ðrÞetmþ‘mr r¼0 d‘;k ðk þ jÞetmþ‘kj ; j¼1 where d‘;k ðk þ jÞ ¼ j X p‘ ðk þ sÞw‘ðkþsÞ ðj sÞ: s¼1 Since fXtg is causal and invertible, 1 X d‘;k ðk þ jÞ ¼ j¼1 ¼ ¼ j 1 X X p‘ ðk þ sÞw‘ðkþsÞ ðj sÞ j¼1 s¼1 1 X p‘ ðk þ sÞ s¼1 1 X 1 X w‘ðkþsÞ ðj sÞ j¼s p‘ ðk þ sÞ s¼1 1 X w‘ðkþsÞ ðrÞ r¼0 P 2 is finite, and hence we also have 1 j¼1 d‘;k ðk þ jÞ < 1. Now apply Lemma 5 with vtmþ‘ ¼ etmþ‘,k etmþ‘ to see that 2 2 !2 3 !2 3 NX k1 N 1 X E4 utmþi ðetmþ‘;k etmþ‘ Þ 5 ¼ E4 uðtkÞmþi ðeðtkÞmþ‘;k eðtkÞmþ‘ Þ 5 t¼0 t¼k 4ðN kÞ2 C 2 D2‘;k g 4ðN kÞ2 C 2 D2k g; where D2k ¼ maxðD2‘;k : 0 ‘ < mÞ and D2‘;k ¼ Ó Blackwell Publishing Ltd 2005 P1 j¼1 d‘;k ðk þ jÞ2 . Next compute 501 PARAMETER ESTIMATION D2‘;k ¼ j 1 X X j¼1 ¼ ¼ j X p‘ ðk þ rÞw‘ðkþrÞ ðj rÞ r¼1 j1 1 X X j¼1 ! ! p‘ ðk þ sÞw‘ðkþsÞ ðj sÞ s¼1 ! j1 X p‘ ðk þ j pÞw‘ðkþjpÞ ðpÞ p¼0 ! p‘ ðk þ j qÞw‘ðkþjqÞ ðqÞ q¼0 1 X 1 X 1 X w‘ðkþjpÞ ðpÞw‘ðkþjqÞ ðqÞp‘ ðk þ j pÞp‘ ðk þ j qÞ p¼0 q¼0 j¼maxðp;qÞþ1 m1 X 1 X 1 X jwi ðpÞwiþqp ðqÞj i¼0 p¼0 q¼0 1 X jp‘ ðk þ j pÞp‘ ðk þ j qÞj; j¼maxðp;qÞþ1 where (without loss of generality we suppose p q) 1 X jp‘ ðk þ j pÞp‘ ðk þ j qÞj ¼ 1 X jp‘ ðk þ jÞp‘ ðk þ j þ p qÞj j¼1 j¼maxðp;qÞþ1 vffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi uX 1 X u1 p‘ ðk þ jÞ2 p‘ ðk þ j þ p qÞ2 t j¼1 1 X j¼1 p‘ ðk þ jÞ2 j¼1 and m1 X 1 X 1 X jwi ðpÞwiþqp ðqÞj i¼0 p¼0 q¼0 Then D2‘;k B P1 j¼1 m1 X 1 X !2 jwi ð‘Þj ¼ B: i¼0 ‘¼0 p‘ ðk þ jÞ2 and (38) follows easily. u ðiÞ Lemma 7. For et,k and Xj ðkÞ as in Lemma 2 we have for some real constant A > 0 that " 2 # N 1 1 X X ðiÞ 1 E Xj ðkÞðeðjkÞmþiþk;k eðjkÞmþiþk Þ p‘ ðk þ jÞ2 : ðN kÞ Ak max ‘ j¼k j¼1 ð39Þ Proof. Rewrite the left-hand side of (39) in the form 2 !2 3 k1 N 1 X X 2 ðN kÞ E4 XðtkÞmþiþs ðeðtkÞmþiþk;k eðtkÞmþiþk Þ 5 s¼0 t¼k and apply Lemma 6, k times with u(tk)mþi ¼ X(tk)mþiþs for each s ¼ 0, . . . , k 1 to obtain the upper bound of (39) with A ¼ 4gC2B. u Ó Blackwell Publishing Ltd 2005 502 P. L. ANDERSON AND M. M. MEERSCHAERT Lemma 8. Suppose that the periodically stationary moving average (2) is causal, invertible, Ee4t < 1, and that for some 0 < g G < 1 we have gz0 z z0 f(k)z Gz0 z for all p k p, and all z in Rm, where f(k) is the spectral density matrix of the equivalent vector moving average process (12). If k ¼ k(N) ! 1 as N ! 1 with k3/N ! 0 and (13) holds then ðiÞ ðiÞ ðN kÞ1=2 bðkÞ0 ðpk þ fk Þ ðN kÞ1=2 bðkÞ0 C1 k;i N 1 X P ðiÞ Xj ðkÞeðjkÞmþiþk ! 0 ð40Þ j¼k ðiÞ for any b(k) ¼ (bk1, . . . , bkk)0 such that kb(k)k2 remains bounded, where fk is from ðiÞ (32) and Xj ðkÞ is from Lemma 4.2. Using (33) the left-hand side of (40) can be written as " # N 1 N 1 X X ðiÞ 1 ðiÞ ðiÞ 1=2 0 1 ^ bðkÞ ðR Þ Xj ðkÞeðjkÞmþiþk;k C Xj ðkÞeðjkÞmþiþk ¼ I1 þ I2 ; ðN kÞ Proof. k;i k j¼k where j¼k " I1 ¼ ðN kÞ 1=2 # N 1 X ðiÞ ðiÞ 1 ^ Þ C1 bðkÞ ðR Xj ðkÞeðjkÞmþiþk;k ; k;i k 0 j¼k " I2 ¼ ðN kÞ 1=2 bðkÞ 0 C1 k;i N 1 X ðiÞ Xj ðkÞðeðjkÞmþiþk;k # eðjkÞmþiþk Þ ; j¼k so that jI1 j ðN kÞ 1=2 NX 1 ðiÞ 1 ðiÞ 1 ^ kbðkÞk kðRk Þ Ck;i k Xj ðkÞeðjkÞmþiþk;k ¼ J1 J2 J3 ; j¼k where J1 ¼ kb(k)k is bounded by assumption, N 1 ðN kÞ1=2 1 X ðiÞ Xj ðkÞeðjkÞmþiþk;k J3 ¼ 1=2 N k j¼k k ^ ðiÞ Þ1 C1 k ! 0 in probability by an argument similar to and J2 ¼ k 1=2 kðR k;i k Theorem 1 in Anderson et al. (1999). In fact, if we let ( ) 1 1 2 X X jwi ðm1 Þjjwj ðm2 Þj < 1; 0 i; j < m ; ð41Þ M ¼ max g 3 m1 ¼0 m2 ¼0 ^ ðiÞ Ck;i k then ðN kÞVarð^sðiÞ where g ¼ Eðe4t Þ and if we let Qk;i ¼ kR m;n Þ M in k general and hence Eðk 1=2 Q2k;i Þ k 3 M=ðN kÞ ! 0 as k ! 1 since k3/N ! 0 as N ! 1, and the remainder of the argument is exactly the same as Theorem 3.1 in Anderson et al. (1999). Using the inequality Var(X þ Y) 2(Var(X) þ Var(Y)) we obtain Ó Blackwell Publishing Ltd 2005 503 PARAMETER ESTIMATION 2 2 3 1 X N 1 N k ðiÞ EðJ32 Þ ¼ Xj ðkÞeðjkÞmþiþk;k 5 2ðT1 þ T2 Þ; E4 N k j¼k k where 2 N k 4 T1 ¼ E N k 2 N k 4 T2 ¼ E N k 2 3 N 1 1 X ðiÞ Xj ðkÞeðjkÞmþiþk 5 k j¼k 2 3 N 1 1 X ðiÞ Xj ðkÞðeðjkÞmþiþk;k eðjkÞmþiþk Þ 5: k j¼k Lemma 7 implies that T2 AðN kÞ max ‘ 1 X p‘ ðk þ jÞ2 j¼1 and 2 !2 3 k1 N 1 X X T1 ¼ k 1 ðN kÞ1 E4 XðjkÞmþiþt eðjkÞmþiþk 5 t¼0 1 1 ¼ k ðN kÞ k1 X " E t¼0 ¼ k 1 ðN kÞ1 j¼k N 1 X ! XðjkÞmþiþt eðjkÞmþiþk j¼k k1 X N 1 X N 1 X N 1 X !# XðrkÞmþiþt eðrkÞmþiþk r¼k EðXðjkÞmþiþt XðrkÞmþiþt eðjkÞmþiþk eðrkÞmþiþk Þ t¼0 j¼k r¼k ¼ k 1 ðN kÞ1 k1 X N 1 X 2 EðXðjkÞmþiþt ÞEðe2ðjkÞmþiþk Þ t¼0 r¼k ¼ k 1 k1 X ciþt ð0Þ ðN kÞ1 t¼0 N 1 X r2iþk r¼k so that T1 D ¼ maxi ðci ð0ÞÞ maxi ðr2i Þ. Hence EðJ32 Þ 2D þ 2AðN kÞ max ‘ P1 1 X p‘ ðk þ jÞ2 ; j¼1 2 where ðN kÞ max‘ j¼1 p‘ ðk þ jÞ ! 0 in view of (13), so that J3 is bounded in probability. Since J2 ! 0 in probability, it follows that I1 ! 0 in probability. ðiÞ Apply Lemma 6 with ‘ ¼ i þ k and uðtkÞmþi ¼ bðkÞ0 C1 k;i Xt ðkÞ to see that EðI22 Þ 4gðN kÞC 2 B max ‘ 1 X p‘ ðk þ jÞ2 ; j¼1 Ó Blackwell Publishing Ltd 2005 504 P. L. ANDERSON AND M. M. MEERSCHAERT P1 where C 2 ¼ maxi k¼0 ci ðkÞ2 comes from the representation u(tk)mþi ¼ P 2 kci(k)e(tk)mþik so that C is finite if and only if Var(u(tk)mþi) < 1. Since ðiÞ ðiÞ ðiÞ ðiÞ 0 0 Ck;i ¼ E½Xt ðkÞXt ðkÞ we have VarðuðtkÞmþi Þ ¼ bðkÞ0 C1 k;i E½Xt ðkÞXt ðkÞ 0 1 1 Ck;i bðkÞ ¼ bðkÞ Ck;i bðkÞ and Theorem A.1 in Anderson et al. (1999) shows that 1 2 kC1 k;i k ð2pgÞ , when the assumed spectral bounds hold, so C < 1. Then it follows from (13) that I2 ! 0 in probability, completing the proof of Lemma 8.u P Proof of Theorem 1. Let Xtm ¼ m j¼0 wt ðjÞetj and define eu(k) to be the k dimensional vector with 1 in the uth place and zero entries elsewhere. Let ðhikiÞ ðmÞ tNm;k ðuÞ ¼ ðN kÞ1=2 eu ðkÞ0 ðCk;hiki Þ1 N 1 X ðhikiÞ Xjm ðkÞeðjkÞmþhikiþk j¼k so that ðhikiÞ tNm;k ðuÞ ¼ ðN kÞ1=2 N 1 X ðhikiÞ ; ð42Þ ðkÞeðjkÞmþhikiþk ; ð43Þ wuj;k j¼k where ðhikiÞ wuj;k ðhikiÞ Xjm ðmÞ ðhikiÞ ¼ eu ðkÞ0 ðCk;hiki Þ1 Xjm ðkÞ ¼ ðXðjkÞmþhikiþk1;m ; . . . ; XðjkÞmþhiki;m Þ0 and ðmÞ ðhikiÞ Ck;hiki ¼ E½Xjm ðhikiÞ ðkÞXjm For each 0 r < k, X(jk)mþhikiþr,m is (e(jk)mþhikiþrs : 0 s m). Hence we write ðhikiÞ wuj;k a ðkÞ0 : linear ð44Þ combination ¼ LðeðjkÞmþhikim ; . . . ; eðjkÞmþhikiþk1 ÞeðjkÞmþhikiþk ; of ð45Þ where L(e1, . . . , eq) denotes a linear combination of e1, . . . , eq. Note that for i, u, k ðhikiÞ fixed, wuj;k are identically distributed for all j since the first two terms in (43) are nonrandom and do not depend on j, while the last two terms are identically distributed for all j by definition, using the fact that Xt is periodically strictly ðhikiÞ ðhikiÞ ðhi0 kiÞ stationary. Also, wuj;k are uncorrelated since E½wuj;k wvj0 ;k ¼ 0 unless (j k)m þ hi ki þ k ¼ (j0 k)m þ hi0 ki þ k, which requires i ¼ i0 mod m. ðhikiÞ ðhikiÞ Since 0 i < m, this implies that i ¼ i0 . Then E½wuj;k wvj0;k ¼ 0 if j 6¼ j0 and otherwise ðhikiÞ E½wuj;k ðhikiÞ wvj;k ðmÞ ¼ r2i eu ðkÞ0 ðCk;hiki Þ1 ev ðkÞ from (44). It follows immediately from (42) that the covariance matrix of the vector ðhikiÞ tNm;k ¼ ðtNm;k ðuÞ : 1 u D; 0 hi ki m 1Þ0 ðm1Þ 2 ð1Þ 2 is Cm ¼ diagðr20 Cð0Þ Þ, where m ; r1 Cm ; . . . ; rm1 Cm Ó Blackwell Publishing Ltd 2005 505 PARAMETER ESTIMATION ðmÞ 0 1 ðCðiÞ m Þu;v ¼ eu ðkÞ ðCk;hiki Þ ev ðkÞ and 1 u, v D. Note that ðh0kiÞ ðh0kiÞ ðhm1kiÞ tNm;k ¼ ðtNm;k ð1Þ; . . . ; tNm;k ðDÞ; . . . ; tNm;k ðhm1kiÞ ð1Þ; . . . ; tNm;k ðDÞÞ0 : We also have for any k 2 RDm that Var(k0 tNm,k) ¼ k0 Cmk. Apply Lemma 1 to the periodically stationary process Xtm to see that ðmÞ eu ðkÞ0 ðCk;hiki Þ1 ev ðkÞ ! ðKðiÞ m Þu;v as N ! 1, where ðKðiÞ m Þu;v ¼ minðu;vÞ1 X piminðu;vÞþs;m ðsÞpiminðu;vÞþs;m ðs þ jv ujÞr2 iminðu;vÞþs s¼0 and et ¼ P1 j¼0 pt;m ðjÞXtj;m (with pt,m(0) ¼ 1). Since Cm ! Km as k ! 1, then lim Varðk0 tNm;k Þ ¼ k0 Km k; N !1 where ðm1Þ 2 ð1Þ 2 Km ¼ diagðr20 Kð0Þ Þ: m ; r1 Km ; . . . ; rm1 Km Next, we want to use the Lindeberg–Lyapounov central limit theorem to show that k0 tNm;k ) N ð0; k0 Km kÞ: Towards this end, define the vector ðh0kiÞ wj;k ¼ k0 ðw1j;k ðh0kiÞ ; . . . ; wDj;k ðhm1kiÞ ; . . . ; w1j;k ðhm1kiÞ 0 ; . . . ; wDj;k Þ for any k 2 RDm so that k0 tNm;k ¼ ðN kÞ1=2 N 1 X wj;k : j¼k Then, Var(wj,k) ¼ Var(k0 tNm,k) ¼ k0 Cmk. Also, fwj,k : j ¼ k, . . . , N 1g are mean zero, identically distributed and uncorrelated. Now, let K ¼ [(k þ m)/m] þ 2, where again [Æ] is the greatest integer function and let N1 be an integer such that N1/(N K) ! 0 and K/N1 ! 0 (the largest integer less than ((N K)K)1/2 suffices). Let M1 be the greatest integer less than or equal to (N K)/N1, so that M1N1/N ! 1 and M1K/N ! 0. Define the random variables Ó Blackwell Publishing Ltd 2005 506 P. L. ANDERSON AND M. M. MEERSCHAERT 1=2 z1N ;k ¼ ðwk;k þ þ wkþN1 K1;k Þ=N1 ; 1=2 z2N ;k ¼ ðwkþN1 ;k þ þ wkþ2N1 K1;k Þ=N1 .. . 1=2 zM1 N ;k ¼ ðwkþðM1 1ÞN1 ;k þ þ wkþM1 N1 K1;k Þ=N1 : i.i.d. with mean zero and The choice of K ensures that z1N,k, . . . , zM1N, k are 1=2 PM1 variance s2M1 ¼ ððN1 KÞ=N1 Þk0 Cm k. Hence, M1 j¼1 zjN ;k is also mean zero with variance s2M1 , where s2M1 ! s2 as M1 ! 1 and s2 ¼ k0 Kmk. The Lindeberg– Lyapounov central limit theorem (see, e.g. Billingsley, 1968, Thm 7.3) implies that 1=2 M1 M1 X zjN ;k ) N ð0; s2 Þ ð46Þ j¼1 if we can show that s4 M1 M1 X 1=2 E ðM1 zjN ;k Þ4 ! 0 j¼1 as N ! 1. Letting N~1 ¼ N1 K, we have Eðz4jN ;k Þ ¼ N12 N~1 Eðw4j;k Þ þ ðN~1 2 N~1 ÞEðw2j;k w2‘;k Þ N12 N~1 Eðw4j;k Þ þ ðN~1 2 N~1 ÞEðw4j;k Þ Eðw4j;k Þ; where Eðw4j;k Þ < 1. Hence, s4 M1 M1 X 1=2 EððM1 1 4 zjN ;k Þ4 Þ s4 M1 M1 Eðwj;k Þ j¼1 s4 M11 Eðw4j;k Þ !0 as N ! 1 (hence N1 ! 1 and M1 ! 1). Thus, (46) holds. Also, 1=2 M1 M1 X j¼1 Ó Blackwell Publishing Ltd 2005 zjN ;k ðN kÞ1=2 N 1 X j¼k wj;k 507 PARAMETER ESTIMATION has limiting variance 0, since its variance is no more than ðM1 þ 1ÞKðVarðwt;k ÞÞ ðM1 þ 1ÞKðVarðwt;k ÞÞ ; N k N which approaches 0 as N ! 1. We can therefore conclude that k0 tNm;k ) N ð0; s2 Þ recalling that s2 ¼ k0 Kmk. Now, an application of the Cramer–Wold device yields tNm;k ) N ð0; Km Þ: ð47Þ Next we want to show that Km ! K as m ! 1. This is equivalent to showing ðiÞ that KðiÞ as m ! 1. Recall that wt(0) ¼ 1 and write m ! K 1 X Xt ¼ wt ð‘Þet‘ ‘¼0 ¼ 1 X wt ð‘Þ 1 r X X r¼1 so that pt ðrÞ ¼ Pr ‘¼1 Xtm ¼ pt‘ ðjÞXt‘j j¼0 ‘¼0 ¼ et þ 1 X ! wt ð‘Þpt‘ ðr ‘Þ Xtr ‘¼1 wt ð‘Þpt‘ ðr ‘Þ for r 1. Similarly m X wt ð‘Þ pt‘;m ðjÞXt‘j;m j¼0 ‘¼0 ¼ et þ 1 X 1 X minðr;mÞ X r¼1 ‘¼1 ! wt ð‘Þpt‘;m ðr ‘Þ Xtr;m Pminðr;mÞ so that pt;m ðrÞ ¼ ‘¼1 wt ð‘Þpt‘;m ðr ‘Þ for r 1. Apply these two formulae recursively to see that pt,m(r) ¼ pt(r) for 1 r m. Then we actually have ðiÞ KðiÞ for all m sufficiently large. m ¼ K Now we want to show that ðhikiÞ lim lim sup P ðjtN ;k m!1 N !1 ðhikiÞ ðuÞ tNm;k ðuÞj > dÞ ¼ 0 ð48Þ for any d > 0, where ðhikiÞ tN ;k ðuÞ ¼ ðN kÞ1=2 eu ðkÞ0 C1 k;hiki N 1 X ðhikiÞ Xj ðkÞeðjkÞmþhikiþk : j¼k Apply the Chebyshev inequality to see that the probability in (48) is bounded above by " # N 1 X ðhikiÞ ðhikiÞ 1=2 d2 VarðtN ;k ðuÞ tNm;k ðuÞÞ ¼ d2 Var ðN kÞ Aj eðjkÞmþhikiþk ; j¼k Ó Blackwell Publishing Ltd 2005 508 P. L. ANDERSON AND M. M. MEERSCHAERT ðhikiÞ ðmÞ ðhikiÞ where Aj ¼ eu ðkÞ0 C1 ðkÞ eu ðkÞ0 ðCk;hiki Þ1 Xjm k;hiki Xj mands are uncorrelated, we have ðhikiÞ VarðtN ;k ðhikiÞ ðuÞ tNm;k ðuÞÞ ¼ ðN kÞ1 ðkÞ. Since the sum- N 1 h i X E A2j e2ðjkÞmþhikiþk ; j¼k ¼ r2i ðN kÞ1 N 1 X EðA2j Þ j¼k ¼ r2i EðA2j Þ because each Aj has the same distribution. Write EðA2j Þ ¼ VarðI1 þ I2 Þ, where h i ðhikiÞ ðhikiÞ ðkÞ Xjm ðkÞ ; I1 ¼ eu ðkÞ0 C1 k;hiki Xj h i ðmÞ ðhikiÞ 1 ðC Þ ðkÞ; I2 ¼ eu ðkÞ0 C1 Xjm k;hiki k;hiki so that VarðI1 Þ ¼ eu ðkÞ0 C1 k;hiki Rk;hiki Ck;hiki eu ðkÞ; where Rk,hiki is the covariance matrix of the random vector (Xtmþhiki þ k1Xtmþhiki þ k1, m, . . . , Xtmþhiki Xtmþhiki, m)0 . Since keu(k)k ¼ 1, Theorem A.1 in Anderson et al. (1999) implies that Var(I1) (G/g)kRk,hikik. In order to show that kRk,hikik!0 as m ! 1, we consider the spectral density matrix fd (k) of the vector P moving average Wt ¼ (Xtmþm1Xtmþm1,m, . . . , XtmXtm,m)0 . If we let Yt ¼ 1 ‘¼0 W‘ Zt‘ , where Yt ¼ (Xtmþm1, . . . , Xtm)0 , Zt ¼ (etmþm1, . . . , etm)0 , and (W‘)ij ¼ wm1i(‘m i þ j) and 0 Y then Wt ¼ Yt Ytm. Since Xt Xtm ¼ Ptm ¼ (Xtmþm1,m, . . . , Xtm,m) , j>m wt(j)etj, the moving average 1 X Wt ¼ ~ ‘ Zt‘ ; W ‘¼dmþ2 m e1 where dxe is the smallest integer greater than or equal to x and ~ ‘Þ ¼ w except that we zero out the entries with ðW m1i ð‘m i þ jÞ ij ‘m i þ j m. Then the spectral density matrix 0 1 0 10 1 1 X 1 @ X ~ ‘ eik‘ AR@ ~ ‘ eik‘ A ; fd ðkÞ ¼ W W 2p mþ2 mþ2 ‘¼d m e1 ‘¼d m e1 where R is the covariance matrix of Zt. As in the proof of Theorem A.1 in Anderson et al. (1999), we now define Cw(h) ¼ Cov(Wt, Wtþh), W ¼ (Wn1, . . . , W0)0 and C ¼ CovðW ; W Þ ¼ ½Cw ði jÞni;j¼0 ¼ CovðXnm1 Xnm1;m ; . . . ; X0 X0;m Þ0 : For fixed i and k, let n ¼ [(k þ hi ki)/m] þ 1. Then Rk,hiki is a submatrix of C ¼ Rnm,0. Fix an arbitrary vector y ¼ (y0, . . . , yn1)0 in Rnm, whose jth entry yj 2 Rm. Then Ó Blackwell Publishing Ltd 2005 509 PARAMETER ESTIMATION y 0 Cy ¼ n1 X n1 X yj0 Cw ðj kÞyk j¼0 k¼0 ¼ n1 X n1 X yj0 Z j¼0 k¼0 ¼ Z p X n1 p Z eikðjkÞ fd ðkÞdk yk p eikj yj0 0 fd ðkÞ j¼0 p dm p p X n1 e ikk yk dk k¼0 X m1 yj0 yj dk ¼ 2pdm y 0 y j¼0 so that kCk 2pdm, where dm is the largest modulus of the elements of fd(k). But dm ! 0 in view of the causality condition (2), hence kRk,hikik kCk 2pdm ! 0, which implies that Var(I1) ! 0 as well. ðmÞ ðmÞ 1 ðmÞ 1 1 Next write VarðI2 Þ ¼ eu ðkÞ0 ½C1 k;hiki ðCk;hiki Þ Ck;hiki ½Ck;hiki ðCk;hiki Þ ðmÞ eu ðkÞ and recall that Ck;hiki is the covariance matrix of (Xtmþhiki þ k1,m, . . . , Xtmþhiki,m)0 for each t. Then as in the preceeding paragraph write Ytm ¼ ðXtmþm1;m ; . . . ; Xtm;m Þ ¼ ½ðm1Þ=mþ1 X ‘ Zt‘ ; W ‘¼0 ‘ Þ ¼ wm1i ð‘m i þ jÞ except that we zero out the entries with where ðW ij ‘mi þ j > m, and let fm(k) denote the spectral density matrix of Ytm. Let B(Zt) ¼ Zt1 denote the backward shift operator, and write Yt ¼ W(B)Zt 1 ik 0 ik and similarly Ytm ¼ WðBÞZ ) and fm ðkÞ ¼ t . Then f(k) ¼ (2p) W(e )RW (e 1 ik 0 ik ðe Þ, whereqasffiffiffiffiffiffiffiffiffiffiffiffiffiffi ð2pÞ Wðe ÞRW before R is the covariance matrix of Zt. Using ffi P 2 the Frobenius norm kAkF ¼ a , we have i;j ij ik ÞRW 0 ðeik Þk kf ðkÞ fm ðkÞkF ¼ ð4p2 Þ1 kWðeik ÞRW0 ðeik Þ Wðe F ik ÞRW0 ðeik Þk ð4p2 Þ1 kWðeik ÞRW0 ðeik Þ Wðe F ik ÞRW0 ðeik Þ Wðe ik ÞRW 0 ðeik Þk þ ð4p2 Þ1 kWðe F 0 ik 2 1 ik ik ð4p Þ kWðe Þ Wðe Þk kRk kW ðe Þk F F F 2 1 ik Þk kRk kW0 ðeik Þ W 0 ðeik Þk þ ð4p Þ kWðe F F F 0 ðeik Þk ! 0 as m ! 1 in view of the causality condition (2) where kW0 ðeik Þ W F and the remaining norms are bounded independent of m, so that |z0 [f(k) fm(k)]z| d(m)z0 z for any z 2 Rm, where d(m) ! 0 as m ! 1. Now Theorem A.1 in Anderson et al. (1999) yields z0 ðg dðmÞÞz z0 f ðkÞz z0 dðmÞz z0 fm ðkÞz z0 f ðkÞz þ z0 dðmÞz z0 ðG þ dðmÞÞz: Let C(h) ¼ Cov(Yt, Ytþh), Y ¼ (Yn1, . . . , Y0)0 and Cnm;0 ¼ CovðY ; Y Þ ¼ 0 ½Cði jÞn1 i;j¼0 , the covariance matrix of (Xnm1, . . . , X0) . For fixed i and k, let Ó Blackwell Publishing Ltd 2005 510 P. L. ANDERSON AND M. M. MEERSCHAERT n ¼ [(k þ hi ki)/m] þ 1 as before. Similarly, let Cm(h) ¼ Cov(Ytm, Ytþh,m), ðmÞ Ym ¼ (Yn1,m, . . . , Y0m)0 and Cnm;0 ¼ CovðYm ; Ym Þ, the covariance matrix of ðmÞ 0 (Xnm1,m, . . . , X0m) . Since Ck,hiki is a submatrix of Cnm,0 and Ck;hiki is a ðmÞ submatrix of Cnm;0 , it follows that ðmÞ ðmÞ kCk;hiki Ck;hiki k kCnm;0 Cnm;0 k ðmÞ ¼ sup jy 0 ðCnm;0 Cnm;0 Þyj kyk¼1 X n1 X n1 0 ¼ sup yj ðCm ðj ‘Þ Cðj ‘ÞÞy‘ kyk¼1 j¼0 ‘¼0 X n1 X n1 Z p 0 ikðj‘Þ ¼ sup yj e ðfm ðkÞ f ðkÞÞdk y‘ kyk¼1 j¼0 ‘¼0 p Z !0 ! n1 n1 p X X ikj ik‘ ¼ sup e yj ðfm ðkÞ f ðkÞÞ e y‘ dk kyk¼1 p j¼0 ‘¼0 is bounded above by 2pd(m), hence ðmÞ ðmÞ ðmÞ 1 1 1 kC1 k;hiki ðCk;hiki Þ k ¼ kCk;hiki ðCk;hiki Ck;hiki ÞðCk;hiki Þ k ðmÞ ðmÞ 1 kC1 k;hiki k kCk;hiki Ck;hiki k kðCk;hiki Þ k 1 1 2pdðmÞ : 2pg 2pðg dðmÞÞ Then VarðI2 Þ 1 1 2pdðmÞ 2pg 2pðg dðmÞÞ 2 2pðG þ dðmÞÞ ! 0 as m ! 1, and so Var(I1 þ I2) ! 0 as well, which proves (48). Together with ðiÞ (47), the fact that KðiÞ as m ! 1, and Theorem 4.2 in Billingsley (1968), m ! K this proves that tN ;k ) N ð0; KÞ ð49Þ as N ! 1, where ðhikiÞ tN;k ¼ ðtN;k ðuÞ : 1 u D; 0 hi ki m 1Þ0 : ðhikiÞ ðhikiÞ ðhikiÞ Applying Lemma 8 yields ðN kÞ1=2 eu ðkÞ0 ðpk þ fk Þ tN ;k ðuÞ ! 0 in ðhikiÞ ðhikiÞ ðhikiÞ probability. Note that pk ¼ ðpi ð1Þ; . . . ; pi ðkÞÞ0 and fk ¼ ðfk;1 ; . . . ; ðhikiÞ ðhikiÞ 0 ðhikiÞ fk;k Þ. Combining statement (49) with eu ðkÞ ðpk þ fk Þ¼ ðhikiÞ pi ðuÞ þ fk;u and the fact that (N k)/N ! 1, implies that Ó Blackwell Publishing Ltd 2005 511 PARAMETER ESTIMATION ðhikiÞ N 1=2 ðpi ðuÞ þ fk;u ðhikiÞ Þ tN;k P ðuÞ ! 0 ð50Þ ðhikiÞ hiki is defined as in (30). Then (14) holds with /^k;u as N ! 1, where fk;u hiki by fk;u . Finally, we must show that ðhikiÞ N 1=2 ðfk;u replaced ðhikiÞ P /^k;u Þ ! 0: ð51Þ Write ðhikiÞ N 1=2 kfk;u ðhikiÞ ðhikiÞ ^ k;hiki Þ1^r k ¼ N 1=2 kðR k /^k;u ðhikiÞ ^ 1 c^ C k;hiki k k ^ 1 k k^rðhikiÞ k ^ k;hiki Þ1 C N 1=2 kðR k;hiki k ^ 1 k k^rðhikiÞ c^ðhikiÞ k þ N 1=2 kC k;hiki k k ðhikiÞ and recall that ðN kÞVarð^sk;j Þ M, where M is given by (41), so that ðhikiÞ N 1=2 k^rk k is bounded in probability. Also, recall from the proof of Lemma 8 ^ 1 C1 k ! 0 in probability, so the same is true with i replaced with that k 1=2 kR k;i k;i 1 ^ 1 hiki. A very similar argument yields k 1=2 kC k;hiki Ck;hiki k ! 0 in probability, 1 1=2 ^ 1 ^ so that k kRk;hiki Ck;hiki k ! 0 in probability as well. Next observe that ^ 1 k kC ^ 1 kC C1 k þ kC1 k, where kC1 k is uniformly bounded k;hiki k;hiki k;hiki k;hiki k;hiki ðhikiÞ by Theorem A.1 of Anderson et al. (1999). Since ^sk;j and c^i ðj kÞ differ by only k terms and a factor (N k)/N, it is easy to check that ðhikiÞ ðhikiÞ ðhikiÞ N 1=2 k^rk c^k k ! 0 in probability. It follows that N 1=2 ðfk;u ðhikiÞ /^k;u Þ ! 0 in probability, which completes the proof of Theorem 1. u Proof of Theorem 2. it follows that ðiÞ From the two representations of X^iþk given by (4) and (8), ðhikiÞ hk;j ¼ j X ðhikiÞ ðhikiÞ hk‘;j‘ ð52Þ /k;‘ ‘¼1 ðhikiÞ for j ¼ 1, . . . , k if we define hkj;0 ¼ 1 and replace i with hi ki. Equation (52) can be modified and written as 0 ðhikiÞ hk;1 1 0 ðhikiÞ /k;1 1 B ðhikiÞ C B ðhikiÞ C B hk;2 C B /k;2 C C C B B B ðhikiÞ C ðhikiÞ B ðhikiÞ C B hk;3 C ¼ Rk B /k;3 C; C C B B B .. C B .. C @ . A @ . A ðhikiÞ ðhikiÞ hk;D /k;D ð53Þ where Ó Blackwell Publishing Ltd 2005 512 P. L. ANDERSON AND M. M. MEERSCHAERT 0 ðhikiÞ Rk 1 ðhikiÞ hk1;1 ðhikiÞ hk1;2 B B B ¼B B B @ .. . ðhikiÞ hk1;D1 0 1 ðhikiÞ hk2;1 .. . ðhikiÞ hk2;D2 0 0 0 0 1 0 .. . ðhikiÞ hkDþ1;1 1 0 0C C C 0C C .. C .A 1 ð54Þ ðiÞ ðiÞ for fixed lag D. From the definitions of h^k;u and /^k;u , we also have 1 0 ðhikiÞ 1 ðhikiÞ h^k;1 /^k;1 B ^ðhikiÞ C B ^ðhikiÞ C B hk;2 C B /k;2 C C C B B B ^ðhikiÞ C ^ ðhikiÞ B ^ðhikiÞ C B hk;3 C ¼ Rk B /k;3 C; C C B B B .. C B .. C @ . A @ . A ðhikiÞ ðhikiÞ h^ /^ 0 k;D where know ðhikiÞ h^ k‘;u ð55Þ k;D ^ ðhikiÞ is defined as in (54) with h^ðhikiÞ replacing hðhikiÞ . From (11) we R k k;u k;u ðhikiÞ P that h^k;u ! wi ðuÞ, hence for fixed ‘ with k0 ¼ k ‘, we have ðhi‘k 0 iÞ P ¼ h^ 0 ! w ðuÞ. i‘ k ;u Thus, P ^ ðhikiÞ ! R RðiÞ ; k ð56Þ where 0 B B RðiÞ ¼ B @ 1 wi1 ð1Þ .. . wi1 ðD 1Þ 0 1 .. . wi2 ðD 2Þ 0 0 .. . 1 0 0C C .. C: .A ð57Þ wiDþ1 ð1Þ 1 We have ^ ðhikiÞ ð/^ðhikiÞ /ðhikiÞ Þ þ ðR ^ ðhikiÞ RðhikiÞ Þ/ðhikiÞ ; h^ðhikiÞ hðhikiÞ ¼ R k k k ð58Þ ðhikiÞ ðhikiÞ ðhikiÞ ðhikiÞ where hðhikiÞ ¼ ðhk;1 ; . . . ; hk;D Þ0 , /ðhikiÞ ¼ ð/k;1 ; . . . ; /k;D Þ0 , and h^ðhikiÞ (hiki) (hiki) ðhikiÞ and /^ are the respective estimators of h and / . Note that ðhikiÞ ^ ðR k ðhikiÞ Rk ^ ðhikiÞ R ^ ðhikiÞ Þ/ðhikiÞ Þ/ðhikiÞ ¼ ðR k k ^ ðhikiÞ RðhikiÞ Þ/ðhikiÞ þ ðR k k ðhikiÞ þ ðRk where Ó Blackwell Publishing Ltd 2005 ðhikiÞ Rk Þ/ðhikiÞ ; ð59Þ 513 PARAMETER ESTIMATION 0 ðhikiÞ Rk 1 B hðhi1kiÞ B k;1 B ðhi1kiÞ h ¼B B k;2 .. B @ . ðhi1kiÞ hk;D1 0 1 ðhi2kiÞ hk;1 0 0 0 0 1 0 .. .. . ðhi2kiÞ hk;D2 1 . ðhiDþ1kiÞ hk;1 0 0C C C 0C C .. C .A 1 ð60Þ ^ ðhikiÞ is the corresponding matrix obtained by replacing hðhikiÞ with h^ðhikiÞ and R k k;u k;u for every season i and lag u. We next need to show that ðhikiÞ ðhikiÞ ^ ðhikiÞ R ^ ðhikiÞ ¼ oP ðN 1=2 Þ. This is equivalent Rk Rk ¼ oðN 1=2 Þ and R k k to showing that ðhi‘kiÞ N 1=2 ðhk;u ðhikiÞ hk‘;u Þ ! 0 ð61Þ ðhi‘kiÞ ðhikiÞ P N 1=2 ðh^k;u h^k‘;u Þ ! 0 ð62Þ and for ‘ ¼ 1, . . . , D 1 and u ¼ 1, . . . , D. Using estimates from the proof of Anderson et al. (1999, Cor. 2.2.4) and condition (13) of Theorem 1, it is not hard ðhikiÞ to show that N 1=2 ð/k;u þ pi ðuÞÞ ! 0 as N ! 1 for any u ¼ 1, . . . , k. This leads to ðhi‘kiÞ N 1=2 ð/k;u þ pi‘ ðuÞÞ ! 0 ð63Þ ðhi‘kiÞ by replacing i with i ‘ for fixed ‘. Letting ak ¼ N 1=2 ð/k;u þ pi‘ ðuÞÞ and ðhikiÞ bk ¼ N 1=2 ð/k‘;u þ pi‘ ðuÞÞ we see that bk ¼ ak‘. Since ak ! 0 then bk ! 0 as k ! 1. Hence, ðhikiÞ N 1=2 ð/k‘;u þ pi‘ ðuÞÞ ! 0 ð64Þ as k ! 1. Subtracting (64) from (63) yields ðhi‘kiÞ N 1=2 ð/k;u ðhikiÞ /k‘;u Þ ! 0; ð65Þ which holds for ‘ ¼ 1, . . . , u 1 and u ¼ 1, . . . , k. Since ðhi‘kiÞ hk;1 ðhikiÞ ðhi‘kiÞ hk‘;1 ¼ /k;1 ðhikiÞ /k‘;1 we have (61) with u ¼ 1. The cases u ¼ 2, . . . , D follow iteratively using (10), (53), and (65). Thus, (61) is established. To prove (62), we need Lemma 9. u Lemma 9. For all ‘ ¼ 1, . . . , u 1 and u ¼ 1, . . . , k, we have ðhi‘kiÞ N 1=2 ð/^k;u ðhikiÞ P /^k‘;u Þ ! 0: ð66Þ Ó Blackwell Publishing Ltd 2005 514 P. L. ANDERSON AND M. M. MEERSCHAERT Proof. where ðhi‘kiÞ Starting from (42), we need to show that tNm;k ðhi‘kiÞ tNm;k ðuÞ ¼ ðN kÞ1=2 N 1 X ðhikiÞ P ðuÞ tNm;k‘ ðuÞ ! 0, ðhi‘kiÞ wuj;k j¼k ðhikiÞ tNm;k‘ ðuÞ ¼ ðN k þ ‘Þ1=2 N 1 X ðhikiÞ wuj;k‘ j¼k‘ ðhi‘kiÞ wuj;k ¼ eu ðkÞ 0 ðmÞ ðhi‘kiÞ ðCk;hi‘ki Þ1 Xjm ðkÞeðjkÞmþhi‘kiþk ðhikiÞ ðmÞ ðhikiÞ wuj;k‘ ¼ eu ðk ‘Þ0 ðCk‘;hiki Þ1 Xjm ðk ‘Þeðjkþ‘Þmþhikiþk‘ and ðhi‘kiÞ Xjm ðhikiÞ Xjm ðkÞ ¼ ðXðjkÞmþhi‘kiþr;m : r ¼ 0; . . . ; k 1Þ0 ; ðk ‘Þ ¼ ðXðjkþ‘Þmþhikiþr;m : r ¼ 0; . . . ; k ‘ 1Þ0 ; ðmÞ ðhi‘kiÞ ðmÞ ðkÞ, and Ck‘;hiki is the covariance Ck;hi‘ki is the covariance matrix of Xjm ðhikiÞ ðhi‘kiÞ ðhikiÞ matrix of Xjm ðk ‘Þ. Note that Eðwuj;k wuj0 ;k‘ Þ ¼ 0 unless j0 ¼ j ‘ þ ðhi ‘ ki þ ‘ hi kiÞ=m; ð67Þ which is always an integer. For each j, there is at most one j0 that satisfies (67) for j0 2 fk ‘, . . . , N1g. If such a j0 exists then ðhi‘kiÞ Eðwuj;k ðhikiÞ ðmÞ ðmÞ wuj0 ;k‘ Þ ¼ r2i‘ eu ðkÞ0 ðCk;hi‘ki Þ1 CðCk‘;hiki Þ1 eu ðk ‘Þ; ðhi‘kiÞ ðhikiÞ where C ¼ EðXjm ðkÞXj0 m ðk ‘Þ0 Þ. Note that the (k ‘)-dimensional ðhikiÞ vector Xj0 m ðk ‘Þ is just the first (k‘) of the k entries of the vector ðhi‘kiÞ ðmÞ Xjm ðkÞ. Hence, the matrix C is just Ck;hi‘ki with the last ‘ columns deleted. ðmÞ 1 Then ðCk;hi‘ki Þ C ¼ Iðk; ‘Þ, which is the k k identity matrix with the last ‘ columns deleted. But then for any fixed u, for all k large, we have ðmÞ eu ðkÞ0 ðCk;hi‘ki Þ1 C ¼ eu ðk ‘Þ0 . Then ðhi‘kiÞ Eðwuj;k ðhikiÞ ðmÞ wuj0 ;k‘ Þ ¼ r2i‘ eu ðk ‘Þ0 ðCk‘;hiki Þ1 eu ðk ‘Þ ðmÞ ¼ r2i‘ ðCk‘;hiki Þ1 uu : Consequently, ðhi‘kiÞ Varðwuj;k ðhikiÞ ðhi‘kiÞ wuj0 ;k‘ Þ ¼ E½ðwuj;k ðhikiÞ wuj0 ;k‘ Þ2 ðmÞ ðmÞ 1 2 ¼ r2i‘ ðCk;hi‘ki Þ1 uu þ ri‘ ðCk‘;hiki Þuu ðmÞ 2r2i‘ ðCk‘;hiki Þ1 uu ðmÞ ðmÞ 1 ¼ r2i‘ ½ðCk;hi‘ki Þ1 uu ðCk‘;hiki Þuu !0 Ó Blackwell Publishing Ltd 2005 515 PARAMETER ESTIMATION by Lemma 1 applied to the finite moving average. Thus, ðhi‘kiÞ VarðtNm;k ðhikiÞ ðhi‘kiÞ ðuÞ tNm;k‘ ðuÞÞ ðN kÞ1 Varðwuj;k ðhikiÞ wuj0 ;k‘ ÞðN kÞ !0 since for each j ¼ k, . . . , N 1 there is at most one j0 satisfying (67) along with ðhi‘kiÞ j0 ¼ k ‘, . . . , N 1. Then Chebyshev’s inequality shows that tNm;k ðuÞ ðhikiÞ P tNm;k‘ ðuÞ ! 0. P ðhikiÞ tN ;k‘ ðuÞ ! 0. Since ðhikiÞ ðhikiÞ tNm;k ðuÞ tN ;k P ðuÞ ! 0 we also get ðhikiÞ tN ;k ðuÞ Now the Lemma follows easily using (50) along with (51). Now, since ðhi‘kiÞ ðhikiÞ ðhi‘kiÞ ðhikiÞ h^k;1 h^k‘;1 ¼ /^k;1 /^k‘;1 we have (62) with u ¼ 1. The cases u ¼ 2, . . . , D follow iteratively using (55), ^ /. ^ Thus, (62) is established. From (58) Lemma 9 and (52) with h, / replaced by h; to (62), it follows that ^ ðhikiÞ ð/^ðhikiÞ /ðhikiÞ Þ h^ðhikiÞ hðhikiÞ ¼ R k ^ ðhikiÞ RðhikiÞ Þ/ðhikiÞ þ oP ðN 1=2 Þ: þ ðR k k ð68Þ To accommodate the derivation of the asymptotic distribution of h^ðiÞ hðiÞ , we need to rewrite (68). Define ðh0kiÞ ðh0kiÞ ðh0kiÞ ðh0kiÞ h^ h ¼ðh^k;1 hk;1 ; . . . ; h^k;D hk;D ; . . . ; ðhm1kiÞ ðhm1kiÞ ðhm1kiÞ ðhm1kiÞ 0 h^k;1 hk;1 ; . . . ; h^k;D hk;D Þ ð69Þ and ðh0kiÞ / ¼ð/k;1 ðh0kiÞ ; . . . ; /k;D ðhm1kiÞ ; . . . ; /k;1 ðhm1kiÞ 0 ; . . . ; /k;D Þ: Using (69) we can rewrite (68) as ^ k ð/^ /Þ þ ðR ^ R Þ/ þ oP ðN 1=2 Þ; h^ h ¼ R k k ð70Þ where /^ is the estimator of / and ðh0kiÞ Rk ¼ diagðRk ðh1kiÞ ; Rk ðhm1kiÞ ; . . . ; Rk Þ and ðh0kiÞ Rk ¼ diagðRk ðh1kiÞ ; Rk ðhm1kiÞ ; . . . ; Rk Þ noting that both Rk and Rk are Dm Dm matrices. The estimators of Rk and Rk are ^ k and R ^ . Now write ðR ^ R Þ/ ¼ Ck ðh^ hÞ, where respectively R k k k Ck ¼ D1 X Bn;k P½DmnðDþ1Þ ð71Þ n¼1 Ó Blackwell Publishing Ltd 2005 516 P. L. ANDERSON AND M. M. MEERSCHAERT and Dn Bn;k n n zfflfflfflffl}|fflfflfflffl{ zfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflffl}|fflfflfflfflfflfflfflfflfflfflfflfflfflfflfflffl{ zfflfflfflffl}|fflfflfflffl{ ðh0kiÞ ðh0kiÞ ¼diagð0; . . . ; 0 ; /k;n ; . . . ; /k;n ; 0; . . . ; 0; ðh1kiÞ ðh1kiÞ ðhm1kiÞ ðhm1kiÞ /k;n ; . . . ; /k;n ; . . . ; 0; . . . ; 0; /k;n ; . . . ; /k;n Þ |fflfflfflffl{zfflfflfflffl} |fflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflffl{zfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflffl} |fflfflfflfflfflfflfflfflfflfflfflfflfflfflfflffl{zfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflffl} n Dn Dn with P the orthogonal Dm Dm cyclic permutation matrix (23). Thus, we write equation (70) as ^ k ð/^ /Þ þ Ck ðh^ hÞ þ oP ðN 1=2 Þ: h^ h ¼ R ð72Þ Then, ^ k ð/^ /Þ þ oP ðN 1=2 Þ ðI Ck Þðh^ hÞ ¼ R so that ^ k ð/^ /Þ þ oP ðN 1=2 Þ: h^ h ¼ ðI Ck Þ1 R ðhikiÞ Let C ¼ limk!1Ck so that C is Ck with /k;u R ¼ limk!1Rk, where ð73Þ replaced with pi(u). Also, let R ¼ diagðRð0Þ ; . . . ; Rðm1Þ Þ P ^k ! and R(i) as defined in (57). Equation (56) shows that R R and then Theorem 1 along with equation (73) yield N 1=2 ðh^ hÞ ) N ð0; V Þ; where V ¼ ðI CÞ1 RKR0 ½ðI CÞ1 0 ð74Þ and K is as in (15). Let 0 1 pi1 ð1Þ .. . B B S ðiÞ ¼ B @ 0 1 .. . pi1 ðD 1Þ pi2 ðD 2Þ 0 0 .. . 1 0 0C C .. C: .A ð75Þ piDþ1 ð1Þ 1 It can be shown that 0 2 2 ðiÞ KðiÞ ¼ S ðiÞ diagðr2 i1 ; ri2 ; . . . ; riD ÞS : P From the equation, wi ðuÞ ¼ u‘¼1 pi ð‘Þwi‘ ðu ‘Þ, it follows that R(i)S(i) ¼ IDD, the D D identity matrix. Therefore, Ó Blackwell Publishing Ltd 2005 517 PARAMETER ESTIMATION 0 0 0 2 2 ðiÞ ðiÞ RðiÞ KðiÞ RðiÞ ¼ RðiÞ S ðiÞ diagðr2 i1 ; ri2 ; . . . ; riD ÞS R 2 2 ¼ diagðr2 i1 ; ri2 ; . . . ; riD Þ and it immediately follows that RKR0 ¼ diagðr20 Dð0Þ ; . . . ; r2m1 Dðm1Þ Þ; 2 2 where DðiÞ ¼ diagðr2 i1 ; ri2 ; . . . ; riD Þ. Thus, eqn (74) becomes V ¼ ðI CÞ1 diagðr20 Dð0Þ ; . . . ; r2m1 Dðm1Þ Þ½ðI CÞ1 0 : Also, from the relation wi ðuÞ ¼ Pu ‘¼1 ðI CÞ1 ¼ ð76Þ pi ð‘Þwi‘ ðu ‘Þ, it can be shown that D1 X En P½DmnðDþ1Þ ; n¼0 where En is defined in (21). Using estimates from the proof of Corollary 2.2.3 in Anderson et al. (1999) along with condition (17) it is not hard to show that N1/2(w h) ! 0. Then it follows that N 1=2 ðh^ wÞ ) N ð0; V Þ; where w ¼ (w0(1), . . . , w0(D), . . . , wm1(1), . . . , wm1(D))0 . We have proved the theorem. u ACKNOWLEDGEMENTS Mark M. Meerschaert is partially supported by NSF Grants DES-9980484, DMS-0139927, DMS-0417869 and by the Marsden Foundation in New Zealand. REFERENCES Adams, G. and Goodwin, C. (1995) Parameter estimation for periodic ARMA models. Journal of Time Series Analysis 16, 127–45. Anderson, P. and Vecchia, A. (1993) Asymptotic results for periodic autoregressive moving–average processes. Journal of Time Series Analysis 14, 1–18. Anderson, P. and Meerschaert, M. (1997) Periodic moving averages of random variables with regularly varying tails. The Annals of Statistics 25, 771–85. Anderson, P. and Meerschaert M. (1998) Modeling river flows with heavy tails. Water Resources Research 34, 2271–80. Anderson, P., Meerschaert, M. and Vecchia, A. (1999) Innovations algorithm for periodically stationary time series. Stochastic Processes and Applications 83, 149–69. Berk, K. (1974) Consistent autoregressive spectral estimates. The Annals of Statistics 2, 489–502. Bhansali, R. (1978) Linear prediction by autoregressive model fitting in the time domain. The Annals of Statistics 6, 224–31. Ó Blackwell Publishing Ltd 2005 518 P. L. ANDERSON AND M. M. MEERSCHAERT Billingsley, P. (1968) Convergence of Probability Measures. New York: Wiley. Brockwell, P. and Davis, R. (1988) Simple consistent estimation of the coefficients of a linear filter. Stochastic Processes and Applications 28, 47–59. Jones, R. and Brelsford, W. (1967) Time series with periodic structure, Biometrika 54, 403–8. Lewis, R. and Reinsel G. (1985) Prediction of multivariate time series by autoregressive model fitting. Journal of Multivariate Analysis 16, 393–411. Lund, R. and Basawa, I. V. (2000) Recursive prediction and likelihood evaluation for periodic ARMA models. Journal of Time Series Analysis 21, 75–93. Pagano, M. (1978) On periodic and multiple autoregressions. The Annals of Statistics 6, 1310–7. Salas, J., Tabios, G. and Bartolini, P. (1985) Approaches to multivariate modeling of water resources time series. Water Research Bulletin 21, 683–708. Tjostheim, D. and Paulsen, J. (1982) Empirical identification of multiple time series. Journal Time Series Analysis 3, 265–82. Troutman, B. (1979) Some results in periodic autoregression. Biometrika, 6, 219–28. Ula, T. (1993) Forecasting of multivariate periodic autoregressive moving average processes. Journal of Time Series Analysis 14, 645. Vecchia, A. and Ballerini, R. (1991) Testing for periodic autocorrelations in seasonal time series data. Biometrika 78, 53–63. Ó Blackwell Publishing Ltd 2005