A FREQUENCY DOMAIN METHODOLOGY FOR TIME SERIES MODELING Hens Steehouwer OFRC Methodological paper No. 2008-02 May 2009 A FREQUENCY DOMAIN METHODOLOGY FOR TIME SERIES MODELING1 Hens Steehouwer2,3 Methodological Working Paper No. 2008-02 May 2009 Ortec Finance Research Center P.O. Box 4074, 3006 AB Rotterdam Boompjes 40, The Netherlands, www.ortec-finance.com ABSTRACT In this paper we have describe a frequency domain methodology for time series modeling. With this methodology it is possible to construct time series models that give a better description of the empirical long term behavior of economic and financial variables, bring together the empirical behavior of these variables at different horizons and observation frequencies and get insight in and understanding of the corresponding dynamic behavior. We introduce the most important frequency domain techniques and concepts, describe and illustrate the methodology and finally give the motivation for doing so. This methodology can contribute to a higher quality of investment decision making, implementation and monitoring in general and for Central Banks and Sovereign Wealth Managers in particular. 1 This paper has been published as Steehouwer, H. (2010), “A Frequency Domain Methodology for Time Series Modeling”, in Interest Rate Models, Asset Allocation and Quantitative Techniques for Central Banks and Sovereign Wealth Funds, edited by Berkelaar, A., J. Coche and K. Nyholm, Palgrave Macmillan. 2 The author is Head of the Ortec Finance Research Center and affiliated with the Econometric Institute of the Erasmus University Rotterdam. Please e-mail comments and questions to hens.steehouwer@ortec-finance.com 3 Copyright © 2011 Ortec Finance bv. All rights reserved. No part of this paper may be reproduced, in any form or by any means, without permission from the authors. Shorts sections may be quoted without permission provided that full credit is given to the source. The views expressed are those of the individual author(s) and do not necessarily reflect the views of Ortec Finance bv. 2 1 Introduction Determining an optimal Strategic Asset Allocation (SAA) in general, and for Central Banks and Sovereign Wealth Managers in particular, is essentially a decision making problem under uncertainty. How good or bad a selected SAA will perform in terms of the objectives and constraints of the stakeholders will depend on the future evolution of economic and financial variables such as interest rates, asset returns and inflation rates. The uncertainty about this future evolution of these variables is traditionally modeled by means of (econometric) time series models. The book by Campbell and Viceira (2002) is an example of this approach. They estimate Vector AutoRegressive (VAR) models on historical time series and derive optimal investment portfolios from the statistical behavior of the asset classes on various horizons as implied by the estimated VAR model. It is also known that the results from (SAA) models that take the statistical behavior of asset classes as implied by these time series models as input, can be very sensitive for the exact specifications of this statistical behavior in terms of for example the expected returns, volatilities, correlations, dynamics (auto- and cross-correlations) and higher order moments. Section 1.2 of Steehouwer (2005) describes an example of this sensitivity in the context SAA decision making for a pension fund. Besides the academic relevance, this observation also has an enormous practical impact since many financial institutions around the world base their actual (SAA) investment decisions on the outcomes of such models. Therefore, it is of great importance to continuously put the utmost effort in the development and testing of better time series models to be used for SAA decision making. This paper intends to make a contribution to such developments. If we now turn our attention to the methodological foundations of these time series models, virtually all model builders will agree that empirical (time series) data of economic and financial variables is (still) the primary source of information for constructing the models. This can already be seen from the simple fact that virtually all time series models are being estimated on historical time series data. On top of that, of course also forward looking information can be incorporated in the models. This is desirable if some aspects of the behavior observed in the (historical) time series data is considered to be inappropriate for describing the possible evolution of the economic and financial variables in the future. Furthermore, it is known that the empirical behavior of economic and financial variables is typically different at different horizons (centuries, decades, years, months, etc.) and different observation frequencies (annual, monthly, weekly, etc.). A first way of seeing this is by thinking about typical and well known economic phenomena such as long term trends, business cycles, seasonal patterns, stochastic volatilities, etc. For example, on a 30 year horizon with an annual observation frequency, long term trends and business cycles are important while on a 1 year horizon with a monthly observation frequency, seasonal patterns need to be taken into account and on a 1 month horizon with a daily observation frequency modeling stochastic volatility becomes a key issue. A second way of seeing the relevance of the horizon and observation frequency is by thinking about the so called ‘term structure of risk and return’ as described by Campbell and Viceira (2002). This simply means that expected returns, volatilities and correlations of and between asset classes are different at different horizons. For example, the correlation between equity returns and inflation rates is negative on short (for example 1 year) horizons while the same correlation is positive on long (for example 25 year) horizons. If we now combine these observations with the described sensitivity of real world (investment) decision making, for the statistical behavior of the time series models, we see that a first important issue in time series modeling for (investment) decision making is how to describe the relevant 3/40 empirical behavior as best as possible for the specific problem at hand. So if, for example, we are modeling for the purpose of long term SAA decision making, what does empirical data tell us about the statistical properties of long term trends and business cycles and how can we model this correctly? A second important issue follows from the fact that the results from time series models are also used in more steps of an investment process than ‘just to determine the SAA. Once the core SAA has been set for the long run, other (time series) models may be used for a medium term horizon in order to further refine the actual asset allocation (also called portfolio construction), for example by including more specialized asset classes or working with specific views that induce timing decisions. Once an investment portfolio is implemented, also monitoring and risk management are required, for example to see if the portfolio continues to satisfy (ex ante) short term risk budgets. Because, as mentioned before, the empirical behavior of economic and financial variables is different at different horizons and observation frequencies, it is typically the case that different (parametric and non-parametric) time series models are used in these various steps of an investment process. That is, the best model is used for the specific problem at hand. This is fine in itself, provided that the different steps in the investment process do not need to communicate with each other which obviously they do need to do. The SAA is input for the tactical decision making and portfolio construction while the actual investment portfolio is input for the monitoring and risk management process. If in these different steps different time series models are used, portfolios that where good or even optimal in one step may no longer be good or optimal in the next step, just by switching from one time series model to another. It is not hard to imagine the problems that can occur because of such inconsistencies. A second important issue in time series modeling for (investment) decision making is therefore how to bring together the empirical behavior of economic and financial variables that is observed at different horizons and different observation frequencies in one complete and consistent modeling approach. In response to the two important issues described above, this paper puts forward a specific frequency domain methodology for time series modeling. We will argue and illustrate that by using this methodology it is possible to construct time series models that 1. Give a better description of the empirical long term behavior of economic and financial variables with the obvious relevance for long term SAA decision making. 2. Bring together the empirical behavior of these variables as observed at different horizons and observation frequencies which is required for constructing a consistent framework to be used in the different steps of an investment process. In addition, by using frequency domain techniques, the methodology supports 3. Better insight in and understanding of the dynamic behavior of economic and financial variables at different horizons and observation frequencies, both in terms of empirical time series data and of the time series models that are used to describe this behavior. The methodology combines conventional (time domain) time series modeling techniques with techniques from the frequency domain. It is fair to say that frequency domain techniques are not used very often in economics and finance, especially when compared to the extensive use of these techniques in the natural sciences. We explain this from the non-experimental character of the economic and finance sciences and the therefore in general limited amount of data that is available for analysis. We show how the corresponding problems of conventional frequency domain techniques can be solved by using appropriate special versions of these techniques that work well in the case of limited data. The methodology builds on the techniques and results described in Steehouwer (2005) as well as subsequent research. Its applications are not limited to that of (SAA) investment decisions as described above but cover in principle all other types of applications of time series models. 4/40 Furthermore, the methodology leaves room for and even intends to stimulate the inclusion and combining of many different types of time series modeling techniques. By this we mean that the methodology can accommodate and combine classical time series models, VAR models, London School of Economics (LSE) methodology, theoretical models, structural time series models, (G)ARCH models, Copula models, models with seasonal unit roots, historical simulation techniques, etc. in one consistent framework. This paper does not intend to give a full in-depth description of every aspect of the methodology. Instead, the objective of the paper is to give a rather high-level overview of the methodology and provide the appropriate references for further information. The remainder of the paper has the following setup. We proceed in Section 2 with an introduction to some basic concepts from the frequency domain together with what we feel is the fundamental reason why frequency domain techniques do not have the widespread use within economics and finance they deserve. In Section 3 we continue with a description of the proposed methodology followed by the main points of motivation for proposing this methodology in Section 4. This motivation consists of a combination of technical as well as more methodological issues. This paper does not include (yet) one complete example application of the described methodology. Instead, separate examples are given throughout the text to illustrate individual concepts. Section 5 closes the paper by summarizing the main conclusions. 5/40 2 Frequency Domain The methodology for time series analysis and modeling that we propose in this paper is based on concepts and techniques from the frequency domain, also known as spectral analysis techniques. Frequency domain techniques are not that well known and are not applied very often in economics and finance. Furthermore, these techniques require a rather different view on empirical time series data and stochastic processes. Therefore, this section briefly introduces some key concepts from frequency domain analysis such as spectral densities, frequency response functions and the leakage effect. We will argue that the latter is the key reason why frequency domain techniques are not used more often in economics and finance. These concepts are used in the description of the proposed methodology in Section 3. Those interested in a historical overview of the development of frequency domain and spectral analysis techniques are referred to Section 2.2.11 of Steehouwer (2005). Further details, proofs and references on the concepts discussed here can be found in Chapter 4 of the same reference. Well known classical works on spectral analysis techniques and time series analysis are Bloomfield (1976) and Brillinger (1981). 2.1 Frequency domain versus time domain All frequency domain techniques are built on the foundations of the Fourier transform. With the Fourier transform any time series {xt, t=0,…,T-1} can be written as a sum of cosine functions T −1 xt = ∑ R j cos(ω j t + φ j ) (2.1) j =0 The parameters {Rj, ωj and φj, j = 0,…,T-1} represent the amplitudes, frequencies and phases of the T cosine functions. The conventional representation {xt, t=0,…,T-1} of the time series is referred to as the representation in the time domain. The representation {Rj, ωj and φj, j = 0,…,T-1} is referred to as a representation in the frequency domain. An important property of this frequency domain representation is 1 T −1 2 T −1 2 Rj ∑ xt = ∑ T t =0 j =0 (2.2) So, if we assume the time series xt to have an average value of zero, then this relation tells us that the frequency domain representation decomposes the total variance of the time series into the squared amplitudes of the set of cosine functions. The higher the Rj for a certain ωj, the more this frequency contributes to the total variance of the time series. 2.2 Spectral densities A periodogram plots the variance per frequency from (2.2) as a function of the frequencies and thereby shows the relative importance of the different frequencies for the total variance of the time series. If one would calculate the periodogram for different samples from some stochastic time series process, this would result in different values and shapes of the periodogram. Doing this for a great number of samples of sufficient length and calculating the average periodogram on all these samples results in what is called the spectral density (or auto-spectrum) of a (univariate) stochastic process. A spectral density describes the expected distribution of the variance of the process over periodic fluctuations with a continuous range of frequencies. The word “spectrum” comes from the analogy of decomposing white light into colors with different wave lengths. The word “density” comes from the analogy with a probability density function. A probability density function describes the distribution of a 6/40 probability mass of 1 over some domain while a spectral density describes the distribution of a variance mass over a range of frequencies. It can be shown that the spectrum and the traditional autocovariances contain the same information about the dynamics of a stochastic process. Neither can give information that cannot be derived from the other. The only difference is the way of presenting the information. An auto-spectrum specifies the behavior of a univariate stochastic process. However, economic and financial variables need to be studied in a multivariate setting. The dynamic relations between variables are measured by the cross-covariances of a stochastic process. A cross-spectral density function (cross-spectrum) between two variables can be derived in the same way as the autospectrum for a single variable. The only difference is that the cross-covariances need to be used instead of the auto-covariances. In the form of the coherence and phase spectra, these cross-spectral densities ‘dissect’ the conventional correlations at the various frequencies into a phase shift and the maximum correlation possible after such a phase shift. Note that various auto- and cross-spectra can also be combined in a straightforward manner into multivariate spectral densities. 2.3 Filters If a linear filter G(L) is applied on a time series xt we obtain a new time series b b yt = ∑ g l xt −l = ∑ g l Ll xt = G ( L) xt l =a l =a (2.3) The Fourier transform of this filter is called the Frequency Response Function (FRF) of the filter because for each frequency ω it specifies how the amplitude and phase of the frequency domain representation of the original time series xt are affected by the filter. The effects of the filter can be split up into two parts. First, the squared Gain gives the multiplier change of the variance of the component with frequency ω in a time series. The squared Gain is therefore often called the Power Transfer Function (PTF). Second, the Phase of a linear filter gives the phase shift of the component of frequency ω in a time series, expressed as a fraction of the period length. Although it is often not recognized as such, probably the most often applied linear filter consists of calculating the first order differences of a time series. Its squared Gain (i.e. PTF) and Phase are shown in Figure 2.1. Figure 2.1: Squared Gain or Power Transfer Function (PTF) and Phase of the first order differencing operator. The PTF (left panel) shows how the first order differencing operator suppresses the variance of low frequency fluctuations in a time series while it strongly enhances the variance of high frequency fluctuations in a time series. The Phase (right panel) shows that that the first order differencing operator also shifts these fluctuations back in time by a maximum of 0.25 times the period length of the fluctuations. 7/40 Assume t is measured in years. The PTF on the left shows that the variance at frequencies below approximately 1/6 cycles per year (i.e. with a period length of more than six years) is being reduced by the first order differencing filter. This explains why the filter is often used to eliminate trending behavior (i.e. very long term and low frequency fluctuations) from time series data. First order differencing also (strongly) emphasizes the high frequency fluctuations. This can be seen from the value of the PTF for higher frequencies. The variance of the highest frequency fluctuations is multiplied by a factor 4. Besides changing the variance at the relevant frequencies, which can be directly thought of in terms of changing the shape of spectral densities, the first order differencing filter also shifts time series in the time domain which corresponds to phase shifts in the frequency domain. The Phase on the right shows that the lowest frequencies are shifted backwards in time by approximately a quarter of the relevant period length while the phase shift for the higher frequencies decreases towards zero for the highest frequency in a linear fashion. 2.4 Leakage Effect The previous sections already demonstrate some clear intuitive appealing properties of a frequency approach for time series analysis and modeling. Spectral densities very efficiently give information about the dynamic behavior of both univariate and multivariate time series and stochastic time series processes. Gains and Phases clearly show what linear filters do to spectral densities at different frequencies. Nevertheless, frequency domain techniques do not have the widespread use within economics and finance one would expect based on these appealing properties, especially given the extensive use of these techniques in the natural sciences. We feel that the fundamental reason for this lies in the fact that in economics and finance the available amount of (historical) time series data is in general too limited for conventional frequency domain techniques to be applied successfully, which in turn is caused by the fact that these techniques typically require large amounts of data. If these conventional techniques are applied anyway on time series of limited sample sizes this can for example result in disturbed and / or less informative spectral density estimates. Another example is that filtering time series according to some FRF can give disturbed filtering results. Fortunately, there exist special (parametric) versions of frequency domain techniques for the estimation of spectral densities and filtering of time series that are especially adapted to (also) work well on short sample time series data and can therefore be successfully applied on economic and financial time series data. These techniques avoid spurious spectral analysis and filtering results in ways that are described in Sections 3.1 and 3.2. The fundamental disturbing consequences of applying standard frequency domain techniques on time series of limited size are caused by what is called the leakage effect. This effect can best be understood by thinking of the Fourier transform of a perfect cosine function of some frequency. Obviously in the periodogram of this cosine function 100% of the variance should be located at the specific frequency of the cosine function. However, if one only has a limited sample of the cosine function available for the Fourier transform, this turns out not to be the case. Instead, a part of the variance at the specific frequency will have ‘leaked’ away to surrounding frequencies in the periodogram. As the sample size increases, the disturbing effects of leakage decrease and the periodogram gets better and better at revealing the true identity of the time series by putting a larger and larger portion of the variance at the specific frequency of the cosine function. In Section 3.1 we will explain how in small samples the leakage effect can result in disturbed and less informative spectral density estimates while in Section 3.2 we will explain how it can cause disturbed filtering results. In both cases we will also describe the appropriate solutions to these problems. 8/40 3 Methodology In this section we first describe the proposed frequency domain methodology for time series analysis and modeling. In Section 4 we will give various motivations for proposing this methodology. We use this aberrant order of presentation because knowing the methodology makes it easier to fully understand and appreciate its motivations. This means that in this section we deliberately say little about why the methodology works in the proposed way and instead limit ourselves to explaining how it works. The methodology consists of the consecutive steps described in the following sub sections and builds on the fundamental frequency domain concepts described in Section 2. Various examples are given throughout the text to illustrate individual concepts. These examples are taken from different sources and are therefore not necessarily consistent 3.1 Time series Decomposition After having collected the appropriate time series (and possibly also cross section) data, the first step of the methodology is to zoom in on the different aspects of the time series behavior by decomposing the time series. The different components of the time series can then be analyzed and modeled separately by zooming in on the behavior of the time series in the different frequency regions. As an example consider Figure 3.1 which shows a decomposition of the long term nominal interest rate in 4 the Netherlands in a trend, low frequency and high frequency component. This decomposition is such that the three components add up to the original time series. The trend component consists of all fluctuations in the time series with a period length longer than the sample length (194 years) which is very natural definition for a trend. The low frequency component consists of all fluctuations with a period length shorter than the sample length but longer than 15 years. 15 years is a wide upper bound on business cycle behavior. The third component consists of all fluctuations with a period length between 15 and 2 years, the shortest period length possible on annual data. Decomposing a time series in such a way is also called filtering the time series. It is not hard to imagine that the way this filtering is implemented is of crucial importance for the subsequent analysis and modeling. Therefore we proceed by giving more information on the available and required filtering techniques. An overview of filtering techniques and their properties can be found in Chapter 5 of Steehouwer (2005). 4 End of year values of Netherlands 10-year Government Bond Yield 1814 - 2007. 1918 and 1945 values based on beginning of next year. Source: GlobalFinancialDatabase (GFD) code IGNLD10D. 9/40 Figure 3.1: Decomposition of a long term interest rate time series for the Netherlands. Source: GlobalFinancialDatabase (GFD). The time series is decomposed into the three dotted lines in such a way that adding these component time series results in the original time series. The component time series distinct themselves by the period length of the fluctuations from which they are constituted. The first component indicated by ‘Trend’ captures all fluctuation in the interest rate time series with a period length between infinity (i.e. a constant term) and the sample length of 193 years. The second component indicated by ‘Low frequencies’ captures all fluctuations with a period length between 193 and 15 years. The third component indicated by ‘High frequencies’ captures all fluctuations with a period length between 15 and 2 years. 3.1.1 Filter Requirements What do we require from a filtering technique that is to be applied to decompose time series in a way as shown in Figure 3.1? An ideal filter should allow for / result in: 1. User defined pass-bands: By this we mean that the user of the filter should be able to freely specify which period lengths (frequencies) of fluctuations to include and exclude in the filtered time series and should not be restricted by the properties of the filter. 2. Ideal pass-bands: The filter should exactly implement the required user defined pass-bands. Often filters do this only in an approximating sense. 3. No phase shifts: The filter should not move time series back or forth in time as this could influence inferences about the lead / lag relations between variables. This means that the Phase of a filter must be zero for all frequencies. 4. No loss of data: The available amount of data in economics and finance is in general limited. Therefore, an ideal filter should not lose observations at the beginning or end of the sample. 3.1.2 Zero Phase Frequency Filter Many conventional filters fail one or more of the four described requirements. There are two main reasons for this. A first reason is that many filters were originally defined in the time domain instead of in the frequency domain in terms of the properties of their PTF and Phase. As an example consider the properties of the simple first order differencing filter described in Section 2.3 which fails on all of the four requirements. To a lesser, though still significant, extent this also holds for other well known filters such as the exponential smoothing and Hodrick Prescott filter. The second reason is the 10/40 leakage effect from Section 2.4 which causes that ideal filtering results cannot be achieved when filters are applied on time series of a limited sample size, even when the filters have been explicitly designed to achieve the required ideal frequency domain properties. To see how this works think a of a conceptually simple direct frequency filter that starts by transforming a time series into the frequency domain. Next, based on the required PTF of the filter, the weights at certain frequencies are set to zero while other are preserved (i.e. the ones that lie within the pass-band of the filter). And finally, the adjusted frequency domain representation of the time series is transferred back into the time domain. For time series of limited sample sizes this approach does not work well because, as explained in Section 2.4, the true frequency domain representation of the (limited sample) time series is disturbed in the first step. If, in the second step, the ideal PTF is applied to this erroneous frequency domain representation, some (frequency) parts of the time series behavior will be deleted that should have been preserved and vice versa. Section 5.4 of Steehouwer (2005) describes a special Zero Phase Frequency Filter that does meet all four requirements by focusing on a solution for the second problem of filtering time series of finite sample sizes as caused by the leakage effect. This filtering technique is very much based on the ideas from Bloomfield (1976) and the filtering approach of Schmidt (1984). For example Hassler et al. (1992), Baxter and King (1999) and Christiano and Fitzgerald (1999) describe different approaches to deal with the problems of filtering finite time series in the frequency domain. The filter algorithm comes down to the iterative estimation of a number of periodic components and multiplying each of these components by the value of the PTF to obtain the filtered time series. The key trick of this filter is that it avoids the disturbing leakage effects by skipping the transformation into the frequency domain but filtering estimated periodic components (sine and cosine functions) directly instead. This is possible because in case we would have had an infinite sample size, the filtering result of a periodic component of a certain frequency would be known exactly beforehand. As an example think of the frequency domain representation of the perfect cosine time series described in Section 2.4. In a sense, by estimating periodic components, the time series is “extrapolated” from the sample size into infinity and thereby the disturbing effects of leakage can be avoided. As shown in Section 6.3.2 of Steehouwer (2005), this Zero Phase Frequency Filter results in very similar filtering results as the popular Baxter and King and Christiano and Fitzgerald filters. The additional advantages are that compared to these filters, the filter is more precise by filtering directly in the frequency domain (requirement 2), it causes no phase shifts (requirement 3) and leads to no loss of data (requirement 4). 3.1.3 Zero Correlation Property An important advantage of decomposing time series in the frequency domain, based on ideal filters and implemented by appropriate filtering techniques, is that all filtered components of non-overlapping pass-bands of some (set of) time series have zero correlation in the time domain, both in a univariate and multivariate context. This (theoretical) property holds for “all” filters which adequately implement an ideal pass-band and is the continuous analogue of the orthogonal property of cosine functions. This theoretical property can also be checked to apply to practical filter output when the Zero Phase Frequency Filter is applied on actual time series. Tests described in Steehouwer (2007) show that although non-zero correlations can actually occur in practical filter output, it can still be concluded that from a theoretical perspective zero correlations between the component time series can (and must) safely be assumed for time series modeling purposes. This can easily be understood by thinking about short samples of several low frequency components. On short samples the correlations between such components can be very different from zero. However, if the time series behavior of these low frequency components would be modeled and simulated on sufficiently long horizons, the fundamental 11/40 correlations would still need to be zero. Although this zero correlation property greatly simplifies the time series modeling process, note that zero correlations do not need to imply that component time series are also independent. In fact, there exist quite complex forms of dependencies between the component time series which need to be taken into account in the modeling process. The perhaps counterintuitive fact is that, despite the zero correlation property, the decomposition approach does not hinder the analysis and modeling of such complex dependencies but actually facilitates it. We say more about these complex dependencies in Section 3.3.3. 3.2 Time series Analysis After having decomposed the time series in order to be able to zoom in on the behavior of the time series variables in the different frequency regions, the second step in the methodology consists of actually analyzing this behavior of the time series. The resulting understanding of this time series behavior then forms the basis for an adequate modeling of this time series behavior. For the trend and low frequency components, as for example shown in Figure 3.1, traditional time series analysis techniques (mean values, volatilities, correlations, cross correlations, etc.) typically suffice for obtaining the appropriate understanding. For higher frequency components, again as for example shown in Figure 3.1, spectral analysis techniques are very powerful for a further unraveling and understanding of the time series behavior. 3.2.1 Maximum Entropy Spectral Analysis There exist two groups of methods for estimating (multivariate) spectral densities as defined in Section 2.2. The first are the traditional non-parametric spectral estimators which estimate the spectral density of a stochastic process by means of its sample counterpart, the periodogram. The periodogram in its pure form can be shown to be an inconsistent estimator in the sense that it does not converge to the true spectrum as the sample size increases. This inconsistency can be repaired by applying so-called spectral windows which comes down to replacing the periodogram values at all frequencies by a weighted average of the periodogram values at adjacent frequencies. However, the most important problem in practice of using the periodogram as an estimator for a spectral density is that because of the finite sample only a limited number of auto-covariances are fed into the formula for the periodogram while the theoretical spectrum contains all (infinite) auto-covariances of the process. In Section 2.4 we saw that the disturbing leakage effect is a direct consequence of the finite sample size. Although the leakage effect can be “reduced” by applying spectral windows, this will always come at the expense of a lower resolution of the estimated spectrum. By resolution we mean the extent to which a spectral estimator is able to differentiate between separate, possibly adjacent, peaks in the theoretical spectrum. A lower resolution is therefore equivalent to a larger bias in the estimate. It is unavoidable that the averaging of the periodogram over adjacent frequencies causes adjacent peaks in the spectrum to be melted together which means a possible loss of valuable information about the dynamic behavior of the time series process under investigation. Especially in economics and finance, where most of the times only limited samples of data are available, this low resolution of the conventional non-parametric estimators is a serious problem. In case of small samples there will be a lot of leakage and hence a lot of smoothing required which leads to a loss of a lot of potentially valuable information. A second group of methods for estimating (multivariate) spectral densities are the less well known parametric spectral estimators. By estimating the spectrum through a cut off sequence of sample autocovariances, implicitly all higher order auto-covariances are assumed to be zero. This also holds if a spectral window is applied, although in that case additionally the sample auto-covariances that are available are modified by the weighting function. In fact, as a consequence of cutting off the auto- 12/40 covariances, the spectrum of an entirely different, to some extent even arbitrary, stochastic process is estimated. Parametric spectral estimators try to circumvent this problem by first estimating the parameters of some stochastic process on the available sample. Once such a stochastic process is “known”, its auto-covariances can be calculated for any order up to infinity. These auto-covariances can then be used to calculate the periodogram of the process as an estimate of the spectral density function. In a way, such a model ”extrapolates” the auto-covariances observed within the sample into auto-covariances for orders “outside” the sample. Note that this parametric approach for estimating spectral densities has a strong analogy with the Zero Phase Frequency filter described in Section 3.1.2 in terms of avoiding the disturbing leakage effects. When estimating a spectral density, the sample auto-covariances are extrapolated by means of a parametric model. When filtering a time series, a periodic component of some frequency present in the time series is extrapolated using a cosine function. A special case of such parametric spectral estimators are the so called autoregressive or Maximum Entropy spectral estimators. These consist of first estimating the parameters of a Vector AutoRegressive (VAR) model on the (decomposed) time series data and then calculating the (multivariate) spectral densities from this estimated model. This autoregressive spectral analysis leads to consistent estimators of spectral densities. Furthermore, a theoretical justification exists for choosing autoregressive models instead of many other possible models (including the non-parametric approach) to estimate the spectral densities. This justification is based on the information theoretical concept of Maximum Entropy. Entropy is a measure of “not knowing” the outcome of a random event. The higher the value of the entropy, the bigger the uncertainty about the outcome of the event. The entropy of a (discrete) probability distribution has its maximum value when all outcomes of the random event have an equal probability. The fundamental idea of the maximum entropy approach as first proposed by Burg (1967) is to select from all possible spectra that are consistent with the available information, represented by a (finite) sequence of (sample) auto-covariances, the spectrum which contains the least additional information as the best spectral estimate. All additional information on top of the information from the available sample is not supported by the data and should therefore be minimized. This is consistent with choosing the spectrum with the maximum entropy from all spectra that are consistent with the observed sample of auto-covariances. The solution to the corresponding optimization problem shows that the best way to estimate a spectrum in terms of this criterion is by estimating an autoregressive model and using the spectrum of this model as the spectral estimate. Furthermore, Shannon and Weaver (1949) show that given a number of auto-covariances, a Normal distributed process has the maximum entropy. So in total, the Maximum Entropy concept comes down to estimating a Normal distributed autoregressive model on the available time series data and calculating the spectral densities of this estimated model. Note that the Maximum Entropy concept itself says nothing about which order to select or which estimation procedure to use. Section 4.7 of Steehouwer (2005) describes Monte Carlo experiments which are set up to find the answers to these two questions. As an example, Figure 3.2 shows a Maximum Entropy spectral density estimate of the high frequency component of the long term interest rate time series as shown in Figure 3.1. Here, an AR(6) model is estimated by means of the Yule-Walker estimation technique on the full 1814-2007 sample. In terms of the further unraveling and understanding of the time series behavior of this long term interest rate time series, we can learn from this spectral density that on average its high frequency behavior is composed of fluctuations with a period length of around 10 years which describe approximately 50% of the high frequency variance. Furthermore, fluctuations with a period length of around 4.5 years seem to be important. Both observations are consistent with what is known about the business cycle behavior of many economic and financial time series. By extending the analysis into a multivariate 13/40 framework, also coherence and phase spectra can be used to further unravel the correlations between variable into frequencies, lead / lag relations and phase corrected correlations. Figure 3.2: Maximum Entropy (or Auto Regressive) spectral density estimate of the high frequency component of the long term interest rate time series for the Netherlands as shown in Figure 3.1. Around 50% of the variance of these high frequency (business cycle type of) fluctuation in the interest rate time series is described by fluctuations with a period length of around 10 (1/0.10) years while also pseudo periodic behavior with a period length of around 4 (1/0.25) to 5 (1/0.20) years can be observed. 3.3 Model Specification and Estimation After having analyzed the decomposed time series in order to obtain an adequate understanding of the total time series behavior, the third step in the methodology consists of actually modeling the component time series in line with this understanding. Because of the zero correlation property described in Section 3.1.3, the trend, low frequency and high frequency (or any other) components can in principle be modeled separately. In a multivariate context, each of these sub models will be a multivariate model. Once the separate trend and frequency models have been constructed, these can be added back together again to model the total time series behavior in terms of for example forecasts, confidence intervals or scenarios. This modeling process is depicted in Figure 3.3. 14/40 Figure 3.3: Frequency domain time series modeling approach which start at the top by decomposing multiple time series just as in the example in Figure 3.1. Next, the corresponding components (f.e. all high frequency components) from all time series are modeled by means of a suitable (multivariate) time series model. In the final step, these models are combined again to obtain a model that adequately describes the behavior of the time series at all frequencies. Historical time series Frequency domain filter Trend Model Frequency Model 1 . .. Frequency Model n Forecasts / Confidence intervals / Samples 3.3.1 Trend Model On purpose, in Figure 3.3 we distinguished the trend model from the models for the other frequency component time series. The reason for this is that the trend component of time series will typically require another modeling approach than the modeling of the other frequency components. For example, the low and high frequency component time series shown in Figure 3.1 can be modeled well by means of conventional time series modeling techniques because the time series show considerable variability. The trend component however, will typically be a straight flat or trending line for which it is clearly of little use to apply conventional time series modeling techniques. What to do then? One possibility would be just to extrapolate these straight lines into the future as being our trend model. Although we have only one observation of the trend value for each time series, we know that it could very well have had a different value and hence could very well have a very different value in our model for the future also. We should therefore prefer to work with a stochastic rather than a deterministic (linear) trend model. One single time series does not offer a lot of information for such a stochastic trend model. After all, even a very long sample time series, only provides us with one observation of the long term trend value. Therefore other sources of information are needed here. One could for example use a Bayesian approach, survey data or theoretical macroeconomic models. Also the distribution of the sample mean estimators could be used. A more direct data oriented approach is to switch from time series data to cross section data to obtain information about the ultra long trend behavior of economic and financial variables. As an example consider the statistics in Table 3.1 and th the top panel of Table 3.2. These are based on cross section data consisting of 20 century annual averages for the five indicated variables for 16 OECD countries (excluding Germany because of the extreme effects of the world wars). The volatilities and correlations observed here could form the basis for constructing an appropriate trend model. From these tables we can for example see that with a standard deviation of 2.0% the long term inflation uncertainty is substantial and that in terms of the trends, a total return equity index is positively correlated (0.55) to this long term inflation rate. Of course, the fundamental underlying assumption for such a cross section based approach is that the countries included in the cross section are sufficiently comparable to form a homogeneous group on 15/40 which to base the long term trend model behavior for each of these countries individually. Finally note that such an approach is very similar to the ones followed in historical long term cross country growth studies, but here it is proposed to extend this into a forward looking stochastic modeling framework. th Table 3.1: 20 century averages and (geometric) average growth rates of annual GDP (volume), CPI, (nominal) short and long term interest rates and (nominal) total equity returns (i.e. price changes plus dividend yields) for 16 OECD countries. Data from Maddison (2006) and Dimson et al. (2002). Short Long Log TRR Interest Rate Interest Rate Equity Index 3.9% 4.5% 5.2% 11.9% 2.2% 5.7% 5.2% 5.1% 8.1% Canada 3.8% 3.2% 4.9% 4.9% 9.7% Denmark 2.8% 4.1% 6.5% 7.1% 9.3% France 2.4% 7.8% 4.4% 7.0% 12.1% Ireland 2.0% 4.4% 5.2% 5.4% 9.4% Italy 2.9% 9.2% 5.0% 6.8% 12.1% Japan 4.0% 7.8% 5.5% 6.1% 13.0% Netherlands 3.0% 3.0% 3.7% 4.1% 9.1% Norway 3.3% 3.9% 5.0% 5.4% 8.0% South Africa 3.2% 4.8% 5.6% 6.2% 12.2% Spain 3.0% 6.2% 6.5% 7.5% 10.2% Sweden 2.7% 3.8% 5.8% 6.1% 12.4% Switzerland 2.6% 2.5% 3.3% 5.0% 7.3% United Kingdom 1.9% 4.1% 5.1% 5.4% 10.2% Unites States 3.3% 3.0% 4.1% 4.7% 10.3% Avg 2.8% 4.8% 5.0% 5.7% 10.3% Stdev 0.6% 2.0% 0.9% 1.0% 1.8% Min 1.9% 2.5% 3.3% 4.1% 7.3% Max 4.0% 9.2% 6.5% 7.5% 13.0% Country Log GDP Log CPI Australia 2.3% Belgium 16/40 3.3.2 Frequency Models As indicated in Figure 3.3, next to the trend model there exist a number of so called frequency models in the proposed methodology. These model the low and high frequency behavior of the time series variables around their underlying long term trends. The number of frequency models is the same as the number of frequency segments used in the decomposition step described in Section 3.1, excluding the trend component. Ideally the number of frequency segments and their size could be determined on a spectral analysis of the data. Due to data limitations this will often be difficult. Furthermore, because in the decomposition step no information is lost or added, we expect the exact choice of the frequency intervals to have a limited impact on the final model behavior. Instead, the split between the frequency models can therefore also be determined based on the sample size and observation frequencies of the data and the economic phenomena that, based on a thorough empirical analysis, need to be modeled. As an example of how this can work, Figure 3.4 shows five-annual observations of low frequency filtered times series of five economic and financial variables (the same as in Table 3.1) for the Netherlands for the 1870-2006 sample period. The second panel of Table 3.2 shows some statistics of the same low frequency component time series. These time series data capture the very long term deviations from the underlying trends and also contain information about changes in economic regimes. The sample needs to be as long as the data allows because here we are interested in the very low frequency behavior and we simply need a long sample to be able to observe this behavior adequately. Five-annual (instead of annual) observations are sufficient here to capture this low frequency behavior and facilitate the modeling of the corresponding time series behavior. If this is required for combining the various frequency models, the observation frequency of such a model can be increased to for example an annual observation frequency by means of for example simple linear interpolation. Figure 3.4: Five-annual observations of low frequency filtered times series of five indicated economic and financial variables for the Netherlands for 1870-2006. Log National Product Log Consumer Price I ndex Log TRR Equity Index Short Interest rate Long Interest Rate 1870 1890 1910 1930 1950 17/40 1970 1990 2010 Figure 3.5 shows annual observations of high frequency filtered times series of the same five economic and financial variables for the Netherlands but now for the 1970-2006 sample period. The third panel of Table 3.2 shows some statistics of the same high frequency component time series. These time series data capture the business cycle behavior of the variables around the sum of the trend and the low frequency components. Although empirically speaking business cycle behavior is surprisingly stable across samples spanning several centuries, a more recent 1970-2006 sample will be considered more relevant for the business cycle behavior in the near future while also more recent data will be of a higher quality. Higher (than five-annual) annual observation frequencies are needed to adequately capture the higher frequency behavior of the time series. Figure 3.5: Annual observations of high frequency filtered times series of five indicated economic and financial variables for the Netherlands for 1970-2006. Log National Product Log Consumer Price Index Short Interest rate Long Interest Rate Log TRR Equity Index 1970 1975 1980 1985 1990 1995 2000 2005 If also the monthly behavior of the variables would need to be modeled, a third frequency model could be included that would run on filtered monthly time series data for an even more recent sample, say 1990:01–2005:12, on a monthly observation frequency. These time series data would capture seasonal patterns and possibly also stochastic volatility patterns. In principle we can continue in this way by including additional frequency models for weekly, daily or over ticker data if required. 18/40 Table 3.2: Statistics of trend, low frequency and high frequency components of five economic and financial variables. Trend data are from Table 3.1. Low frequency data are five-annual observations of low frequency filtered times series for 1870-2006. High frequency data are annual observations of high frequency filtered time series for 1970-2006. Original time series are updates of data for the Netherlands from Steehouwer (2005). Trend Avg Corr and stdev Log GDP 2.8% 0.6% Log CPI 4.8% 0.07 2.0% Short Interest Rate 5.0% 0.08 0.31 0.9% Long Interest Rate 5.7% -0.01 0.67 0.70 1.0% Log TRR Equity Index 10.3% 0.19 0.55 0.27 0.44 Low Frequencies Avg 1.8% Corr and stdev Log GDP 0.0% 12.2% Log CPI 0.0% 0.23 15.7% Short Interest Rate 0.0% 0.70 0.50 1.6% Long Interest Rate 0.0% 0.64 0.31 0.86 1.8% Log TRR Equity Index 0.0% -0.18 0.39 -0.25 -0.27 High Frequencies Avg 33.2% Corr and stdev Log GDP 0.0% 1.7% Log CPI 0.0% -0.37 1.3% Short Interest Rate 0.0% 0.39 -0.22 1.8% Long Interest Rate 0.0% 0.10 0.10 0.74 0.9% Log TRR Equity Index 0.0% -0.12 -0.20 -0.37 -0.56 24.0% For each of the frequency models the most appropriate time series modeling and estimation techniques can be used to model the corresponding economic phenomena and time series behavior as good as possible. In principle, these models need not to come from the same class of models. For example, a structural business cycle model could be used for the high frequency components from Figure 3.5 while a model with seasonal unit roots could be used for the seasonal components in the monthly time series and a (G)ARCH model could be used to model the stochastic volatility in the seasonally corrected part of the monthly time series. Other time series modeling techniques to use in the different frequency models could be classical time series models, VAR models, models from the London School of Economics (LSE) methodology, theoretical models, Copula models, historical simulation techniques, etc. 19/40 3.3.3 State Dependencies In Section 3.1.3 we already described the usefulness of the zero correlations between the component time series that are obtained from the filtering process in the sense that it simplifies subsequent time series modeling. However, zero correlations do not need to imply that component time series are also independent. An example of a complex dependency between component time series is the so called ‘level effect’ in high frequency interest rate volatility. In the left panel of Figure 3.6 we see again the long term nominal interest rate in the Netherlands together with the filtered low and high frequency components. The low frequency component is shown here as the sum of the trend and low frequency components from Figure 3.1. If we define this sum as the underlying ‘level’ of the interest rate then it is clear that there exists a positive relation between the volatility of the high frequency component and the level of the interest rate. A similar effect can be found in short and long term interest rates for other countries and also in inflation rates. Figure 3.6: Level effect in the high frequency volatility of interest rates. The left panel shows the empirical positive relation between the underlying (low frequency) level of the nominal long term interest rate in the Netherlands from Figure 3.1 and its short term (high frequency) volatility. The right panel shows three samples from a model that explicitly captures this level effect by dynamically linking the volatility of the high frequency model to the level of the simulated underlying low frequency model according to a simple linear relation that has been estimated between the value of the low frequency component and the volatility of the high frequency component from the left panel. The perhaps counter intuitive fact is that, despite the zero correlation property, the decomposition approach actually facilitates the modeling of such complex dependencies between the component time series. If we can define and estimate some functional relation between the volatility of the high frequency component and the level or ‘state’ of the lower frequency models that describe the trend and low frequency components, then it is easy to implement this relation in for example a simulation framework of the model. What we need to do then is just start by simulating from the trend and low frequency models and then to simulate from the high frequency model while constantly updating the volatility of this high frequency model based on the observed level of the trend and low frequency simulations. The right hand panel of Figure 3.6 shows three example simulations of the long term interest rate that include this level effect in the high frequency volatility. The functional relation used here is taken from Section 13.2.1 of Steehouwer (2005). The approach described here can equally well be applied to more complex types of state dependencies, as for example state (with time as a special case) dependent business cycle dynamics and state dependent asset correlations. 20/40 3.3.4 Higher Moments and Non-Normal Distributions Spectral analysis techniques as described in Section 2 focus on the decomposition of the variance of time series and stochastic processes over a range of frequencies. Spectral densities are calculated as the Fourier transform of auto- and cross-covariances which shows that spectral analysis focuses on the second moments of the behavior of time series and stochastic processes. Although these second moment are of course very important, they do not cover all relevant information. Furthermore, in Section 3.2.1 we gave a justification for using Normal distributions in the Maximum Entropy spectral analysis framework while we know that the behavior of economic and financial variables will often be far from Normal. When studying the low and high frequency properties of decomposed time series data, one soon finds many forms of non-Normal distributions with aberrant third and fourth moments. As an example look at the Skewness and (excess) Kurtosis numbers of the cross section trend data in Table 3.1. Another example are monthly high frequency components of equity returns in which we often find empirical distributions that are more peaked with thinner tails than the Normal distribution (leptokurtic). A first way of modeling these kinds of non-Normal distributions in the proposed modeling framework is an explicit modeling approach. As an example, note that the modeling of the level effect in high frequency interest rate volatility along the lines as described in Section 3.3.3 will result in a skewed overall distribution. A second example is modeling a stochastic volatility in the monthly high frequency component of equity prices which will result in leptokurtic distributions for the corresponding frequency model. A second way of modeling non-Normal distributions would of course be to use other than Normal distributions in the trend and frequency models. 3.3.5 Missing Data Solutions Ideally the methodology described here would run on an abundance of empirical time series data consisting of samples of hundreds of years on say a daily observation frequency of all economic and financial variables. Of course this kind of time series data will not be available in most cases. Especially the trend and low frequency models require cross section or very long time series which will not be available for all variables. This poses a problem because every variable needs to be specified in the trend model and in each of the frequency models to be able to describe the complete behavior of the variables. One standpoint could be that if long term data is not available for certain variables, one should not try to model the long term behavior of these variables in the first place (also see Section 4.3) but this is not always a feasible standpoint. Another solution to this missing data problem is to describe the behavior of variables for which we have insufficient data in certain frequency ranges as a function of the variables for which we do have sufficient data. The functional form and parameters of such relations can be determined in two ways. The first is to perform an empirical analysis on a shorter sample for which data for all the required variables is available and estimate an appropriate relation. The second possibility is to base the relations on economic theory. 21/40 As examples of this theoretic approach consider • • • to describe the low frequency behavior of a real interest rate as the low frequency behavior in a nominal interest rate minus the low frequency behavior of price inflation (Fisher relation), to describe the low frequency behavior of an exchange rate as the difference in the low frequency behavior of the involved price indices (Purchasing Power Parity) or to describe the low frequency behavior in private equity prices as the low frequency behavior of public equity prices plus some error term. 3.4 Model Analysis After in the previous steps having decomposed time series data, analyzed the decomposed time series and modeled the component time series, the fourth step in the methodology consists of analyzing the constructed models. Such an analysis is required for example to check whether the models adequately describe the intended behavior or to use the properties of the models as input for the framework in which they are applied. An example of the latter would be to calculate the relevant moments of the implied stochastic behavior of the economic and financial variables as input for a SAA optimization routine across various horizons. In the described methodology, traditional methods to perform such a model analysis can still be applied. Depending on the types of time series models that are used, for example unconditional and conditional distribution characteristics such as means, standard deviations, correlations and percentiles can still be calculated for the trend and various frequency models separately. The only extension here is that these characteristics have to be combined to obtain characteristics of the total model. In most cases this is fairly straightforward because of the zero correlation property described in Section 3.1.3. For example the total covariance between two variables is simply the sum of the covariances between those two variables in the trend and each of the frequency models. In addition to the traditional methods, two additional types of analysis are available in the described frequency domain methodology which consist of spectral analysis and variance decomposition techniques. 3.4.1 Spectral Analysis In Section 3.2.1 we described that spectral analysis techniques are very powerful for unraveling and understanding the behavior of filtered time series components. However, spectral analysis techniques can also be fruitfully applied to analyze the behavior of the constructed time series models. Multivariate spectral densities in terms of the auto-, coherence and phase spectra can equally well be calculated for each of the frequency models separately as for the total model. The resulting model spectral densities can for example be compared to the spectral densities of the (decomposed) time series data. Calculating spectral densities for estimated models can be done in two ways. The first is a direct calculation of the spectral densities based on the parameters of the model. Section 4.5 of Steehouwer (2005) for example gives the spectral density formulas for a simple white noise process and moving average and autoregressive models. A second and more flexible possibility is to first calculate, possibly numerically, the auto- and cross-covariances of a sufficient high order and then to apply the Fourier transform to transform these auto- and cross-covariances into the corresponding spectral densities. Because of the zero correlation property, spectral densities for the total model in terms of the trend and various frequency models can easily be obtained by first summing the autoand cross-covariances across the sub models before transforming these into the frequency domain. 22/40 3.4.2 Variance Decomposition Just as a spectral density decomposes the variance of one of the frequency models over the whole frequency range, a similar decomposition can be made of the variance of the total model into the variance of the individual trend and frequency models. Again because of the zero correlation property, constructing such a variance decomposition is fairly straightforward. The variance of the sum of the 5 models is simply the sum of the individual variances . A variance decomposition can be constructed for both the unconditional distributions and the conditional distributions. The latter gives insight into the contribution of the different sub models to the total variance at different horizons. As an example consider Figure 3.7 which shows a conditional variance decomposition for a long term interest rate and a log equity total rate of return index. The model consists of a stochastic trend model, a low frequency model, a business cycle model, a seasonal monthly model and a seasonally corrected th monthly model. For every 12 month in a 35 year horizon, the two panels show in a cumulative fashion the proportion of the total conditional variance of that month that can be attributed to the various sub models. From the right hand panel we can see for example that the business cycle model (m=2, which is here the same as m=3) describes around 80% of the variance of the log equity total return index on a 1 year horizon while on a 35 year horizon this holds for the trend and low frequency model (m=1). That is, high frequency models are typically important on short horizons while low frequency models are important on long horizons. th Figure 3.7: Variance decompositions evaluated at every 12 month in a 35 year horizon. Left panel for a long term interest rate and right panel for a log equity total rate of return index. The lines show in a cumulative fashion which portion of the conditional variance at the various horizons is caused by a trend model and four frequency models. In general such variance decompositions show for models of economic and financial that low (high) frequency behavior causes a relatively larger part of the variance at long (short) term horizons. 5 Note that some kind of interpolation method might be needed to align all frequency models on the same (highest) observation frequency. 23/40 4 Motivation Why do we propose the specific frequency domain methodology for time series analysis and modeling? The motivation for this consists of several points, some of which have already been implicitly given in the description of the methodology in Section 3. In the following sub sections we make these points of motivation more explicit, extend them with a further argumentation and also put forward a number of new points of motivation. 4.1 Understanding Data and Model Dynamics A first reason for proposing the frequency domain methodology for time series modeling is that it provides very powerful tools for understanding the dynamic behavior in historical time series data as well as for analyzing the dynamic properties of models that describe time series behavior for the future. If there is one thing about the behavior of economic and financial variables we know, then it is that they move up and down and never move in straight and stable paths. Therefore, frequency domain techniques are the most natural to analyze how exactly they move up and down. What types of fluctuations dominate the behavior of a variable and what are the correlations and lead / lag relations with other variables at the various speeds of fluctuations? As a first element, decomposing time series into different components allows us to zoom in on the behavior in various frequency regions which provides us with a clearer and more focused insight in the corresponding dynamic behavior. We can for example focus our attention on the business cycle behavior of economic and financial variables, something which is very common in business cycle research but much less common in a time series modeling context. As a second element, estimating and analyzing spectral densities is a very efficient way of summarizing the dynamic behavior of component time series within a certain (high) frequency region. Although spectral densities contain exactly the same information as conventional time domain auto- and cross-correlations, spectral densities represent this information in a more efficient and intuitive manner. These appealing properties of frequency domain approaches are certainly not new to economists. It has long been recognized that when studying macroeconomics one has to make a clear distinction as to which aspect of macroeconomics one is interested in. It seems likely that we will be dealing with very different ‘forces‘ when we are studying for example the long term growth of economies, comprising many decades, compared to the intra day trading effects on a stock exchange. If different forces are at work, also different models or approaches may be needed to adequately analyze and describe the relevant economic behavior. The first formalization of this idea dates back to Tinbergen (1946) who proposed a decomposition of time series of economic variables as Time Series = Trend + Cycle + Seasonal + Random (4.1) The first economic applications of spectral analysis date as far back as Beveridge (1922) who used a periodogram to analyze the behavior of a wheat price index. However, for a successful application of frequency domain techniques on economic and financial time series, it is crucial to use appropriate and special versions of these techniques that can deal with the leakage problem of Section 2.4 as caused by the in general limited sample sizes available in economics and finance. This is exactly what the Zero Phase Frequency filter from Section 3.1.2 and the Maximum Entropy spectral analysis from Section 3.2.1 provide us with. 24/40 4.2 Different Economic and Empirical Phenomena A second reason for proposing the frequency domain methodology for time series modeling concerns the fact that at different horizons (centuries, decades, years, months, etc.) and different observation frequencies (annual, monthly, weekly, etc.) the empirical behavior of economic and financial variables is typically different and dominated by different well known economic phenomena such as long term trends, business cycles, seasonal patterns, stochastic volatilities, etc. First of all, decomposing time series into different components allows us to analyze these economic phenomena separately while spectral analysis techniques can be used for a further unraveling of the corresponding behavior. Second, by also following a decomposed modeling approach, as summarized by Figure 3.3, we are also able to adequately model the different economic phenomena simultaneously by using the most appropriate time series models for each of them. The potential benefits of this decomposition approach to time series modeling can perhaps best be illustrated by the following extreme example. Suppose one wants to model realistic behavior of economic and financial variables up to a horizon of several decades but on a daily basis. Such a model should at the same time give an adequate description of the trending, low frequency, business cycle, seasonal and daily behavior, each with their very specific properties and for each of the variables. Obviously, achieving this within one single conventional time series modeling approach is very difficult. Estimating for example a conventional Vector AutoRegressive (VAR) model on a sample of daily observations will probably not produce the intended result. However, with the proposed decomposition approach to time series modeling this is exactly what is possible in a theoretically well founded, simple and flexible way. Thereby, the proposed methodology brings together the empirical behavior of economic and financial variables that is observed at different horizons and different observation frequencies in one complete and consistent modeling approach. Furthermore, this way of modeling also stimulates the incorporation of economic reasoning and intuition into what can become purely statistical time series modeling exercises. 4.3 Appropriate Samples and Observation Frequencies A third reason for proposing the frequency domain methodology for time series modeling is that it allows us to use appropriate time series data in terms of samples and observation frequencies to analyze and model the various economic phenomena mentioned in Section 4.2. Furthermore, the methodology allows us to optimally combine all available sources of time series information. In some cases it is evident what type of time series data is required for modeling a certain economic phenomenon. For example, if we want to model seasonal patterns it is clear that monthly data for a sufficient number of years is required. However, in some cases what data to use seems less evident. This especially holds in the case of modeling the very long term (low frequency) behavior of economic and financial variables. As an example consider Figure 4.1 which shows two samples of the same long term nominal interest rate in the Netherlands as used in Section 3.1. Suppose we would use the left hand 38 year long annual 1970-2007 sample as a basis for modeling the long term behavior of this nominal interest rate, suppose up to an horizon of 30 years into the future. A sample of almost 40 years certainly seems like a long sample at first sight. However, from a frequency domain point of view, we are modeling the behavior at very low frequencies, which may have period lengths of 40 years or more, based on one single sample of 40 years. In a sense, by doing this, we are estimating a model for the low frequency behavior based on one single observation of this low frequency behavior. This is clearly inadequate and can lead to something Reijnders (1990) calls perspective distortion which means that if one looks at too short a sample, one can be misled about the behavior of economic and financial variables. Based on the left hand panel of Figure 4.1 we could for example conclude that the average interest rate is somewhere around 7%, the interest rate has a strong downward trend (which might lead us to the conclusion that the interest rate is a non-stationary 25/40 process that needs to be modeled in terms of the annual changes of the interest rate instead of the levels) and always has a large amount of short term volatility. However, if we look at the full 194 year 1814-2007 sample, as shown in the right hand panel of Figure 4.1, entirely different information about the behavior of the long interest rate is revealed. We see for example that the average interest rate is actually more around 4 to 5%, does not have a downward trend but during the 1970-2007 period was returning back from ‘exceptionally’ high post war levels to this more regular level of 4 to 5% (which shows us that the interest rate is actually a stationary process with a very slow rate of mean reversion and should be modeled as such in terms of the levels) and short term volatility can also be quite low, especially at low interest rate levels (hence, the ‘level effect’ discussed in Section 3.3.3). Figure 4.1: The risk of ‘perspective distortion’ by using short samples illustrated on a long term interest rate time series for the Netherlands. Left hand panel shows sample 1970-2007, right hand panel shows full 1814-2007 sample. From the left panel one could be lead to believe in a high level, downward trending and high (short term) volatility behavior of this interest rate variable while the right panel shows that the true behavior of the interest rate actually consists of long term fluctuations around a lower level with short term volatilities that are positively related to the underlying level of the interest rate. Therefore, by considering only a relatively small part of the available sample one can be misled about the behavior of time series variables and hence can be led to construct erroneous models to describe this behavior. 26/40 So in short we can say that ‘short’ samples can give insufficient information for modeling the long term behavior of economic and financial variables. This is a first example of the observation that conventional modeling approaches tend to model the long term behavior (which should be based on long samples) of economic and financial variables based on an ‘extrapolation’ of the short term behavior of these same variables (by using short samples). One could wonder whether this is important. Is there really a big difference between the long and short term behavior? The answer to this question is a strong yes. As an example consider the correlation between equity returns and inflation rates as shown in Table 3.2. We see that this long term correlation in terms of the trend and low frequency components, is very positive (0.55 and 0.39 respectively) while this short term correlation in terms of the high frequency component is actually negative (-0.20). For a discussion of the literature and these data findings about the inflation hedging capacities of equities at different horizons, we refer to Section 16.4.3 of Steehouwer (2005). Here we just want to stress the potential danger and impact of modeling long term behavior based on short term data. Conventional annual equity returns and annual inflation rates will show the same type of negative short term high frequency correlation. If we estimate a model on this data there is a risk that the implied long term correlation will also be negative, instead of being positive as the data tells us. It is not hard to imagine that in terms of for example SAA decision making for a pension plan with inflation driven liabilities, working with a negative instead of a positive long term correlation between equity returns and inflation rates will have an enormous negative impact on the amount of equities in the optimal SAA. To avoid this ‘perspective distortion’ and have sufficient information for modeling both the long term (low frequency) and short term (high frequency) behavior of economic and financial time series, we would ideally have very long sample time series (say covering several centuries) with high observation frequencies (say daily). Based on this data we would then apply the decomposition, time series analysis, modeling and model analysis steps of the methodology described in Section 3. Of course such ideal time series data is not available in most cases. However, by accommodating the use of samples of different sizes and observation frequencies for the long and short term behavior in terms of the trend and various frequency models, the methodology does allow an optimal use of all the time th series data that is available. In the examples given in Section 3.3.1 and 0, we used 20 century cross section data for 16 countries for the trend model, five-annual 1870-2006 data for the low frequency model, annual 1970-2006 data for the high frequency (business cycle) model and monthly 1990:012005:12 data for the monthly frequencies. A consistent use of these different data sources in the methodology is achieved by means of applying an appropriate decomposition approach. For example, if we use the annual sample above to describe the business cycle behavior as all fluctuation in the time series with a period length between 15 and 2 years, we should make sure that these types of business cycle fluctuations are excluded from the monthly sample above by filtering out all fluctuation longer than 24 months from the monthly time series. Note that some kind of interpolation method might be needed to align all frequency models on the same (highest) observation frequency. An obvious comment on the use of using very long samples of data, in some cases covering several centuries, might be to ask how relevant and representative data this far back still is for the current and future behavior of economic and financial variables. There are two answers to this question. The first is that based on thorough empirical analysis of historical time series data, one might be surprised about the amount of ‘stability’ there actually is in the behavior of economic and financial variables, not only across time but also across countries. A nice quote to illustrate this in the context of business cycle behavior comes from Lucas (1977) who states that “though there is absolutely no theoretical reason to anticipate it, one is led by the facts to conclude that, with respect to the qualitative behavior of co-movements among series, business cycles are all alike”. This remarkable stability of the business cycle mechanism has also been reported more recently by for example Blackburn and Ravn (1992), Backus and Kehoe (1992) and Englund et al. (1992) and Steehouwer (2005). Nevertheless 27/40 also business cycle behavior has gradually changed over time and therefore we would be inclined to use a relative recent sample (say 1970-2006) to model the business cycle behavior. The second answer to this question of the relevance of long term data is to realize that all science starts with the analysis and understanding of data that give us information about the phenomenon that we are studying. Therefore, to understand and model the long term behavior of economic and financial variables, by definition we have to start by studying long term time series data. Of course we can start deviating from what the data tells us by incorporating theoretic or forward looking information but we should start from the data at the very least. Jorion and Goetzmann (2000) illustrate this approach by stating that “Financial archaeology involves digging through reams of financial data in search for answers”. It is known that different economic regimes and historical circumstances are underlying the behavior observed in long term historical time series data. One could argue that this is not a problem of using long term data but that in fact such changes in economic regimes and historical circumstances are exactly what drive the uncertainty and behavior of financial variables in the long run. So by modeling directly on long term historical data we are taking a kind of stochastic approach to regime switching. Just as we view annual business cycle observations for a 1970-2006 sample as realizations of some underlying business cycle process, we can also view five-annual observations of low frequency behavior for a 1870-2006 sample as realizations of some underlying long term process. As far as regime changes affect the short term high frequency behavior of economic and financial variables, this is exactly what the high frequency models should adequately describe. It is for example well known that at the business cycle frequencies consumer prices have changed from a coincidental behavior into a lagging behavior when compared to the GDP. We can model this (lagging) behavior adequately by using a sufficiently recent sample for the business cycle frequencies. In other cases we might also want to model the dependency of the high frequency behavior on the low frequency behavior directly, for example in terms of the ‘level effect’ in the short term volatility of interest rates described in Section 3.3.3. In this way the described methodology allows for an optimal use of the available time series data and allows us to use the appropriate samples and observation frequencies for modeling the various long and short term economic phenomena. 28/40 4.4 Equal Importance of all Frequencies A fourth reason for proposing the frequency domain methodology for time series modeling is that it considers the behavior of economic and financial variables in all frequency ranges, and thereby also in terms of the long and short term behavior, of equal importance. Therefore the methodology does not put the focus on either the long term low frequency behavior or the short term high frequency behavior, but allows us to focus on the long and short term behavior at the same time. We can best explain this point by again using the long term interest rate time series introduced in Section 3.1. In Section 4.3 we argued that a first reason that conventional modeling approaches tend to model the long term behavior of economic and financial variables based on an ‘extrapolation’ of the short term behavior is that too short samples are being used to be able to contain fundamental information about the long term behavior in the first place. However, even if long samples are used, there is still a second risk of ‘extrapolating’ the short term behavior into the long term. This is so because in addition to the sample, also the representation of the time series data used for the modeling plays an important role. Especially, the effects of the often applied first order differencing operator are important here. This filter is often applied to model for example period to period changes of variables such as interest rates or period to period returns of variables such as equity prices. From the PTF of the filter in Figure 2.1 we already saw that in terms of a frequency domain point of view, the first order differencing filter suppresses the low frequency behavior and amplifies the high frequency behavior in time series. This effect is clearly visible in the left hand panel of Figure 4.2 which shows the original (year end) levels together with annual changes of the interest rate. In the annual changes, the information about what maximum and minimum levels the interest rate has achieved or what type of long term fluctuations it has experienced is missing while at the same time the short term fluctuations dominate the time series of annual changes. Because of the clearly different types of fluctuations that dominate the variance of the level and the annual changes of the interest rate time series (just by using another representation of the data), it is not hard to imagine that models estimated on the level will do well at describing the long term behavior while models estimated on the annual changes will do well at describing the short term behavior. So, by focusing on long sample time series but still using representations of the time series in terms of period to period changes or returns, conventional approaches are still focusing on the short term high frequency behavior and still run the risk of not adequately capturing the true long term low frequency behavior, simply because the latter type of information has been severely suppressed in the time series data used for estimating the models. Of course the short term behavior of interest rates is important in some applications, for example for modeling the returns on fixed income portfolios. However, in other applications the long term behavior of interest rates is important, for example to determine how large the uncertainly about the mark to market value of accrued pension liabilities is. That is, the behavior of economic and financial variables at all frequencies is in principle of equal importance and we should therefore avoid to be (implicitly) putting more importance on the behavior in one specific frequency range. How does the proposed methodology solve this problem? The answer is by using an appropriate decomposition approach as the one described in Section 3.1. The frequency domain filter proposed for these purposes, does not (unintended) amplify nor suppresses the behavior in certain frequency ranges. Instead, it just cuts up the behavior into different frequency regions. In technical terms, the PTF of the filters described that are used, only have values of 0 or 1 and together cover exactly the range of all possible frequencies. We can see the benefits of this approach in the right hand panel of Figure 4.2. This is the same filter output as shown in Figure 3.1 with the only exception that we added the trend and low frequency components. If we add the resulting low and high frequency components 29/40 we again get the original interest rate time series. This is not possible for the annual changes because in that case there is no second component time series which contains for example the suppressed part of the low frequency behavior. Comparing the right and left hand panel it is easy to see that the low frequency component captures the long term behavior of the original level of the interest rate while the high frequency component captures the short term behavior in terms of the annual changes of the interest rate. By constructing separate models for the low and high frequency components we are capable of modeling both the long term low frequency and short term high frequency behavior of economic and financial variables adequately at the same time instead of focusing on of the two while we know that both can be of equal importance. Figure 4.2: The benefits of the decomposition approach. The left hand panel shows the original level and the annual changes of the long term nominal interest rate for the Netherlands from Figure 3.1. The right hand panel shows the sum of the trend and low frequency component together with the high frequency component from the same Figure 3.1. Comparing the two panels shows how the low frequency component captures the behavior of the long term level of the interest rate while the high frequency component captures its short term annual changes. If one models both the low and high frequency component separately, one can therefore adequately model both the level and annual changes of the interest rate at the same time. The left hand panel shows that if one for example tries to model the complete behavior in terms of the annual changes, this becomes rather difficult as the information about the long term level of the interest rate has been suppressed and is therefore hardly visible in the annual changes. 4.4.1 Forecasting consequences What are the possible consequences of using (conventional) modeling approaches that (implicitly) give an unequal importance to the different frequency ranges? A first possible consequence concerns the forecasting performance, and especially the long term forecasting performance. To see this, assume that indeed there is some valuable information in the long term low frequency behavior of economic and financial time series and let us compare the forecasts of two modeling approaches. The first is a conventional modeling approach that ‘ignores’ this low frequency information by either using relatively short samples and / or using (first order differencing) representations of the data that suppress this low frequency information. The second approach uses the proposed decomposition approach that explicitly takes the (unchanged) low frequency information into account. For the first approach we estimated an autoregressive model on the annual (log) GDP growth rates for the Netherlands and produced forecasts and confidence intervals for a horizon of 50 years into the future. The results are shown in the left hand panel of Figure 4.3. The right hand panel shows the same 30/40 results but now for the case in which we first decomposed the (log) GDP time series into low and high frequency components on which we then estimated separate autoregressive models. The most striking difference between the two forecasts (the median central solid line), is that for the conventional approach the forecast soon becomes a rather uninformative flat trending line while for the decomposition approach it still fluctuates and is therefore informative during the complete 50 year horizon. Although this is a topic for future research, preliminary results from a formal backtesting procedure have already indicated that the decomposition approach can indeed lead to smaller forecast errors when compared to conventional approaches in which no decomposition is applied. This indicates that there is some valuable information in the low frequency behavior of time series that can be (better) exploited by the decomposition approach. Figure 4.3: Out of sample forecasts and confidence intervals of log GDP in the Netherlands. Left hand panel show results of a conventional modeling approach based on annual growth rates. Right hand panel shows results of a decomposition approach based on separate low and high frequency components in the log GDP time series. Fluctuations with long period lengths are suppressed by the PTF of the first order differencing operator and are therefore hardly present in the growth rate time series. As a result, the forecasts of the model estimated on these growth rates as shown in the left panel are very uninformative in terms of the low frequency behavior of the GDP series. The right panel clearly also shows low frequency information in the forecasts from the decomposition approach. 4.4.2 Monte Carlo experiment In addition to the backtesting experiments that point in the direction of a superior forecasting performance of the decomposition approach, we have performed a formal Monte Carlo experiment to show that the decomposition approach works better in terms of modeling both the long term low frequency and short term high frequency behavior well at the same time. We already explained the intuition why this may be so based on the results shown in Figure 4.2 and also the results of the tests in Chapter 19 of Steehouwer (2005) already pointed in this direction. Here we describe and give the results of a Monte Carlo experiment which adds formal evidence to this claim. We assume the Data Generating Process (DGP) for a variable zt defined by (4.2) where t is measured in years and of which the dynamic properties are inspired by the empirical behavior of the long term interest rate time series shown in Figure 3.1. This process consists of two independent components, a low frequency component xt and a high frequency component yt which are both driven by a specific autoregressive process. Table 4.1 show the means and variances of these two components from which we see that the low frequency component describes 80% of the total variance and the high frequency component 20%. 31/40 zt = xt + yt (4.2) with xt = 1.89 xt −1 − 0.90 xt − 2 + ε t ε t ~ N (0,0.0027 2 ) yt = 0.65 yt −1 − 0.34 yt −2 − 0.26 yt −3 − 0.20 yt −4 + 0.26 yt −5 − 0.46 yt −6 + γ t γ t ~ N (0,0.08102 ) Table 4.1: Mean and variance of low, high frequency and total model from (4.2). The total model has a mean of 0 and a variance of 1 which for 80% comes from the low frequency model and for 20% from the high frequency model. xt yt zt = xt+yt ∆zt Mean 0.00 0.00 0.00 0.00 Variance 0.80 0.20 1.00 0.19 Table 4.2: Complex roots of low and high frequency models from (4.2). The low frequency model (xt) describes pseudo period behavior with a period length of around 50 years. The low frequency model (yt) describes fluctuations with a period length of around 10, 5 and 2.5 years. Root 1 xt Root 1 yt Root 2 yt Root 3 yt Modulus 0.95 0.92 0.87 0.85 Frequency 0.02 0.10 0.20 0.40 Period length 50 10 5 2.5 Table 4.2 shows the modulus, frequency and corresponding period length of the complex roots of the two autoregressive polynomials. From this we see that the low frequency model describes pseudo periodic behavior with a period length of around 50 years while the high frequency model is composed of three types of pseudo periodic behavior with a period length of around respectively 10, 5 and 2.5 years. These types of pseudo period behavior are clearly visible in the (non-normalized) spectral densities in the top two panels of Figure 4.4 which notably integrate to the total variances of 0.80 and 0.20. The bottom left panel shows the implied total spectral density of the DGP (4.2) in terms of the level of zt. The bottom right hand panel of Figure 4.4 shows the implied spectral density of the annual changes of zt, that is of ∆zt. Note how the PTF of the first order differencing operator (∆) shown in Figure 2.1 has rescaled the spectral density by reducing the importance of the low frequencies and increased the importance of the high frequencies in the process. The latter two spectral densities are accompanied by the spectral densities of good approximating autoregressive models. This shows that the DGP of both zt and ∆zt can be well described by (different) autoregressive models. Figure 4.5 shows an example simulation of 200 years from the xt, yt, zt=xt+yt and ∆zt processes from which we can clearly see the behavior described by the roots and spectral densities of the models. Based on the DGP (4.2), the Monte Carlo experiment is now set up as follows. We generate 1000 simulations of sample sizes of 50, 100, 200, 500 and 1000 years. In all cases we simulate a presample of 500 years to guarantee that all simulations adequately represent the unconditional distributions. For each of the individual simulations we use the Yule-Walker estimation technique to try to back out the original GDP of zt from the simulation. Tests described in Section 4.7 of Steehouwer 32/40 (2005) show that this estimation technique shows the best performance in terms of estimating spectral densities, especially in small samples. In large samples many estimation techniques show a similar performance because of identical asymptotic properties. We compare three approaches: 1. Level approach: Estimate an autoregressive model on the simulated level zt 2. Delta approach: Estimate an autoregressive model on the simulated annual changes ∆zt 3. Decomposition approach: Estimate separate autoregressive models on the underlying simulated low and high frequency components xt and yt For each simulation we use each of these approaches to estimate the spectral densities of both zt and ∆zt. The spectral densities can be calculated directly from the parameters of the autoregressive parameters or by applying the Fourier transform on the auto-covariances of the estimated model up to a sufficiently high order. If the estimation has been done in terms of zt, the spectral density of ∆zt is calculated by applying the PTF of the first order differencing filter on the estimated spectral density of zt. If, the other way around, the estimation has been done in terms of ∆zt, the spectral density of zt is calculated by applying the PTF of the inverse first order differencing filter on the estimated spectral density of ∆zt. We then calculate the errors between the estimated spectral densities and the true spectral densities given in the bottom two panels of Figure 4.4 as b S error = ∫ Sˆ (ω ) − S (ω ) dω a b ∫ S (ω )dω (4.3) a Here S (ω ) and Sˆ (ω ) are respectively the non-normalized auto-spectrum of the known underlying DGP and its estimated counterpart. The integration interval [a,b] defines the frequency range over which the error is calculated. The error (4.3) measures the ‘distance’ between the estimated and the DGP spectra on the indicated frequency range. The smaller this distance, the better the estimated process corresponds to the original DGP and hence the better is the performance of the relevant approach. We calculate these errors for in total six cases that are combinations of the DGP representation in terms of the level zt or the annual changes ∆zt on the one hand and different frequency regions on the other hand. These combinations are indicated in Table 4.3. The split between the low and high frequency ranges is made at a frequency of 0.0667 which is equivalent to fluctuations with a period length of 15 years. Table 4.3: Six combinations of DGP representation and frequency ranges for which error (4.3) is calculated. Frequencies hold in terms of cycles per year. Frequency range Level zt Annual change ∆zt Total [0.0000, 0.5000] [0.0000, 0.5000] Low [0.0000, 0.0667] [0.0000, 0.0667] High [0.0667, 0.5000] [0.0667, 0.5000] To be sure we get the best out of each approach, for each estimation we calculate the errors for autoregressive models of all orders between 1 and 25 (this range includes the orders needed to closely approximate the model spectra as shown in the bottom two panels of Figure 4.4) and select the order which gives the smallest error in terms of the representation of the DGP used in the different approaches (levels, annual changes or separate low and high frequency components). Finally, we calculate and compare the mean errors from the calculated errors for each of the 1000 simulations. 33/40 Figure 4.4: Non-normalized spectral densities of xt (top left panel), yt (top right panel), level of zt (bottom left panel) and annual changes of zt (bottom right panel) for models from (4.2). The latter two are accompanied by the spectral densities of close approximating autoregressive models. The complex roots from Table 4.2 are clearly visible in the spectral densities. 34/40 Figure 4.5: Example simulation of 200 years of xt (top left panel), yt (top right panel), level of zt (bottom left panel) and annual changes of zt (bottom right panel) for models from (4.2). The results of the experiment are reported in Table 4.4. A mean error of for example 0.50 for the level approach (second column in each block) says that the level approach on average results in a wrong allocation of 50% of the variance over the frequencies in the estimated spectral densities. The mean errors for the delta and decomposition approaches (third and fourth columns in each block) are reported as a percentage of the mean error of the level approach. A value of for example -0.25% for the delta approach says that the delta approach results in a mean error which is 25% lower than the mean error for the level approach. In case of the example, the mean error of the delta approach is 75%×0.50=0.375. 35/40 Table 4.4: Mean errors (4.3) for each of the six combinations in Table 4.3 based on 1000 simulations. The mean errors of the delta and decomposition approaches (third and fourth columns in each block) are reported as a percentage of the mean error of the level approach (second column in each block). Mean error delta (∆zt) total Mean error level (zt) total Sample Level Delta Decomp Sample Level Delta Decomp 50 0.63 6889% -3% 50 0.65 -26% -20% 100 0.49 7894% -6% 100 0.50 -20% -24% 200 0.37 9655% -8% 200 0.36 -12% -26% 500 0.25 13995% -13% 500 0.25 -5% -31% 1000 0.18 15372% -16% 1000 0.19 0% -34% Mean error delta (∆zt ) low Mean error level (zt) low Sample Level Delta Decomp Sample Level Delta Decomp 50 0.51 8415% -1% 50 0.06 12% -12% 100 0.40 9742% -3% 100 0.04 32% -14% 200 0.30 11992% -4% 200 0.03 68% -14% 500 0.20 17901% -8% 500 0.02 132% -20% 1000 0.14 20003% -11% 1000 0.01 184% -22% Mean error delta (∆zt) high Mean error level (zt ) high Sample Level Delta Decomp Sample Level Delta Decomp 50 0.11 -13% -14% 50 0.60 -30% -21% 100 0.09 -10% -21% 100 0.46 -25% -25% 200 0.07 -8% -25% 200 0.33 -19% -27% 500 0.05 -9% -32% 500 0.24 -16% -32% 1000 0.04 -9% -34% 1000 0.18 -12% -35% From the results in Table 4.4 we observe the following: 1. Larger samples lead to smaller errors in terms of reproducing the original DGP spectra. 2. In terms of the total frequency range (top two panels in Table 4.4), the level approach is best at reproducing the spectral density of the levels while the delta approach is best at reproducing the spectral density of the annual changes. 3. The delta approach is better at reproducing the high frequency part of the spectral density of the levels (bottom left panel in Table 4.4) while the level approach is better at reproducing the low frequency part of the spectral density of the annual changes (middle right panel in Table 4.4). 36/40 4. In virtually all cases, the decomposition approach is better at reproducing the spectral densities of the levels and annual changes, both in terms of the total as in terms of the separate low and high frequency ranges. Observation 1 is of course no surprise. Observation 2 confirms what we explained about the left hand panel of Figure 4.2. That is, models estimated on the levels will do well at describing the long term behavior of the levels while models estimated on the annual changes will do well at describing the short term behavior of the annual changes because of the clearly different types of fluctuations that dominate the variance of the processes of the levels and the annual changes. Note that the errors for the delta approach in terms of the low (and total) frequency ranges of the spectral density of the levels are so large because of the PTF of the inverse first order differencing filter which approaches infinity at very low frequencies. Another way of seeing this is that in those cases we are modeling a stationary process in terms of a non-stationary (integrated) process. Observation 3 is somewhat more of a surprise but only further emphasizes the same point as observation 2. As we saw in Section 2.3, the PTF of the first order differencing filter suppresses the low frequency behavior and amplifies the high frequency behavior. Therefore the delta approach performs even better than the level approach itself at the high frequencies of the level DGP. The other way around, the delta approach itself performs even worse than the level approach at the low frequencies of the annual changes DGP. Observation 4 obviously closes this experiment by confirming what we explained about the right hand panel of Figure 4.2. That is, by constructing separate models for the low and high frequency components we are capable of modeling both the long term low frequency (levels) behavior and the short term high frequency (annual changes) behavior adequately at the same time instead of doing well in terms of the low frequency behavior and poorly in terms of the high frequency behavior or the other way around. Thereby, the Monte Carlo experiment gives formal support for the claim that a decomposition approach can lead to superior modeling results in terms of describing the behavior of economic and financial variables in all frequency ranges. 4.5 Complex Dependencies between Frequency Ranges A fifth and final reason for proposing the frequency domain methodology for time series modeling is that it facilitates the modeling of complex dependencies between the behavior of economic and financial variables in different frequency ranges. At first sight, the zero correlation property described in Section 3.1.3 may seem like a restrictive simplifying feature of the decomposition approach. As explained in Section 3.3.3, by modeling different frequency ranges ‘separately’ we actually get more rather than less possibilities for modeling complex behavior by explicitly modeling relations between the properties of the different frequency models. Examples mentioned in Section 3.3.3 were the ‘level effect’ in the short term volatility of interest and inflation rates, state dependent business cycle dynamics and state dependent asset correlations. 37/40 5 Conclusions In this paper we have described a frequency domain methodology for time series modeling. With this methodology it is possible to construct time series models that in the first place give a better description of the empirical long term behavior of economic and financial variables which is very important for SAA decision making. In the second place, the methodology brings together the empirical behavior of these variables as observed at different horizons and observation frequencies which is required for constructing a consistent framework to be used in the different steps of an investment process. In the third place, the methodology gives insight in and understanding of the corresponding dynamic behavior, both in terms of empirical time series data and of the time series models used to describe this behavior. In the various parts of the paper we introduced the most important frequency domain techniques and concepts, described and illustrated the methodology and finally gave the motivation for doing so. We hope that based on the contents of this paper, more people will be inclined to explore the possibilities of using the appropriate frequency domain techniques for analyzing and modeling time series data and time series processes of economic and financial variables. We are convinced that this can contribute to a higher quality of investment decision making, implementation and monitoring in general, and for Central Banks and Sovereign Wealth Managers in particular. 38/40 References 1. Backus, D.K. and P.J. Kehoe (1992), “International evidence on the historical properties of business cycles”, American Economic Review, 82, 864-888. 2. Baxter, M. and R.G. King (1999), “Measuring business cycles: Approximate band-pass filters for economic time series”, The Review of Economics and Statistics, November, 81(4), 575-593. 3. Beveridge, W.H. (1922), “Wheat prices and rainfall in Western Europe”, Journal of the Royal Statistical Society, 85, 412-459. 4. Blackburn, K. and M.O. Ravn (1992), “Business cycles in the United Kingdom: Facts and Fictions”, Economica, 59, 383-401. 5. Bloomfield, P. (1976), “Fourier analysis of time series: An introduction”, New York. 6. Brillinger, D. R. (1981), “Time Series: Data Analysis and Theory”, San Francisco: Holden-Day. 7. Burg, J.P. (1967), “Maximum entropy spectral analysis”, Paper presented at the 37th Annual International S.E.G. Meeting, Oklahoma City. Reprinted in Childers, D.G. (1978), “Modern Spectral Analysis”, IEEE Press, New York. 8. Campbell, J.Y. and L.M. Viceira (2002), “Strategic asset allocation: Portfolio choice for long-term investors”, Oxford University Press, Oxford. 9. Christiano, L.J. and T.J. Fitzgerald (1999), “The Band Pass Filter”, NBER Working Paper No. W7257. 10. Dimson, E., P. March and M. Staunton (2002), “Millennium book II: 101 years of investment returns”, London: ABN-Amro and London Business School. 11. Englund, P., T. Persson and L.E.O. Svensson (1992), “Swedish business cycles: 1861-1988”, Journal of Monetary Economics, 30, 343-371. 12. Hassler, J., P. Lundvik, T. Persson and P. Soderlind (1992), “The Swedish business cycle: Stylized fact over 130 years”, Monograph, 22, (Institute for International Economic Studies, Stockholm). 13. Jorion, P. and W.N. Goetzmann (2000), “A century of global stock markets”, NBER Working Paper No. W 7565. 14. Lucas, R.E. Jr. (1977), “Understanding business cycles”, in Brunner, K. and A.H. Meltzer (editors), Stabilization of the domestic and international economy, volume 5 of Carnegie-Rochester Series on Public Policy, North-Holland, 7-29. 15. Maddison, A. (2006), “The World Economy: Historical Statistics”, Volume 2, OECD. 16. Reijnders, J. (1990), “Long waves in economic development”, Brookfield. 17. Schmidt, R. (1984), “Konstruktion von digitalfiltern and ihre verwendung bei der analyse ökonomischer zeitreihen”, Bochum. 18. Shannon, C. E. and W. Weaver (1949), “The mathematical theory of communication”, University of Illinois Press, Urbana Illinois. 19. Steehouwer H. (2005), “Macroeconomic Scenarios and Reality. A Frequency Domain Approach for Analyzing Historical Time Series and Generating Scenarios for the Future”, PhD thesis, Free University of Amsterdam. 20. Steehouwer, H. (2007), “On the Correlations between Time Series Filtered in the Frequency Domain”, ORTEC Centre for Financial Research (OCFR) Working Paper. 21. Tinbergen, J. (1946), “Economische bewegingsleer”, N.V. Noord-Hollandsche Uitgeversmaatschappij, Amsterdam. 39/40