Yield Curve Modeling under Cyclical Influence Mats Levander January 2010 Abstract There are few models for long-term yield forecasting and especially in the Real World Measure. This thesis uses Principal Component Analysis (PCA) to analyze the yield curves and gives an update of precedent studies. The conclusion is still that the first three components is enough to describe the variation of the yield curve. For simulation of the yield curves PCA and a semi parametric approach are evaluated. These models fail to yield plausible simulations. Therefore a new model is developed. The model is influenced by a business cycle and a relationship is derived from historical data between the yield curve and the cycle. Nelson-Siegel’s model and OrnsteinUhlenbeck processes are some of the features used in the model. Tools are also introduced to cope with the non-negativity problems in the aftermath of the financial crisis. The model will be used as input to calculate the Swap Breakage Exposure upon an early termination of an interest rate swap. Acknowledgements I would like to thank Nils Hallerström at PK AirFinance for giving me this opportunity and supporting me by sharing his wide knowledge of the financial markets. I would also like to thank the whole PK AirFinance Office in Luxembourg for baring with me during this period, and especially Christophe Beaubron, Frank Withofs and Bo Persson for guidance and discussions. Finally, I would like to thank my supervisor at KTH, Henrik Hult, for his support and ideas during the course of this thesis. Stockholm, January 2010 Mats Levander ii Contents 1 Introduction 1.1 PK AirFinance . . . . . . . . . . . . 1.2 SAFE . . . . . . . . . . . . . . . . . 1.3 The Aircraft Value Cycle . . . . . . 1.3.1 Introduction . . . . . . . . . 1.3.2 Historical evidence . . . . . . 1.3.3 Why this behavior? . . . . . 1.3.4 Amplitude of Cyclical Swings 1.3.5 Prediction of the cycle . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1 1 2 7 7 7 8 9 9 2 Literature Study 13 3 Theoretical Background 3.1 Inflation . . . . . . . . . . . . . . . . 3.2 Present Value . . . . . . . . . . . . . 3.2.1 Compounding of rates . . . . 3.3 Forward Rates . . . . . . . . . . . . 3.4 Yield . . . . . . . . . . . . . . . . . . 3.4.1 Yield Curve . . . . . . . . . . 3.5 Interest Rate Swap . . . . . . . . . . 3.6 Swap Breakage . . . . . . . . . . . . 3.6.1 Example of Swap Breakage . 3.7 LIBOR rate . . . . . . . . . . . . . . 3.7.1 Historical Background . . . . 3.7.2 BBA-LIBOR . . . . . . . . . 3.7.3 Definition . . . . . . . . . . . 3.7.4 Applications . . . . . . . . . 3.7.5 Selection of panels . . . . . . 3.7.6 Calculation . . . . . . . . . . 3.8 U.S. Interest Rate Swaps . . . . . . . 3.9 US Government Debt . . . . . . . . 3.9.1 Treasury Bills . . . . . . . . . 3.9.2 Treasury Bonds and Treasury 17 17 18 18 19 19 20 20 21 21 23 23 24 24 25 25 26 26 26 26 27 iii . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Notes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3.10 Smoothing Spline . . . . . . . . . . . 3.11 Correlation . . . . . . . . . . . . . . 3.11.1 Cross-Correlation . . . . . . . 3.11.2 Cross-Correlation Coefficient 3.12 Least Squares Method . . . . . . . . 3.13 Maximum Likelihood Method . . . . 3.14 Linear Regression . . . . . . . . . . . 3.14.1 The Simple Linear Model . . 3.14.2 Multivariate Models . . . . . 3.14.3 Goodness of Fit . . . . . . . . 3.14.3.1 R2 and adjusted R2 3.15 Principal Component Analysis . . . 3.16 Ornstein-Uhlenbeck process . . . . . 3.17 Nelson-Siegel’s Yield Curve Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28 28 29 30 30 31 31 32 33 33 33 34 35 36 4 Data Selection 39 5 Results 5.1 Principal Component Analysis . . . . . . . . . . . 5.1.1 LIBOR and Swaps . . . . . . . . . . . . . . 5.1.1.1 Different Valuation Basis . . . . . 5.1.1.2 Weekly data . . . . . . . . . . . . 5.1.2 US Treasuries Constant Maturity . . . . . . 5.1.2.1 Daily Data . . . . . . . . . . . . . 5.1.2.2 Weekly data . . . . . . . . . . . . 5.1.2.3 Monthly Data . . . . . . . . . . . 5.1.2.4 Conclusion PCA . . . . . . . . . . 5.1.3 Correlation with the PURCS Cycle . . . . . 5.1.3.1 Crosscorrelation with PURCS . . 5.2 Simulation: PCA . . . . . . . . . . . . . . . . . . . 5.2.1 LIBOR and US Swaps . . . . . . . . . . . . 5.2.1.1 Nominal . . . . . . . . . . . . . . 5.2.1.2 Real . . . . . . . . . . . . . . . . . 5.2.2 US Treasuries Constant Time to Maturity . 5.3 Simulation: Rebonato . . . . . . . . . . . . . . . . 5.3.1 T-notes . . . . . . . . . . . . . . . . . . . . 5.4 Simulation: Brute Force Model . . . . . . . . . . . 5.4.1 Calibration of Ornstein-Uhlenbeck processes 41 41 41 44 46 47 47 51 52 53 53 56 56 57 57 62 63 65 65 68 74 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6 Discussion 81 6.1 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81 6.2 Future Studies . . . . . . . . . . . . . . . . . . . . . . . . . . 82 A PCA Results 85 iv B Maximum Likelihood Estimation Ornstein-Uhlenbeck 95 vi Chapter 1 Introduction 1.1 PK AirFinance PK AirFinance started in 1979 when the Swedish bank PKbanken set up a subsidiary, PK Finans, in Luxemburg. In 1983 PK Finans made their first aircraft financing - a DC-10-30 to Martinair. A couple of years later a separate Air Group was established. Crédit Lyonnais bought 85% of the Air Group in 1988 and it was renamed to Crédit Lyonnais/PK AirFinance. Under the following years the company expanded, e.g. it set up offices in New York and Tokyo, started an Asset Management Group and a first version of a deal analysis model was launched. In 1991 Crédit Lyonnais bought the remaining 15% of Air group. In 2000 the company was bought by GECAS (General Electric Commercial Aviation Services) and renamed to PK AirFinance. An office was set up in Toulouse in 2001 because of the proximity with Airbus. As of today PK AirFinance is a leading provider and arranger of asset based loans that are secured by commercial jet aircraft. The company’s customers are airlines, aircraft traders, lessors, investors, financial institutions and manufacturers worldwide. PK AirFinance’s business activity is to; invest for its own account in aircraft backed loans and securities. This is done by providing or purchasing; senior1 loans or finance leases, junior loans, and minor equity investments in special purpose companies that own aircraft and guarantees. The parent company, GECAS, is the world’s largest aircraft lessor. PK AirFinance has tailor-made financing deals involving more than 75 Airlines worldwide. At the end of 2008 the company’s portfolio volume reached USD 7.7 billion. As of today PK AirFinance has 15 employees. The deal 1 a senior loan has higher priority than a junior loan in a repayment structure 1 analysis model, SAFE2 , was created in 1993. This tool has been subject to continuous refinement during the years. 1.2 SAFE PK AirFinance (PK Air) uses SAFE to analyze, price, underwrite and to continuously monitor deals throughout their term in the portfolio. The model is made up by several modules that contribute to the output of the model. The model has a wide range of applicability, e.g. for determining risk effectiveness, value creation and portfolio management. The graphical user interface is in an Excel environment supported by C++ and databases. In this interface PK Air can alter many different parameters in the analysis of new or existing deals. Examples of parameters are jurisdiction, margins, type of lease, aircraft type and many others. Some of these parameters are analyzed in different subroutines. For example they model the future aircraft value, the probability of default of the obligor and the industrial cycle in subroutines. To put it short, the model boils down to calculation of the distribution of the Net Present Value (NPV) of a deal. In this calculation PK Air generates many different scenarios of the outcome of the NPV. The risk in the deals is when the NPV turns negative and PK Air will experience losses. To value different deals PK Air uses different metrics such as Risk Reward Ratio (Expected Loss divided by Expected Present Value), Average Downside Risk (the average of the negative outcomes in the NPV scenarios) and Value At Risk (VaR). The fundamentals of PK Air’s Asset Backed Financing are • Default Risk • Asset Value • Deal Structure Default Risk is the risk that the airline will default (not being able to meet its payment obligations). The probability of default is linked to the credit ratings of the airline. Another thing that affects the probability of default is the Aircraft Value Cycle which is explained thoroughly in Section 1.3, briefly it can be said to adjust for the economical cycles. Asset value risk is the risk that is attached to the uncertainty of an aircraft’s future value. PK Air has found a way of predicting the aircraft’s 2 Statistical Aircraft Finance Evaluation 2 future values. This model is also adjusted for the Aircraft Value Cycle and is in addition affected by a projected inflation path. The deal structure is the way the deal is set up. This describes the outstanding loan amount or book value over the tenor, fees, rent and interest rate margins, priority ranking over the collateral (aircraft, deposits, guarantees etc.), rights to upside in the aircraft, and swap breakage exposure. There are numerous variations of the these structures, and when the counter party defaults the pay-off function of the lease or the loan will depend on how the deal was set up. This pay-off function is a function of the asset’s resale price. Different deal structure examples are now presented, and the pay-offs are illustrated under the assumption that the borrower/lessee immediately defaults and that the loans are non-recourse3 . • Lease of aircraft bought for USD 20mm • Senior loan USD 20mm • Shared Pari-Passu loan USD 10mm each • Senior Shared loan USD 15mm and a Junior Shared loan USD 5mm First we have the lease contract. If the lessor buys an aircraft for USD 20mm and leases it to an airline the pay-off function takes the shape of Lease USD 20mm in Figure 1.1. Upon default the lessor repossess the aircraft and may sell it on the market since they own it. Therefore the lessor can at most experience a loss of USD 20mm but the upside is unlimited in theory. A Senior loan is conducted by lending USD 20mm to the airline to finance their aircraft with the aircraft as collateral. If the airline defaults, they have to sell the aircraft in the market to repay their debt. The best scenario is that the lender gets his USD 20mm back, and the worst is loosing all USD 20mm. The lender is not entitled to any upside since they do not own any aircraft in this deal. If two lenders enter a shared Pari-Passu4 loan, lending out USD 10mm each to the airline. This results in the payoff function of the Senior loan being ”halfed”, e.g. if the aircraft is sold for USD 10mm both will experience losses of USD 5mm. 3 non-recourse means that you can only go after the aircraft to recover your money, not the owner 4 Pari-Passu is latin and may be translated as ”with equal force”, for a loan it means that you have equal recovery rights 3 In a shared Senior-Junior loan settlement the pay-off shape changes. If we have the Senior loan of say USD 15mm and the other lessor has USD 5mm as Junior, we will only experience losses after a sale price of less than USD 15mm, since the Junior tranche experiences the first loss of USD 5mm. If we instead have the Junior tranche the pay-off will have the shape of Junior USD 5mm. There is a vast amount of hybrids of these kind of structures but these lie outside of the scope of this thesis. Pay−off Deal Structures 10 Pay−off USD mm 5 0 −5 −10 Lease USD 20mm Senior Loan USD 20mm Pari−passu USD 10mm Senior USD 15mm Junior USD 5mm −15 −20 0 5 10 15 20 25 30 Sale Price USD mm Figure 1.1: Examples of Deal Structure Pay-offs upon default Another part of the deal structure is the Repayment Schedule. It is the agreement on how the borrower will repay its debt to the lender, i.e. the points in time and the magnitude of the cash flows from the borrower to the lender. As an example a fictitious schedule is shown in Figure 1.2 (additional fees and margins has been left out in this example). The total outstanding debt is 25 million USD that PK Air has lent to a borrower. To finance this a lender would typically take a loan of 25 million USD in the floating rate market, and to hedge themselves they will swap a fixed rate for floating matching the cash flows. The fixed rate is calculated as an effective rate so that the net present value of the cash flows is zero, see Section 3.6 4 for an example. The repayment schedule is as follows: the borrower repays 0.5 million USD every 6 months and after 15.5 years he pays the remaining residual of 10 million USD. PK Air has in turn matched this repayment setting with their lender. Fictitious Repayment Schedule 25 Debt [Million USD] 20 15 10 5 0 0 2 4 6 8 10 12 14 Time [years] Figure 1.2: Fictitious Repayment Schedule 5 16 If PK Air’s borrower defaults or there is an early repayment option of the loan PK Air will not receive the planned cash flows. Instead they have to finance their upcoming repayments with the present market setting, that is with the yield curve at that time. Depending on the yield curve, there might be a gain or loss from this refinancing procedure when the net present value and the corresponding fixed rate is calculated. Therefore the Exposure at Default is uncertain and dependent on the propagation of the yield curve. A fictitious illustration of repayment schedule outcome simulations with percentiles is shown in Figure 1.3. It shows that the uncertainty is greatest in the middle of the repayment period if we are at time zero today. In the beginning and in the end of period the uncertainty diminishes to become certain at the endpoints. Figure 1.3: Fictitious Repayment Schedule Simulation At today’s date PK Air uses a model that uses a worst case scenario, the yield curve simulation decays exponentially with time. PK Air is not pleased with this way of modeling and wants to improve this simulation. They wish to have a model that is tractable, has few parameters, is fast to simulate and it should be influenced by the cycle described in the next Section. 6 1.3 1.3.1 The Aircraft Value Cycle Introduction The common view is that the Aircraft business is cyclical. For example the delivery of aircraft, aircraft orders, air traffic and many other factors vary cyclically with time. The same thing holds for aircraft values. The outcome of buying, selling, leasing and/or financing aircraft will be very dependent on the timing of an investment or divestment compared to the aircraft value cycle. 1.3.2 Historical evidence PK Air has built up a database with historical resale prices of aircraft, which at the moment contains around 4000 sales transactions since 1970. By looking at specific data for each aircraft type PK Air has been able to analyze depreciation patterns, cyclicality and volatility statistically. In Figure 1.4 the resale price of the Boeing 737 300 series expressed in constant 1984 USD value is shown. The pattern around the trend line is cyclical. Boeing 737−300 Series Resale Price in 1984 $ Boeing 737 resale price Trend 25 Price 20 15 10 5 1990−01−01 2000−01−01 Date Figure 1.4: Boeing 737-300 Resale Price in constant 1984 USD over time PK Air has from their broad study of different aircraft type values found out that the cyclical variation is almost perfectly correlated in time between different types. 7 1.3.3 Why this behavior? Classic economic theory is built upon the interaction between supply and demand. When the demand for an asset is higher than the supply the price will rise to find an equilibrium and vice versa. The supply and demand mechanisms for aircraft seems not to be in equilibrium for most of the time. The explanation to this is the lag when the aircraft manufacturing industry tries to adapt to the current market situation. Their pace of manufacturing is directly dependent on the predictions of future traffic made by the airlines. The manufacturing time is normally 12-18 months. This explains the adjustment lag; by the time of delivery of the aircraft, the market might have changed and therefore the predictions and so on. This makes the traffic very hard to predict in the short term. The demand for aircraft is estimated by air traffic, which can be expressed as revenue passengers miles per year. The relationship between production (revenue passenger miles; RP M ) for one aircraft, load factor LF , block speed V , utilisation hours U and number of seats S is defined as: RP M = LF · V · U · S (1.1) The time series of the RP M growth are shown in Figure 1.5 RPM Growth 0.2 0.15 0.1 0.05 0 −0.05 1980−01−01 2000−01−01 Date Figure 1.5: RPM Growth 1968 to 2009 This relationship also holds for airlines and the global air traffic market. Explanation of the parameters: 1. The load factor has been improved due to management system and new scheduling techniques. On the one hand a high load is good for 8 the airline’s profit, on the other hand opportunity costs may arise from high load factors. For example the airline might have to reject passengers in peak times and vice versa if there is a trough. 2. The block speed was improved greatly during the 1960’s when the aircraft obtained pressurised cabins, and later by first the turbine engines and then the jet engines. Today the block speed is considered as constant. 3. Utilisation has improved over the years due to longer range aircraft and optimized scheduling. There are limiting factors though, in general people do not fancy flights departing or arriving in the middle of the night or early morning. On average the utilisation is rising since the long haul traffic is increasing faster than the short haul traffic. The reason is that long haul planes spends more hours per day than the short haul planes. For the short term airlines are capable to adjust load factor and utilisation to shifts in demand and supply. There is a however a point in time when they reach the maximum of these parameters, and they must therefore expand by buying more seats or the other way around if the times are tough. Fortunately, aircraft are globally traded assets and may therefore easily be transfered between operators. 1.3.4 Amplitude of Cyclical Swings In Figure 1.6 the annual growth of physical installed seats (delivered less scrapped) is shown and the cumulative difference between RP M and Seat growth adjusted for the trend is shown in Figure 1.7. We can see the cyclical behavior in the graph which is the driving process for the aircraft value swings. Amplitudes greater than zero is a measure of the pent-up relative capacity shortage in the global aircraft fleet, and an amplitude less than zero describes a pent-up surplus. Let us denote this measure pent-up relative capacity shortage/surplus (PURCS ). In the peaks, airlines have high load factors and the aircraft is in use many hours per day. There will be a demand for aircraft so prices will go up. In troughs, the converse holds. 1.3.5 Prediction of the cycle To know whether to invest or not we need an prediction about the future. The (PURCS ) cycle is affected by; traffic growth and seat growth. If we 9 Seat Growth 0.35 0.3 0.25 0.2 0.15 0.1 0.05 0 1980−01−01 2000−01−01 Date Figure 1.6: Seat Growth 1968 to 2009 PURCS Cycle 0.15 0.1 0.05 0 −0.05 −0.1 −0.15 −0.2 1980−01−01 2000−01−01 Date Figure 1.7: PURCS Cycle 1968 to 2009 10 could predict those for the time horizon of a cyclical swing (approximately 5 years) we would know when the peaks or troughs appear. Looking at the RPM growth in Figure 1.5 above we could see that the volatility is high. The notoriously hard predicted GDP Growth drives the short term variations. For prediction PK Air has two in-house teams, which consists of members from both the parent company GECAS and PK Air. The first team is responsible for predicting the RPM Growth by predicting the parameters in equation (1.1). The second team predicts Seat Growth that is defined as Seat Growth = F + P − C − R (1.2) where F is current fleet seats, P is produced seats, C is converted seats (origins from aircraft being converted from passenger to cargo aircraft (the converse does not happen)) and R is retired seats. In the simulations in the model, RPM Growth and Seat Growth are not simulated independently, instead PK Air starts out in the latest known peak or trough and simulates the coming peaks and troughs. In the simulation there are stochastic variables representing the uncertainty in the occurrence (time) and the level of the next peak or trough. The predictions are revised and the PURCS Cycle is updated and improved on a quarterly basis. 11 12 Chapter 2 Literature Study There is an enormous amount of models and research on interest rate modeling. Most of the papers focus on short rate models, e.g. Hull-White, Ho-Lee, Vasicěk and Heath Jarrow Morton (HJM) and Cox Ingersoll Ross (CIR). All these models is priced under the risk neutral measure. The parameters in the models (eg. reversion speed, jump frequencies) implicitly embed, after being calibrated to market prices, a component related to risk aversion and therefore contain drift terms different from the objective measure. The drift terms creates a gap between the risk-neutral and the real world evolution of the yield curve [17]. Using these short rate models in long horizons results in evolutions that becomes totally dominated by the no-arbitrage drift terms and do hardly possess any virtual resemblance to the real-world evolution. For example Vasicěk and CIR cannot even recover an arbitrary exogenous yield curve, and the market price of risk is specified (proportional or proportional to the square root of the short rate) have been chosen to allow analytical tractability. Another problem with these models is that they are often supposed to be used with short time horizons such as days or weeks, e.g. for banks that trade with interest rates. PK Air’s typical deals have an average life of around 9 years, i.e. the time horizon is much longer than for banks. In addition their simulations are done in the real world measure, not in a riskneutral measure, and they price the risk at the end of their simulations. To limit the scope of this study and to adapt to PK Air’s time horizon, I have chosen in agreement with my academic supervisor and PK Air, not to look at the short rate models. Instead the focus is on Principal Component Analysis (PCA) and curve fitting techniques to understand and model the movements of the yield curve. Litterman and Scheinkman [11] investigated the common factors that affect the returns on U.S. government bonds and similar securities. They compared duration analysis with PCA for hedging purposes. The duration 13 concept is used to describe parallel shifts of the yield curve, but in real life, the yields do not always move in parallel shifts. With PCA they found three factors affecting the returns of the bonds and that they explained 96 % of the variation. The factors are referred to as the level, the steepness and the curvature. The first factor explained 89.5% of the variation, the steepness 8.5% and the curvature 2.0%. They concluded that this PCA representation is better than traditional duration hedging. Gloria Soto [21] evaluated the performance and stability of PCA to explain term structure movements. Comparisons were done with typical oneand two-factor interest rate models. She focused on the Spanish government debt market from January 1992 to December 1999. The result was that the PCA with with three factors (level, slope, curvature) outperformed the oneand two-factor models. There were however some concerns with the stability of the PCA over time. The model’s performance deteriorated significantly when the model is estimated from the most recent data in contrast to longsample data. This points towards the use of factor models with dynamic volatility structures. Scherer and Avellaneda [20] used PCA to study the Brady Bond Debt of Argentina, Brazil, Mexico and Venezuela from 1994 to 1999. They found that the two first principal components explains up to 90% of the variation. The analysis was split up into windows to show how the components vary with time and tried to link those to different economic events. Because of the short series of data, they looked at daily yield changes, since their results with weekly or monthly observations were dominated by noise and lacked structure. For the analysis they looked at absolute yield changes. They also verified that the results were invariant to valuation basis, that is if they e.g. used relative changes. Their opinion was that PCA on long observation windows was best for mature markets with stable economic cycles, which is compatible with the mathematical definition of stationarity. This was not the case for this market and they therefore used 120-day windows. They conclude that there was a high risk for under-estimating the behavior of the market if considering only a static PCA over the whole period. Wesley Phoa [16] also compared traditional interest rate management such as duration management with PCA on the US Treasury market. The PCA was performed with daily changes in yield on Constant Maturity Treasuries. He pointed out that the result will only be meaningful if a consistent set of yields is used and therefore he used Constant Maturity Treasury yields. An alternative approach was to use historical swap rates, since these are par yields by definition. The results said that the most important shifts of the yield curve were the level and the slope. These two shifts can be estimated quite precisely and robust over time. The curvature shifts tended to be 14 more varying and were highly dependent on the dataset used. When using shorter term to maturities the variance explained by the level diminished a little and the slope and curvature increased slightly. The importance of the three components for bonds was also shown to be similar for all large countries e.g. the U.S., Germany, France, UK, Australia and Japan. He also pointed out that some risk factors are ignored with PCA, for example if there would be a shift in the curve for maturities from 30 to 100 years relative to shorter yields. This would not show up in the PCA since the range only goes to 30 years. Another yield curve risk that does not show up using Constant Maturity Treasuries is the the yield spread between liquid and illiquid assets. This was a large factor in the US Treasury market in 1998 that was greater than the curvature factor. Rebonato, Mahal, Joshi, Buchholz and Nyholm [17] evolved a method to construct a future yield curve over a period of years using a simple semiparametric approach. They used PCA to analyze percentage changes of USD LIBOR and Swap yields from 29th of September 1993 to 4th of December 2001. They found that the first component explains more than 90% of the variation. To model future yields they tried different models. The first one, referred to as the naive simply used historical simulation to evolve future yields. The drawing of historical yields, was done by randomly choosing a starting date and then randomly choosing a window of consecutive yields that should be included. In this manner a yield curve was evolved. This procedure preserved the co-dependence structure of the original data across maturities. But they found that the assumption that the increments were IID did not hold. The method fails to capture the serial autocorrelation structure and the cross-sectional co-dependence. The difference from IID behavior was so strong that the procedure was unsatisfactory. To cope with this they introduced mean-reversion for the shortest and longest maturities and also introduced ”springs” for maturities in between. The springs were introduced to match the curvatures of the empirical data. The last refinement consisted of a jump frequency in the sampling windows, that is a small probability of jumping out of the sampling window and start a new sampling. By these enhancements they managed to replicate the statistics of the empirical data very well. They claimed that one possible application of the model is off-balance sheet transaction, e.g. swaps, which exposes the counter-party to potential future credit exposure, which is dependent of the deal will be in the money at the time of a possible default. Rebonato and Nyholm [18] extends their paper from 2005 by comparing their semi-parametric approach with a parametric approach for long-horizon yield curve forecasts. The parametric approach is a model by Bernadell, Coche and Nyholm [3] that is a three-factor representation of observed yields built upon the Nelson and Siegel methodology [14]. They found that their semi15 parametric model discovers the same features as the parametric model without the use of Markov chains. The data focused on was US Treasury yields with maturities; 3 months, 6 months, 1 years, 2 years, 5 years, 10 years, 20 years and 30 years from 1st of January 1986 to 7th of May 2004 at a daily frequency. Fiori and Iannotti [7] used Principal Component Analysis and Monte Carlo Simulation to assess Italian banks interest rate exposure. They looked at the Euro Area par yield curve with maturities from 1 month to 30 years from 4th of January 1999 to 30th of September 2003. The first three components explained 95% of the total variation. Instead of simulating from the normal distribution, which have been proved to be empirical incorrect, they focused on heavy-tailed distributions. In particular they used an Gaussian kernel estimator with optimal bandwidth. The distribution functions of these were obtained by integrating the kernel densities. From these Value-a-Risk calculations were made. They compared the parametric simulation (Normal) with the non-parametric (Gaussian kernel). The results showed that the parametric approach has limitations when the interest rates are increasing, especially when interest rates were low as in December 2001. Duffee [6] investigated the idiosyncratic variation of Treasury Bill Yields and found that Treasury bills of three month or less term to maturity exhibits price movements that are idiosyncratic, i.e. they are not related to changes in other interest rates. He believed that this is caused by an increased market segmentation. Nelson and Siegel [14] introduced a simple and perspicuous model for the shapes of the yield curve. They applied it to US Treasury bill and Treasury bonds yields. The parametric model was successful in representing typical yield curve shapes: monotonic, humped and S-shaped. They found that the model explained 96 % of the variation in bill yields across maturities from 1981 to 1983. The model is explained more thoroughly in Section 3.17. Bank of Canada investigated the Nelson Siegel and Svensson parametric yield curve models [5]. They investigated the optimization problem, the robustness and data filtering applied to Canadian Government Securities. They found that the Svensson model was the best model. 16 Chapter 3 Theoretical Background 3.1 Inflation Inflation is the rate of change in general prices over time [12]. It can be expressed in terms of an inflation rate i. The prices 1 year in the future can on average be expressed as today’s prices multiplied by a factor (1 + i). The inflation compounds in a similar way to interest rates, that is after k years at the same inflation rate i, the future prices will be the original prices times (1 + i)k . In reality the inflation rate changes over time. Inflation can also be seen as it reduces the purchasing power of money. A classic example is that a dollar today does not yield as much bread and milk as a dollar did 10 years ago. Therefore we can think of the prices as increasing or the value of money as decreasing. If we assume that the inflation rate today is i, then the value of a dollar next year expressed in today’s purchasing power will be 1/ (1 + i). When conducting a study over a certain time period, it can be very convenient to express prices in the same kind of dollars. Therefore we consider constant (real) dollars, which are defined relative to a reference year (typically the starting year). This yields that the dollars will maintain the purchasing power of the dollars for the reference year over time. These dollars should be distinguished from the actual (nominal) dollars that are used in daily transactions. From this we are able to define a new interest rate, namely the real interest rate, defined as the rate that real dollars increase if they were put into a bank account that pays nominal interest rate. To illustrate this, think of putting money into a bank account at time zero and then withdrawing them 17 after a year. The purchasing power of the bank balance can be expressed as 1 + r0 = 1+r , 1+i (3.1) where r is the nominal interest rate, i is the inflation rate and r0 is the real interest rate. This equation tells us that the money in the bank grows with 1 (1 + r) nominally but it is at the same time deflated by 1+i . Rearranging the expression we obtain r−i r0 = . (3.2) 1+i For small inflation levels the real rate is approximately equal to the nominal rate minus the inflation rate. 3.2 Present Value The idea behind the concept of Present Value (PV) [12] is to describe the time value of money. If you invest 1 USD at the bank today at a rate rf (consider as the risk-free rate) you money will be worth 1 · erf ·1 after one year using continuous compounding. On the other hand 1 USD recieved one r ·1 f year from now is worth 1/ e USD today which is also referred to as the present value. The procedure of transforming future cash-flows into today’s value is called discounting. 3.2.1 Compounding of rates There are two ways of compounding rates; frequent and continous compounding. If we assume that the risk-free is rf and the interest is compounded at m equally spaced times per year. We do also assume that the cash flows start out at time zero and at the end of each period, and that the total number of periods is n. Then we obtain cashflows as follows (x0 , x1 , ..., xn ) and this yields the following formula for the present value PV = n X xk [1 + r/m]k k=0 (3.3) If we instead assume that the interest is compounded continuously and that the cash flows occur at times t0 , t1 , ..., tn . Then the formula boils down to PV = n X x (tk ) e−rtk k=0 18 (3.4) 3.3 Forward Rates Forward rates [12] are interest rates for money to be borrowed between two future time points, but the terms are agreed upon today. If we for example consider a 2-year horizon. Assume that the spot rates s1 and s2 are known and given on a yearly basis. After two years 1 USD would have grown to (1 + s2 )2 USD in a 2-year account. One other option is to first put the money in a 1-year account for one year, which will yield (1 + s1 ) after one year. If we then already have agreed to borrow the money for 1 more year at a rate f , then we will recieve (1 + s1 ) (1 + f ) at the end of year 2. f is called the forward rate. These two ways of borrowing money should be equal if there is no arbitrage and therefore we get the following equation (1 + s2 )2 = (1 + s1 ) (1 + f ) (3.5) which can be solved for f f= (1 + s2 )2 −1 (1 + s1 ) (3.6) as can be seen the forward rate is determined by the two spot rates s1 and s2 . 3.4 Yield The yield [12] of a bond is the rate of interest that is implied by the payment structure. More precise it is the interest rate for which the present value of the bond’s cash flow streams (coupons and face value) is exactly equal to its current price. The more formal term of yield is yield to maturity (YTM) to separate it from other yield measures. Yields are always quoted on annual basis. The yield is exactly the internal rate of return of a bond at its current price, but yield is the term used in the market. The calculation for the yield, y, of a bond maturing at time T using continuous compounding is −y(T −t) PV = F · e + N X Ci · e−y(T −ti ) , (3.7) i=1 where F is the face value of the bond, N is the number of coupons and Ci is the coupon paid at time ti . Equation (3.7) must be solved for y to determine its yield. It can only be solved by hand for very simple cases, otherwise it has to be done by an iterative process, which can easily be implemented on a computer. 19 3.4.1 Yield Curve The general conditions in the fixed-income securities is closely related to the yield of a bond [12]. In general all yields move together in this market, but all yields are not exactly the same. The reason for the variation in yield over bonds is that they have different credit quality ratings. An AAA-bond (highest rating) has in general a higher price and thus has a lower yield than a B-bond with the same maturity. There are however other explanations to the differences in bond yields. Time to maturity is another factor that affects the yield of a bond. Intuitively bonds with long time to maturity tend to offer higher yields than bonds with shorter time to maturity with the same rating. Therefore a ”normally” shaped yield curve has higher yield with longer time to maturity. It does in fact happen that bonds with longer time to maturity have lower yields than bonds with shorter time to maturity, this is termed inverted yield curve. This shape appears when the shortterm rates increases rapidly, but the investors believe that the increase is only temporary, and therefore the long-term rates (which is an expectation about the future) remain close to their former level. 3.5 Interest Rate Swap A swap [10] is a an agreement between two counter-parties to exchange cash flows according to a prearranged formula in the future. The first swaps were settled in 1981 and since then the market has grown fast. Many billions of dollars of swap contracts are negotiated each year. The ”plain vanilla” interest rate swap is the most common one. The setup is that one party, A, agrees to pay another party, B, cash flows of a predetermined fixed rate on a given notional principal at specific points in time during an interval. A the same time party B pays cash flows to party A, based on a floating rate on the given notional amount over the same time period. The currency is the same for the two interest rate cash flows. The swaps have duration for 2 to 15 years. Swaps appear because companies have comparative advantages. One company may have an advantage in the fixed rate market and the other in the floating market. The companies tend to borrow money in the market they have an comparative advantage. However they might want to match a cash flow in the other market, this is where the swap is used to transform the rate. The floating rate used in these agreements is typically the LIBOR rate, see Section 3.7 for further details. Financial institutions have a natural advantage funding themselves in the floating rate market. This implies that a interest rate swap occurs when 20 they want to match for example a deal with an airline who wants to borrow money with a fixed rate. Then PK Air turns to their lender and matches a floating rate with a fixed rate that is offered to the airline. All of PK Air’s interest rate swaps agreements so far has been in US Dollars. Figure 3.1: Interest Rate Swap 3.6 Swap Breakage Swap breakage occurs when a swap agreement is terminated early, for PK Air there could be two causes of this; one is an early repayment and the other is a default of the counter party. The fixed rates are of course at the same level throughout the agreement. The floating rate will naturally change over time. The fixed rate was calculated from the yield curve at time t0 = 0. On the time of default tdef ault , the yield curve will look different and the floating rates would imply another fixed rate for the remaining term. In the present value calculation there is a high probability that the risk-free rate that is used for discounting has also changed. The present value can be both positive and negative and of course also zero, resulting in an income, cost or zero result breakage. 3.6.1 Example of Swap Breakage Say that we want to swap a floating rate for a fixed rate for a time interval of 5 years and that the swaps are made on a yearly basis. The swap is on 21 a notional amount, N , of 50 000 USD. Assume that today’s yield curve has the following setting: Table 3.1: Fictional Yield Curve Term to maturity 1 yr 2 yr 3 yr 4 yr rate 0.0058 0.0124 0.0186 0.0233 5yr 0.0270 where the rates are compounded on a yearly basis. From this we can calculate the implied forward rate fij between year i and j. For the second year this is (1 + 0.0124)2 = (1 + 0.0058) (1 + f12 ) if solved for f12 f12 = (1 + 0.0124)2 − 1 = 0.0190 (1 + 0.0058) in the same manner we can calculate all forward rates. Table 3.2: Implied Forward Rates interval 0-1 yr 1-2 yr 2-3 yr 3-4 yr forward rate 0.0058 0.0190 0.0312 0.0375 4-5 yr 0.0419 To calculate the fixed rate, we calculate the Present Value (PV) in thousands USD of the cashflow streams and we set the PV to zero and solve for rf ixed 50 (0.0058 − rf ixed ) e−0.0058·1 + ... + (0.0419 − rf ixed ) e−0.0270·5 = 0 0.0058 · e−0.0058·1 + ... + 0.0419 · e−0.0270·5 = 0.0265 e−0.0058·1 + ... + e−0.0270·5 If we now assume that the agreement is terminated after 1 year and we have the following yield curve at that time ⇒ rf ixed = Table 3.3: New Fictional Yield Curve Term to maturity 1 yr 2 yr 3 yr 4 yr rate 0.0122 0.0152 0.0223 0.035 22 As before we back out the implied forward rates Table 3.4: New implied Forward Rates interval 0-1 yr 1-2 yr 2-3 yr 3-4 yr forward rate 0.0122 0.0182 0.0367 0.0740 From this we obtain the following Present Value P V = (0.0122 − 0.0265) · e−0.0122·1 + ... + (0.0740 − 0.0265) · e−0.035·4 · 50 = +1.7058 If the counter-party defaults we will have a Swap Breakage Gain of 1705.8 USD. 3.7 LIBOR rate The information in this section was found at one of the British Bankers Association (BBA) webpages [2]. 3.7.1 Historical Background In the early 1980’s an increasing number of banks in the London market were actively trading new instruments as currency options and interest rate swaps. Most of the banks found these new instruments attractive but they were at the same time bothered about that the underlying rates that had to be agreed on before entering the contract. The BBA were therefore asked by the banks they represented to create a uniform measure for the market and produce a benchmark index. The concept was that banks could now reference their contracts against a standard rate which removed the negotiation of the underlying rate. This improved the generation of instruments. In 1984 BBA cooperated with for example the Bank of England and others which resulted in the BBAIRS, that is the BBA standard for Interest Swap rates. In this standard the fixing of BBA Interest Settlement rates, the predecessor of BBA-LIBOR, was included. BBAIRS became standard market practice 23 the 2nd September 1985. The first BBA-LIBOR rates were given in 1986 and in three currencies; US Dollars, Japanese Yen and Sterling. Today it is calculated in 10 currencies for 15 maturities. Since the BBA-LIBOR was introduced there has been two significant changes. Firstly in 1998 the question was changed from ”At what rate do you think interbank term deposits will be offered by one prime bank to another prime bank for a reasonable market size today at 11 am?”. The new question, which is still used today is ”At what rate could you borrow funds, were you to do so by asking for and then accepting inter-bank offers in a reasonable market size just prior to 11 am?”. Secondly in 1999 with introduction of the Euro on the 1st of January 1999 the number of currencies BBA-LIBOR was calculated for obviously diminished. 3.7.2 BBA-LIBOR BBA-LIBOR stands for London InterBank Offered Rate. The rate is derived for 10 currencies with 15 maturities for each, the shortest being overnight and the longest being 12 months. The BBA-LIBOR should be interpreted as a benchmark, it provides you with an idea of what the average rate of a prominent bank, for a specific currency, can get unsecured funding for a certain term in the specific currency. In other words, it yields the lowest real-world cost of unsecured funding in the London market. The rates are derived by a calculation based on submissions from a panel, consisting of the most active banks in the specific currency. 3.7.3 Definition It is important to know that BBA-LIBOR is based upon the offered rate, and not the bid rate. Each contributing bank is ask to base their BBALIBOR submission on the the question mentioned above: ”At what rate could you borrow funds, were you to do so by asking for and then accepting inter-bank offers in a reasonable market size just prior to 11 am?”. In this way the submissions are based upon the lowest perceived rate that a bank that participates in a certain currency panel could go into the inter-bank money market and get sizable funding for a certain maturity. The BBA-LIBOR is not derived from actual transactions, it would be hard to have this as an requirement since not all banks require funds in marketable size each day for the currencies and maturities they quote. This should not be interpreted as that the rates does not reflect the true cost of 24 interbank funding. The banks are aware of their credit and liquidity risk profiles for the rates it has dealt with, and are therefore able to build a curve that accurately predicts the correct rates for currencies or maturities which it has not quoted at the moment. All BBA-LIBOR rates are quoted as annualized interest rates, which is a market convention. For an example say that the overnight rate is given as 2 %, this does not imply that that the overnight loan interest rate is 2 %. Instead it means that the annual loan interest rate is 2 %. 3.7.4 Applications BBA-LIBOR is the primary benchmark for short term interest rates in the world. The rate is also used as an indicator of the strain in money markets and it is also often used as a measure of the market’s expectations of future central banks interest rates. Different studies show that approximately 350 trillion USD of swaps and 10 trillion USD of loans are indexed to BBALIBOR. It is the foundation for settling interest rate contracts on most of the world’s major futures and options exchanges. 3.7.5 Selection of panels The contributing banks are chosen for currency panels with the goal of reflecting the market balance for a specific currency based upon these three guidelines 1. Scale of market activity 2. Credit rating 3. Perceived expertise in the currency concerned Each of the 10 panels, with 8 to 16 contributors, is chosen by the independent Foreign Exchange and Money Markets Committee (FX & MM Committee) to obtain the best way of representing the activity in the London money market for a certain currency. Therefore BBA-LIBOR submissions from panel members will on average be the lowest interbank unsecured loan offers from the ones available on the money market. The FX & MM Committee evaluates each panel every year, based upon a review from BBA of the contributors. The review evaluates each bank according to the criteria listed above. The review is not limited to only present contributors as any Bank can submit itself to the evaluation process for any currency. 25 3.7.6 Calculation The calculations are made by Thomson Reuters. They analyze the data from the banks and calculate the rates according to the definitions given by FX & MM Committee. The whole procedure is supervised by BBA. In the contributing banks every cash desk has an application from Thomson Reuters installed. Each morning the currency dealer takes their own rates of the day and puts them in to the application between 11.00 and 11.20. The banks are not able to see the other banks rates until they are submitted by Thomson Reuters. The data is also submitted by nine other data vendors. All the BBA-LIBOR rates from Thomson Reuters are calculated in the same way using an trimmed arithmetic mean. The contributors are ranked by Thomson Reuters and sorted in descending order. The trimming is done by dropping the top and bottom quartiles. The remaining 50 % of the data are then averaged to get the BBA-LIBOR quote. The reason to drop the top and bottom quartiles is to improve the accuracy of the quotes by removing outliers. By dropping these the possibility for any individual contributor to influence the quote is removed. 3.8 U.S. Interest Rate Swaps These rates can be used as a proxy for the continuation of the LIBOR rates. These are quoted as the fixed rates swapped for 3 month LIBOR semiannually. 3.9 3.9.1 US Government Debt Treasury Bills The U.S. government offers debt obligations called Treasury bills [19] , also known as T-bills. Of the total U.S. government marketable debt the T-bills occupy approximately one-fourth. In the middle of 1995, there was around three-quarters of a trillion dollars worth of bills outstanding. Since the bills are liabilities of the government, these obligations are considered as default free. The secondary market for these are one of the most active, characterized by its low bid-ask spread and extremely high liquidity. The T-bills are issued at discount from the face value and have no coupons or stated interest. The earnings for the holder is simply the difference between the discount issue price and the face value, which is paid out by the Treasury at time of maturity. The discount is determined in an auction, where new bills are offered to dealers and other investors. The auction is conducted by the Federal Reserve Bank of New York acting on behalf of 26 the U.S. Treasury. The bills are sold to those who offer the highest price, and this auction ensures that the resulting interest costs of the Treasury is minimized. The discount is related to the term of the bond and what investors believe about the market future. T-bills are offered in the following maturities: 91-day (3-month), 182-day(6month) and 364-day(1-year). The three- and six-month bills are auctioned every week, and the one year every fourth week to meet the huge demand for funds from the U.S. government of refinancing an outstanding debt. Since 1993, the bills that have been issued in multiples of 1000 USD, with 10000 USD as the smallest amount. The amount which has the lowest commission rates is the one known as a round lot and is 5 million USD. These are normally traded by large market participants. The commission rates is generally in between 12.50 to 25 USD per 1 million USD and is affected by the maturity of the bill, three-months have the lowest commission. No certificates or papers of the debt is issued when the T-bills are issued. The claim is only registered in the computer system of the Federal Reserve. Due to their low risk and short maturity, T-bills are very popular instruments for market participants. The range of holders goes from individuals to governments and everything in between. For individuals and commercial investors the T-bills is interesting since they are relieved from state and local taxes. Foreign banks are large holder of bills, their interest is mostly based on the safe return of the bills. For the U.S. government the T-bills are an important way to raise funds to finance the U.S. debt outstanding. The Federal Reserve does also hold a lot of bills to be able to use its monetary instruments to influence the economy. 3.9.2 Treasury Bonds and Treasury Notes Contrary to the T-bills, the Treasury bonds (T-bonds) and Treasury notes (T-notes) [19] are interest-bearing securities with maturities greater than one year. The difference between notes and bonds is that they are issued in different maturities. Notes are issued with maturities of 2, 3, 4, 5, 7 and 10 years and bonds with maturities greater then 10 years, e.g. 30 years. The primary market for these securities is a bit complicated. Almost all Treasury debt that is negotiable is offered at auctions on a yield basis. The speculators submit sealed bids with the lowest yield to maturity that they would accept, instead of regular bidding auction. The bids are conducted at local Federal Reserve Banks and they are opened at a prespecified time in the future. The bids specify the volume and the yield (in percent in two decimal places). There is also a possibility for small investors to place a noncompetitive bid, which means that they will pay a price equal to the mean of the accepted bids. The Treasury itself does intervene the process 27 by setting ”stop-out” levels, which is the highest level that they are willing to issue debt. At the prespecified times the average level of the accepted bids are calculated. Then the Treasury sets a coupon on the debt closest to 1/8 of 1 percent, where the level will yield an average price of the successful bids equal to 100 USD (per 100 USD of face value) or lower. 3.10 Smoothing Spline If the available data set is very noisy, e.g. daily yields of a certain maturity, it might be a good idea to fit your data with a smoothing spline [13]. A smoothing spline, s, is constructed for a given smoothing level p and weights wi . The objective of the spline is to minimize p X Z 2 wi (yi − s (xi )) + (1 − p) i d2 s dx2 2 dx if the weights, wi , are unspecified all points are assigned a weight of 1. The smoothing level, p, is defined between 0 and 1. The lower limit produces a least-square straight line fit to the data. If you set it to the upper boundary it produces a cubic-spline interpolant. An application of this procedure is given in Section 5.1.3 later on. 3.11 Correlation To get an idea of what correlation is, consider the joint density function of two random variables [1]. This density could be said to describe a mountainlike shape. If the mountain is symmetric around the two axes of the random variables, little information about one variable is obtained by knowing the value of the other variable, i.e. the correlation is low between the variables. On the other hand, if the shape of the joint density function is ridge-like between the axes, the correlation will be high. To get a feeling for the correlation, scatter plots is useful. A scatter plot is a plot of two synchronous returns of financial time series against each other, e.g. gold returns against crude oil. A low correlation corresponds to symmetrical dispersed scatter plot, i.e. a high value of one variable does not imply a high/low value of the other. A high correlation is described by a ridge between the axes, if it has negative the slope the correlation is negative, and vice versa. Correlation measures the co-movements between two return series. Strong positive correlation says that an upward movement in one return series is accompanied by an upward movement in the other series. For strong negative 28 correlation an upward movement in one series corresponds to a downward movement in the other series. The covariance measure can be used to measure the co-movements of two random variables X and Y , it is defined as Cov (X, Y ) = E [(X − µX ) (Y − µY )] (3.8) where µX and µY are the means of the random variables. The drawback with this measure is that it is not only determined by the co-movements of the returns but also by the size of them, that is in general monthly returns will have higher covariance than daily returns. Since covariance is not independent of the units of measurement, it is not suitable to make comparisons. To cope with this the correlation measure is introduced. It is a standardized form of the covariance that it is independent of the units of measurement. For two random variables it is standardized by dividing the covariance by the product of their standard deviations, that is Cov (X, Y ) Corr (X, Y ) = p V (X) · V (Y ) (3.9) where V (X) and V (Y ) are the variances. This standardization procedure will yields a correlation value of -1 to +1, where the first refers to perfect negative correlation and the latter to perfect positive correlation. If the two random variables are statistically independent their correlation coefficient should be insignificantly different from zero, and the variables are referred to as orthogonal. Notice that the converse is not necessarily true, if two random variables are orthogonal does not imply independence (they could have zero covariance but still be related by the higher moments of their joint density function). 3.11.1 Cross-Correlation The autocorrelation function could be used to determine a sequence’s structure in the time domain. Cross-correlation uses the same concept but instead of comparing a time shifted version of the signal with itself, it compares to different sequences with each other [9]. The cross-correlation function of two sequences {Xt } and {Yt } is defined as T X 1 Xt Yt+h T →∞ 2T + 1 φX,Y (h) = E (Xt Yt+h ) = lim (3.10) t=−T The function is a statistical comparison of two sequences as a function of the time shift between them. It is very useful as a practical tool for determining timing differences between the sequences. 29 3.11.2 Cross-Correlation Coefficient The Cross-Correlation Coefficient is closely linked to the usual correlation coefficient and is defined as φX,Y (h) `X,Y (h) = p φX,X (h = 0) φY,Y (h = 0) (3.11) as with the traditional correlation coefficient, the value of the cross-correlation coefficient lies between -1 and 1. It can be used to determine for which lag h to sequences obtain the highest correlation. 3.12 Least Squares Method Let x1 ,...,xn be outcomes of independent stochastic variables X1 ,...Xn . The means of the stochastic variables are known except for an unknown parameter. Therefore it is assumed that E (Xi ) = µi (θ) for i = 1, 2, ..., n where µ1 , µ2 , ..., µn are known and θ is an unknown parameter with parameter space ΩΘ . This implies that Xi = µi (θ) + i , i.e. Xi = ”known function of θ” + ”error”. These errors are assumed to have zero means. In general their variances are assumed to be identical. Let Q (θ) = n X [xi − µi (θ)]2 (3.12) i=1 be the sum of the observations deviations from the µi (θ)’s. The expression between the hard brackets are recognized as the error of observation i if θ is the correct parameter value. The method is to use the θ-value that minimizes the sum of squares as θ-estimate. To find this Least Squares Estimate of θ [4] the derivative dQ (θ) /dQ is calculated and set equal to zero to find the minimum. A more general method can be applied if the distribution contains k unknown parameters θ1 , θ2 , ..., θk . The expression for Q is now a function of these parameters Q (θ1 , θ2 , ..., θk ) = n X wi [xi − µi (θ1 , θ2 , ..., θk )]2 (3.13) i=1 where w1 , ..., wn are weights that could be assigned to the observations. In general these are set to 1, which corresponds to equal uncertainty for all data. The Least Square Estimate is as before determined by minimization, e.g. by setting the partial derivatives of Q to zero and solving for θi . 30 3.13 Maximum Likelihood Method The idea behind the Maximum Likelihood Method is to as an estimate of θ use the value that makes our sample data as likely as possible [4]. The value of θ will therefore naturally depend on our sample data and will be ∗ of θ. the Maximum Likelihood Estimate, θobs Let x1 ,...,xn be outcomes of independent stochastic variables X1 ,...Xn that has a distribution that depends on an unknown parameter θ with parameter space Ωθ . In most cases X1 ,...Xn are assumed to be independent and that they have the same distribution, but the method also works without these assumptions. If X is a continuous stochastic variable it has density function f (x; θ), and in the discrete case it has probability mass function p (x; θ). If the sample is from a discrete distribution and all the Xi0 s are independent, will the probability of obtaining P (X1 = x1 , ..., Xn = xn ) be P (X1 = x1 , ..., Xn = xn ) = P (X1 = x1 ) · · · P (X1 = x1 ) = pX1 (x1 ; θ) pX2 (x2 ; θ) · · · pXn (xn ; θ) (3.14) (3.15) The Likelihood Function is defined as P (X1 = x1 , ..., Xn = xn ; θ) (discrete) L (θ) = fX1 ,...fXn (x1 , ..., Xn = xn ; θ) continuous The idea behind the Maximum Likelihood method is to let the argument θ take all values in Ωθ and find the value of θ that maximizes the function. ∗ and is the estimate. In the discrete case the idea beThis value is called θobs hind the method is to maximize the probability of obtaining the sample data. When maximizing L (θ) it is often convenient to maximize lnL (θ). Since the logarithm is a monotonic increasing function will both L (θ) and lnL (θ) have maximums at the same point. The Likelihood function is normally a product and therefore the logarithm yields a sum which facilitates the maximization which is done by taking the derivative and solving for it equal to zero. 3.14 Linear Regression Linear regression models are formulated in this way Y = β1 X1 + β2 X2 + ... + βk Xk (3.16) On the left side is the dependent variable, Y , and on the right there are the independent variables X1 , X2, ..., Xk [1]. The independent variables can 31 also be called explanatory variables. The coefficients β1 , β2 , ..., βk are model parameters and they measure the influence of its corresponding independent variable on Y . X1 is in general assumed to be equal to 1 so the model includes a constant β1 . The goal of the regression is to find estimates of the true parameter values and predictions of the dependent variable by using historical data on the dependent and independent variables. 3.14.1 The Simple Linear Model The simplest case is when k = 2 and X1 = 1, i.e. we have a constant for all t. For notational purposes the constant is set to α (interpreted as the intercept with the vertical axis), and β2 is denote by β (the slope of the line) and X2 is denoted X. This results in the following setting Yt = α + βXt (3.17) To see the relationship between the data all pairs of (Xt , Yt ) is plotted against each other in a scatter plot. All the points will not lie along a straight line so an error process is introduced to the equation Yt = α + βXt + t (3.18) An estimated straight line through the scatter plot yield a predicted or fitted value of Yt for each Xt from Ŷt = α̂ + β̂Xt (3.19) where α̂ and β̂ are the estimates of the intercept and the slope. The difference between the real value of Y and the fitted value Y at time t is denoted t and is called the residual at time t, i.e. t = Yt − Ŷt . So each real data point Yt is described by Ŷt = α̂ + β̂Xt + t (3.20) The estimates are obtained by minimizing the sum of the squares of the residuals, which is known as ordinary least squares (OLS) criterion. OLS is an unbiased estimation. The OLS estimates for the Simple Linear Model are given by X Xt − X̄ Yt − Ȳ β̂ = 2 P t t Xt − X̄ α̂ = Ȳ − β̂ X̄ where Ȳ and X̄ denotes the sample means of X and Y . 32 (3.21) (3.22) 3.14.2 Multivariate Models In the multivariate case the model takes the form Yt = β1 X1t + β2 X2t + ... + βk Xkt + t (3.23) and assuming that the model has a constant term yields X1t = 1 for all t = 1, ..., T . There will be T equations with k unknown parameters. To simplify things the model can be written in matrix notation. Let the dependent variable column vector be y = (Y1 , Y2 , ..., YT )0 and put the independent variables into a matrix X where the jth column of X corresponds to the data on Xj . The first column of X will be a column of 1s if there is a constant in the model. Denote the vector of true parameters β = (β1 , β2 , ..., βk )0 and let = (1 , 2 , ..., T )0 be the vector of error terms. The representation of the model then boils down to Y = Xβ + (3.24) The matrix form of the OLS estimators of β is given by −1 0 β̂ = X0 X Xy 3.14.3 (3.25) Goodness of Fit There are numerous ways of measuring how well the model describes the data. Some measures are presented below. 3.14.3.1 R2 and adjusted R2 R2 is the coefficient of determination of a model. It is a product of the concept of analysis of variance (ANOVA). It has three key metrics;the total sum of squares, T SS = (y − ȳ)T (y − ȳ), the explained sum of squares, ESS = (ŷ − ȳ)T (ŷ − ȳ) and the residual sum of squares, RSS = T . R2 is the proportion of the total sum of squares that is explained by the model T SS RSS R2 = =1− (3.26) ESS T SS From this measure it is possible to tell how much of the variation in y that is explained by βX. Adjusted R2 takes into account the number of explanatory variables in the 2 increases only if the new term improves the model. Different from R2 , Radj model more than expected by chance. The measure can be negative and it is less or equal to R2 . The definition is 2 Radj =1− RSS/ (T − k − 2) T SS/ (T − 1) (3.27) where T is the sample size and k is the number of explanatory variables. 2 is in general used for determining if an explanatory variable improves Radj the model 33 3.15 Principal Component Analysis This section follows the derivation in the book by Alexander [1]. Before the analysis can be done, the data must be stationary. This is done by taking the absolute difference from day to day in the T × K rate data matrix X. Then you normalize your data matrix X by subtracting each column with its mean µ and standard deviation σ to obtain a new matrix X∗ with mean 0 and standard deviation 1. Principal Components Analysis (PCA) starts out from the symmetric correlations matrix of the variables in X∗ V= X∗T · X∗ T (3.28) the components are found by calculating the eigenvalues and eigenvectors of the correlation matrix V. It will be shown that the principal components can be described as a linear combination of these columns. The components will be orthogonal to each other since they are eigenvectors. The first principal component will describe the most of the variance in X∗ , the second the second most and so forth. This is achieved by choosing the weights from the k × k eigenvector matrix of V in the following way VW = WΛ (3.29) where Λ is the k × k diagonal matrix of eigenvalues from V. Next the columns of W is ordered according to the size of their corresponding eigenvalue. Therefore if wij for i, j = 1, ...k, then the mth column of W, wm = (w1m , ..., w1k ) is the k × 1 eigenvector that corresponds to eigenvalue λm . The columns of W is ordered so that λ1 > λ2 > ... > λk . The mth principal component is defined in the following way Pm = w1m · X1∗ + w12 · X2∗ + ...w1k · Xk∗ (3.30) where Xi∗ is the ith column of the normalized matrix X∗ . This can be written as Pm = X∗ · wm (3.31) The complete T × m matrix of principal components can thus be expressed as P = X∗ · W (3.32) To understand that this leads to orthogonal (uncorrelated) components, notice that 0 PP = WT X∗ T X∗ W = {X∗ T · X∗ = T V} = T WT WΛ = {W is orthogonal WT = W−1 } = T Λ ,where Λ is the diagonal matrix of eigenvalues 34 Therefore the vectors of PT P are uncorrelated. Each principal components variance is determined by its corresponding eigenvalues proportion of the total variation. For example the variation in X∗ explained by Pm is given by λm λm = Σλii k where k is the number of variables in the system. (3.33) If the components in the original system is highly correlated the first eigenvalue will be much greater than the others, i.e. the first principal component explains the most of the variation. Equation (3.32) can be expressed 0 0 as X∗ = P · W because W = W−1 , in vector form Xi = wi1 P1 + wi2 P2 + ... + wik Pk (3.34) It is therefore possible to represent each input data vector as a linear combination of the principal components. This is known as the Principal Component Representation of the original variables. It is often enough to only include the first few principal components, since these explain almost all the variation. 3.16 Ornstein-Uhlenbeck process An Ornstein-Uhlenbeck process is a mean-reverting process given by the stochastic differential equation dXt = κ (θ − Xt ) dt + σdWt (3.35) where κ, θ and σ are parameters and Wt is a standard Brownian Motion. κ is the speed of mean-reversion and θ is the mean and σ is the volatility. To solve the equation we introduce a change of variable, letting Yt = Xt − θ. Then Yt satisfies the following Stochastic Differential Equation dYt = dXt = −κYt dt + σdWt (3.36) In equation 3.36 the process dYt has a drift towards zero at an exponential rate κ. Therefore a change of variable is motivated to remove the drift. This is done by Yt = e−κt Zt ⇔ Zt = eκt Yt , (3.37) Applying Itô’s formula to equation 3.37 yields dZt = κeκt Yt dt + eκt dYt κt (3.38) κt (3.39) = 0dt + σe dWt (3.40) = κe Yt dt + e (−κYt dt + σdWt ) κt 35 The solution is obtained by integrating from s to t Z t eκu dWu Zt = Zs + σ (3.41) s Retransforming the change of variables yields Yt = e −κt Zt = e −κ(t−s) −κt t Z eκu dWu , (3.42) e−κ(t−u) dWu . (3.43) Ys + σe s and finally Xt = Yt + θ = θ + e −κ(t−s) Z (Xs − θ) + σ t s 3.17 Nelson-Siegel’s Yield Curve Model Nelson and Siegel formulated their model to fit the shape of the yield curve in the following way m m 1 − e− τ 1 − e− τ m y (m) = β0 + β1 + β2 − e− τ (3.44) m/τ m/τ where m is term to maturity, β0 is the long term component, β1 is the short term component, β2 is the medium term component. The long term component is a constant throughout the maturity, the medium component starts out at zero and decays to zero and is therefore only affecting the medium term. The short term component has the fastest decay monotonically to zero. τ determines the if the fit will be best at short or long maturities and the location of the maximum of the medium component β2 . A graphical illustration is shown in Figure 3.2. Some constrains has to be imposed on the model to obtain plausible results when fitting it to empirical yields. First β0 , the long term component has to be greater than zero. Secondly, β0 + β1 has to be greater than zero to guarantee non-negative yields at the short end. Thirdly, the decay parameter, τ has to lie in the maturity spectrum, i.e. be greater than zero. 36 Influence of Nelson−Siegel Components with τ = 2 1 0.8 β0 0.6 β 1 β2 0.4 0.2 0 0 5 10 15 20 25 30 Term to Maturity Figure 3.2: Nelson-Siegel components influence 37 38 Chapter 4 Data Selection The floating rate that PK Air uses as a reference is the LIBOR rate, this data was obtained from Bloomberg. Since the LIBOR rates longest maturity is 12 months, US Swap rates have been used as a proxy for longer time to maturity. The idea was to look at the most homogeneous data sets. The time series of data were available in the following setting for daily quotes: Table 4.1: LIBOR and US Swap Time Series 1988-2009 LIBOR 1 m 2 m 3 m 6 m 12 m USSWAPS 2 yr 3 yr 4 yr 5 yr 7 yr 10 yr Since the time series only go back to 1988, there was a need for time series who stretched further back in time. Therefore US Government Securities yields were also studied, specifically Treasuries with constant time to maturity [15]. The longest homogeneous time series were available from 2nd of January 1962 and were composed in this way and were available in daily, weekly and monthly quotes. Table 4.2: US Treasuries Constant Time to Maturity Series 1962-2009 Treasury 1 yr 3 yr 5 yr 10 yr Another setting with more maturities were available from 1982 Table 4.3: US Treasuries Constant Time to Maturity Series 1982-2009 Treasury 3 m 6 m 1 yr 3 yr 5 yr 10 yr 39 40 Chapter 5 Results 5.1 5.1.1 Principal Component Analysis LIBOR and Swaps Starting out with the LIBOR and Swaps rates some problems appeared with the raw data. It turned out that the rates had sometimes been quoted on different days, typically one rate was missing out on a certain day. To deal with this problem interpolation was applied to obtain homogeneous data sets. It was now possible to construct a yield surface Figure 5.1: Nominal Yield Surface LIBOR and Swaps 1988 to 2009 41 Secondly the rates were transformed to real rates using the relation (3.1). The US inflation data was given on a monthly basis [8] and was applied to the rates that corresponded to that month. This yielded a real yield surface as can bee seen the yield surface shape changes. Figure 5.2: Real Yield Surface LIBOR and Swaps 1988 to 2009 Principal Component Analysis has an application to term structures [1]. Term structures are special because they yield an ordering of the system that gives an intuitive interpretation of the principal components as shall be seen. Before performing the PCA it is important that the time series are stationary, the yields are in general not stationary , therefore absolute differences of the yield data were considered, X, as shown in Figure 5.3. As a last step the differences are normalized, that is subtracting the mean and dividing by the standard deviation for each of the columns to obtain a new matrix X∗ , a graphical illustration of this for 1 month LIBOR can be seen in Figure 5.4. 42 1 month nominal LIBOR absolute changes 0.015 0.01 0.005 0 −0.005 −0.01 −0.015 1990−01−01 2000−01−01 date Figure 5.3: 1 month nominal LIBOR absolute changes 1 month nominal LIBOR normalized absolute changes 20 15 10 5 0 −5 −10 −15 −20 1990−01−01 2000−01−01 date Figure 5.4: 1 month nominal LIBOR normalized absolute changes 43 From the normalized absolute changes the correlation matrix was obtained Table 5.1: Correlation Matrix LIBOR and Swaps 1 m 2 m 3 m 6 m 1 yr 2 yr 3 yr 4 yr 5 yr 7 yr 10 yr 1 m 1 0.620 0.555 0.463 0.377 0.112 0.108 0.105 0.089 0.086 0.087 2 m 3 m 6 m 1 yr 2 yr 3 yr 4 yr 5 yr 7 yr 10 yr 1 0.648 0.556 0.439 0.129 0.105 0.096 0.091 0.089 0.083 1 0.741 0.624 0.196 0.163 0.148 0.143 0.139 0.125 1 0.757 0.229 0.205 0.186 0.179 0.171 0.160 1 0.273 0.249 0.228 0.222 0.212 0.202 1 0.934 0.899 0.887 0.858 0.7924 1 0.935 0.928 0.906 0.842 1 0.936 0.918 0.860 1 0.957 0.916 1 0.936 1 The correlations shows typical yield curve behavior. The correlation diminishes with the spread between the maturities. Another observation is that the longer term to maturity, the higher the correlation with the neighboring maturities than for shorter maturities. Table 5.2: Eigenvalues of Correlation Matrix LIBOR and Swaps Principal Component Eigenvalue Cumulative R2 P1 5.8070 52.8 P2 3.0414 80.5 P3 0.7745 87.5 P4 0.3842 91.0 The first three components, which are referred to as the level, the slope and the curvature explains more than 87% of the variation. The table does also show that the first component corresponds to the largest eigenvalue. Figure 5.5 shows a graphical illustration of the first three principal components (PCs). A hump occurs for the shorter maturities, this could be explained as shall be seen later on by the inclusion of the shorter maturities. The first PC, the level, is approximately constant over the maturities. The second PC, the slope, approximately raises the yield for maturities up to 2 years and lowers it for longer. The third PC, the curvature, approximately raises the yield for all maturities except the interval 3 months to 3 years. 5.1.1.1 Different Valuation Basis What happens if the valuation basis is changed from absolute returns to log returns? Log returns is defined as Xn = logSn − logSn−1 44 (5.1) Principal Component Analysis 0.8 1st PC 2nd PC 3rd PC 0.6 Change in yield 0.4 0.2 0 −0.2 −0.4 −0.6 0 1 2 3 4 5 6 7 8 9 10 Time to maturity Figure 5.5: Principal Components nominal LIBOR and Swaps where Si is the yield at day i. The answer is that the results are almost the same as shown in Figure 5.6 and Table 5.3 45 Principal Component Analysis 0.8 1st PC 2nd PC 3rd PC 0.6 Change in yield 0.4 0.2 0 −0.2 −0.4 −0.6 −0.8 0 1 2 3 4 5 6 7 8 9 10 Time to maturity Figure 5.6: Principal Components nominal log LIBOR and Swaps Table 5.3: Variance Explained log LIBOR and Swaps 1988-2009 Component Cumulative R2 P1 54.03 P2 84.11 P3 90.79 P4 93.68 The results are therefore indifferent of the valuation basis, the small differences can be explained by numerical errors in MATLAB. 5.1.1.2 Weekly data The same procedure was done for weekly data on the yields of LIBOR and US Swaps. The Correlation Matrix changed as can be seen in Table 5.4. 46 Table 5.4: Correlation Matrix LIBOR and Swaps Weekly 1 m 2 m 3 m 6 m 1 yr 2 yr 3 yr 4 yr 5 yr 7 yr 10 yr 5.1.2 5.1.2.1 1 m 1 0.679 0.734 0.637 0.469 0.310 0.274 0.257 0.229 0.209 0.191 2 m 3 m 6 m 1 yr 2 yr 3 yr 4 yr 5 yr 7 yr 10 yr 1 0.839 0.753 0.577 0.346 0.303 0.284 0.271 0.247 0.222 1 0.857 0.726 0.442 0.394 0.365 0.343 0.323 0.311 1 0.824 0.591 0.537 0.502 0.482 0.447 0.423 1 0.641 0.598 0.562 0.547 0.515 0.494 1 0.979 0.941 0.926 0.890 0.873 1 0.968 0.963 0.938 0.914 1 0.964 0.946 0.922 1 0.976 0.951 1 0.962 1 US Treasuries Constant Maturity Daily Data The Treasury yield curve go back from 1962, but from then data are only available for maturities from 1 to 10 years. This longest yield surface is shown in Figure 5.7. Figure 5.7: T-notes 62-09 Nominal Yield Surface 47 If the whole period is analyzed at once, the following results are obtained 1 1st PC 2nd PC 3rd PC 0.8 Change in yield 0.6 0.4 0.2 0 −0.2 −0.4 −0.6 −0.8 1 2 3 4 5 6 7 8 9 10 Time to maturity Figure 5.8: Principal Components Nominal T-notes 62-09 48 Table 5.5: Variance Explained T-Notes 1962-2009 Component Cumulative R2 P1 89.05 P2 96.43 P3 98.89 The level explains almost all the variation. The time interval between 1962 to 2009 was also split into smaller windows of 5 years. The visual results of how the principal components propagates over time can be seen in Appendix A. The total variance explained is shown in Table 5.6. Table 5.6: Variance Explained T-notes 1962-2009 5 year windows Cum. R2 P1 P2 P3 62-67 80.66 91.99 96.91 67-72 86.94 94.81 98.42 72-77 81.28 91.95 97.50 77-82 91.26 96.61 98.82 82-87 91.46 96.90 98.93 87-92 91.37 97.76 99.25 92-97 92.66 98.02 99.40 97-02 89.21 97.15 99.20 02-07 89.81 97.52 99.48 07-09 84.48 96.73 99.43 From the table is clear that the level explains more than 80 % of the total variation and all three components explains more than 96 %. This implies that when leaving out the shorter term to maturities the explanatory power increases. From the graphical illustrations of the components the ”hump” that occurred with the LIBOR and US Swaps has now disappeared. The explanation is that the shorter term to maturities are not included in the T-notes and that they do not covary with longer maturities as stated by Duffee [6]. Another observation is that the level is approximately constant over the different time windows. The same relation holds for the slope and the curvature has the same shape except for the intervals 1967 to 1977. The next dataset is from 1982 to 2009 and the nominal yield surface is shown in Figure 5.9 and compared to the previous set, maturities of 3 months and 6 months were included. 49 Figure 5.9: Nominal T-bond Yield Surface 1982-2009 The principal components are shown in Figure 5.10 Principal Component Analysis 0.8 1st PC 2nd PC 3rd PC 0.6 Change in yield 0.4 0.2 0 −0.2 −0.4 −0.6 0 1 2 3 4 5 6 7 8 9 10 Time to maturity Figure 5.10: T-bond 1982-2009 Principal Components 50 Here the ”hump” resurrects again and could as before be originated to the shorter maturities. According to Table 5.7 the same pattern continues, the level explains the most of the variation, although it has diminished compared to the window-study above, one possible explanation for this is the inclusion of shorter maturities. Table 5.7: Variance Explained T-bonds 1982-2009 Component Cumulative R2 P1 78.47 P2 93.54 P3 96.60 5.1.2.2 Weekly data For the weekly data I started out with the T-notes (1 to 10 year maturities) from 1962-2009. The principal components are shown in Figure 5.11 Principal Component Analysis T−notes Weekly 1962 − 2009 0.8 1st PC 2nd PC 3rd PC 0.6 Change in yield 0.4 0.2 0 −0.2 −0.4 −0.6 −0.8 1 2 3 4 5 6 7 8 9 10 Time to maturity Figure 5.11: Principal Components absolute changes Nominal Weekly Tnotes 1962-2009 51 Table 5.8: Variance Explained T-Notes Weekly 1962-2009 Component Cumulative R2 P1 91.83 P2 97.95 P3 99.42 From Table 5.8 we see that the level itself explains almost all variation. Another observation is that explanatory power is higher than for the daily data. 5.1.2.3 Monthly Data The principal components of the monthly data for the period 1962-2009 is shown in Figure 5.12 Principal Component Analysis T−notes Monthly 1962 − 2009 0.8 1st PC 2nd PC 3rd PC 0.6 Change in yield 0.4 0.2 0 −0.2 −0.4 −0.6 −0.8 1 2 3 4 5 6 7 8 9 10 Time to maturity Figure 5.12: Principal Components absolute changes Nominal Monthly Tnotes 1962-2009 52 Table 5.9: Variance Explained T-Notes Monthly 1962-2009 Component Cumulative R2 P1 93.43 P2 98.84 P3 99.77 Table 5.9 shows that the variation explained increases when the frequency of the data decreases, e.g. from daily to monthly data. 5.1.2.4 Conclusion PCA The results are indifferent of valuation basis, i.e. with relative or absolute changes. The explanatory power increases when the data frequency decreases. When shorter maturities were added to the analysis, the explanatory power decreased. This can be observed by comparing the result of LIBOR and Swaps with T-notes. It was also confirmed that the first three Principal Components explains almost all the variation in the system. These results are analogue with previous research. 5.1.3 Correlation with the PURCS Cycle Since the PURCS Cycle is the driver of all different subsimulations in SAFE, PK Air would also want it to affect the evolution of the yield curve in some sense. Therefore the correlation between PURCS and the yields was studied. First PURCS was compared against 1 month Nominal LIBOR, the series are shown in Figure 5.13 and a scatter point of their absolute changes is shown in Figure 5.14. As can be seen in the scatter plot they do not seem to co-vary at all and the correlation coefficient is calculated to 0.0592. So there is no correlation between these two time series? It turned out that there actually is! The problem lies in the scope of Signal Theory. Since the PURCS Cycle is given on a yearly basis and the 1 month LIBOR on a daily basis, the latter will be very noisy compared to former. Therefore the 1 month LIBOR should be smoothed to give a fair description of the correlation. For this purpose Smoothing Splines were used in MATLABs Curve Fitting Toolbox (cftool), see Section 3.10. An illustration of the different smoothing levels is shown in Figure 5.15. The next step was to analyze how the correlation was affected by the smoothing level and the result is given in Figure 5.16. As can bee seen the correlations is about 0.6 and to retain the most information in the original data the smoothing factor was chosen to 0.995. 53 PURCS and 1 month Nominal Libor time series 0.12 1 month Libor PURCS 0.1 0.08 0.06 0.04 0.02 0 −0.02 −0.04 −0.06 −0.08 1980−01−01 1990−01−01 2000−01−01 2010−01−01 Date Figure 5.13: PURCS and 1 month Nominal LIBOR Time Series abs PURCS vs abs 1 month Libor 0.015 0.01 0.005 0 −0.005 −0.01 −0.015 −3 −2.5 −2 −1.5 −1 −0.5 0 0.5 1 1.5 −3 x 10 Figure 5.14: absolute changes of PURCS vs 1 month Nominal LIBOR 54 Smoothed Libors vs Interpolated PURCS 0.12 0.1 0.08 0.06 rate 0.04 0.02 0 Smooth 0.1 Smooth 0.2 Smooth 0.5 Smooth 0.8 Smooth 0.995 Smooth 0.99964 0.999993 Smooth 1 Smooth PURCS 0 −0.02 −0.04 −0.06 −0.08 1990−01−01 2000−01−01 Date Figure 5.15: PURCS and 1 month Nominal LIBOR with different smoothing Correlation with PURCS vs Smoothing Factor 0.7 0.6 0.5 ρ 0.4 0.3 0.2 0.1 0 −0.1 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Smoothing Factor Figure 5.16: Correlation between PURCS and 1 month Nominal LIBOR with different smoothing 55 5.1.3.1 Crosscorrelation with PURCS To determine if the correlation could be higher if the two time series were shifted crosscorrelation was also studied for the different absolute changes of the smoothing factors. It could be determined that the highest correlation was obtained for lag h = 0, i.e. for the original times series not shifted. A graphical illustration is shown in Figure 5.17. Smooth 0 vs PURCS Smooth 0.5 vs PURCS 0.1 1 0.05 0.5 0 0 −0.05 −0.1 −1 −0.5 0 0.5 −0.5 −1 1 lag x 10 Smooth 0.995 vs PURCS −0.5 0 0.5 1 4 lag x 10 Smooth 1 vs PURCS 4 1 0.05 0.5 0 0 −0.5 −1 −0.5 0 0.5 lag 1 −0.05 −1 −0.5 0 lag 4 x 10 0.5 1 4 x 10 Figure 5.17: Crosscorrelation Coefficient between PURCS and 1 month Nominal LIBOR with different smoothing factors 5.2 Simulation: PCA The idea was to simulate future yield curves using the The Principal Component Representation Xi = wi1 P1 + wi2 P2 + ... + wik Pk . (5.2) Since the first three principal components explained 93.68 % of the total variation, focus was only on these and (5.2) boils down to Xi = wi1 P1 + wi2 P2 + wi3 P3 + , where ∼ N (0, V ar (wi4 P4 + ... + wik Pk )) 56 (5.3) (5.4) To do the simulation we keep the eigenvectors, wi1 , wi2 , wi3 fixed and study the distributions of the first three principal components, P1 , P2 , P3 . 5.2.1 5.2.1.1 LIBOR and US Swaps Nominal The distribution of the first PC is shown in Figure 5.18, the best fits for the distributions are obtained with a t4.57 -location-scale1 distribution with parameters as shown in Table 5.10. The estimation was done in MATLAB using dfittool. 0.25 1st PC Normal t−location−scale Density 0.2 0.15 0.1 0.05 0 −10 −5 0 5 10 Data Figure 5.18: Distribution of Principal Component 1 LIBOR and Swaps 1 If the random variable X is tp distributed and µ and σ are estimated then the corresponding location-scale variable is Y = µ + σX. 57 Table 5.10: Estimate of t-location-scale PC 1 LIBOR and Swaps Parameter Estimate Standard Error µ 0.0089 0.029 σ 1.837 0.032 ν 4.571 0.300 The distribution of the second PC is shown in Figure 5.19 and the parameter estimation in Table 5.11 0.45 PC 2 Normal t−location−scale 0.4 0.35 Density 0.3 0.25 0.2 0.15 0.1 0.05 0 −15 −10 −5 0 5 10 15 Data Figure 5.19: Distribution of Principal Component 2 LIBOR and Swaps 58 Table 5.11: Estimate of t-location-scale PC 2 LIBOR and Swaps Parameter Estimate Standard Error µ 0.0680 0.014 σ 0.7841 0.016 ν 2.014 0.073 For the third component the following t-location-scale estimation was obtained Table 5.12: Estimate of t-location-scale PC 3 LIBOR and Swaps Parameter Estimate Standard Error µ -0.003 0.0059 σ 0.33 0.0074 ν 1.89284 0.071 From these findings the simulation was done by Monte Carlo simulation from the corresponding t-location scale distributions. The number of days simulated was 8000 days (about 16 years) since this is the average length of PK Air’s deals. To make the simulations more stable the simulations are done 1000 times and then averaged. The positive thing with averaging is that the simulations become more stable, but the negative thing is that the yields flatten out and lose some of their original shape. The simulation is started from the following yield curve from the simulaTable 5.13: Fictional Starting Yield Curve TTM yield 1m 0.01 2m 0.015 3m 0.0175 6m 0.02 1yr 0.023 2yr 0.025 3yr 0.027 4yr 0.028 5yr 0.029 7yr 0.0295 10yr 0.0296 ∗ tion according to (5.3) the normalized changes Xsim were obtained. These were tranformed back to Xsim using ∗ Xsim = Xsim · V (Xi ) + E (Xi ) (5.5) where Xi is the original changes column i, e.g. the column for changes of 3 month LIBOR over time. The next step was to transform back the yield data by Si+1 = ∆i,i+1 + Si (5.6) where Si are the yields at time i and ∆i,j is the absolute yield change from time i to j. A simulation for 8000 days conducted 1000 times and then averaged resulted in the following yield surface The result is not that pleasing since the yields go negative. The surface 59 Figure 5.20: Simulated Surface for 8000 days Nominal LIBOR and Swaps also shows a hump for shorter maturities. Lots of the movements have been flattened out because of the averaging. A Principal Component Analysis on the simulated changes correlation matrix is visible in Figure 5.21. Principal Component Analysis Simulated Libor and Swaps 0.8 1st PC 2nd PC 3rd PC 0.6 Change in yield 0.4 0.2 0 −0.2 −0.4 −0.6 0 1 2 3 4 5 6 7 8 9 10 Time to maturity Figure 5.21: Principal Components Simulated Nominal LIBOR and Swaps 60 If comparing with Figure 5.5 the level component have shifted from approximately -0.4 to 0.4, the slope has also been inverted, but the curvature remains in the same setting. The correlation matrix of the simulated data is shown in Table 5.14 Table 5.14: Correlation Matrix Simulated Nominal LIBOR and Swaps 1 m 2 m 3 m 6 m 1 yr 2 yr 3 yr 4 yr 5 yr 7 yr 10 yr 1 m 1 0.975 0.662 0.324 0.104 -0.015 -0.036 -0.030 -0.047 -0.042 -0.034 2 m 3 m 6 m 1 yr 2 yr 3 yr 4 yr 5 yr 7 yr 10 yr 1 0.811 0.525 0.320 -0.023 -0.057 -0.060 -0.079 -0.078 -0.074 1 0.923 0.813 0.025 -0.035 -0.063 -0.081 -0.090 -0.098 1 0.974 0.064 -0.002 -0.041 -0.055 -0.069 -0.082 1 0.113 0.049 0.006 -0.004 -0.020 -0.036 1 0.998 0.994 0.993 0.991 0.989 1 0.999 0.998 0.998 0.996 1 0.999 0.999 0.998 1 0.999 0.999 1 0.999 1 61 5.2.1.2 Real The normalized original change matrix X ∗ had the setting shown in Table 5.15 The simulation was conducted in the same manner as earlier and it Table 5.15: Correlation Matrix Real LIBOR and Swaps 1 m 2 m 3 m 6 m 1 yr 2 yr 3 yr 4 yr 5 yr 7 yr 10 yr 1 m 1 0.876 0.869 0.839 0.794 0.677 0.667 0.662 0.662 0.666 0.673 2 m 3 m 6 m 1 yr 2 yr 3 yr 4 yr 5 yr 7 yr 10 yr 1 0.901 0.873 0.822 0.691 0.677 0.670 0.674 0.678 0.683 1 0.938 0.893 0.744 0.728 0.720 0.725 0.728 0.732 1 0.928 0.746 0.733 0.723 0.727 0.729 0.732 1 0.738 0.725 0.714 0.717 0.718 0.720 1 0.974 0.959 0.955 0.944 0.920 1 0.974 0.971 0.962 0.938 1 0.974 0.967 0.944 1 0.983 0.967 1 0.975 1 yielded the yield curve in Figure 5.22. It showed the same pattern as for the nominal yields, although not as neagtive, but still inplausible. Figure 5.22: Simulated Surface for 8000 days 1000 times Real LIBOR and Swaps A PCA on the simulated changes yielded the following results The variance explained is displayed in Table 5.16 The simulated correlation matrix is given in Table 5.17 62 Principal Component Analysis Simulated Libor and Swaps 0.6 1st PC 2nd PC 3rd PC 0.4 Change in yield 0.2 0 −0.2 −0.4 −0.6 −0.8 0 1 2 3 4 5 6 7 8 9 10 Time to maturity Figure 5.23: Principal Components Simulated Real LIBOR and Swaps Table 5.16: Variance Explained Simulated Real LIBOR and Swaps 2008 2009 Component Cumulative R2 P1 89.17 P2 94.88 P3 100 5.2.2 US Treasuries Constant Time to Maturity For comparison purposes the US Treasuries were studied. First the longest set of data is analyzed, that is from 1962 to 2009. For this interval T-notes were only available with maturities of 1, 3, 5 and 10 years. The correlation matrix had the following setting The simulated surface can be seen in Figure 5.24 The resulting surface for this setting is more pleasing than for the LIBOR and Swaps rates. An upward trend is apparent. PCA of the simulated data are shown in Figure 5.25 and the correlation matrix in Table 5.19 From these results the conclusion was that the simulations were highly dependent on the propagation of the historical datas used as a foundation for the simulations. For the LIBOR and Swaps there has been a downward trend since 1988 and the simulations therefore continue in the same pattern. For the T-notes the historical data were Λ-shaped but with more upward trends than downward and the simulations exhibit an upward trend. 63 Surface for 8000days 0.06 0.05 rate 0.04 0.03 0.02 0.01 0 10 5 0 Term to maturity (years) 6000 4000 2000 0 8000 Days Figure 5.24: Simulated Surface for Nominal T-notes 1962-2009 Principal Component Analysis Simulated T−notes 62−09 1 1st PC 2nd PC 3rd PC 0.8 Change in yield 0.6 0.4 0.2 0 −0.2 −0.4 −0.6 −0.8 1 2 3 4 5 6 7 8 9 10 Time to maturity Figure 5.25: Principal Components Simulated Yields 64 Table 5.17: Correlation Matrix Simulated Real LIBOR and Swaps 1988 2009 1 m 2 m 3 m 6 m 1 yr 2 yr 3 yr 4 yr 5 yr 7 yr 10 yr 1 m 1 0.982 0.86 0.730 0.595 0.768 0.765 0.770 0.770 0.776 0.789 2 m 3 m 6 m 1 yr 2 yr 3 yr 4 yr 5 yr 7 yr 10 yr 1 0.938 0.846 0.736 0.832 0.825 0.826 0.825 0.830 0.842 1 0.978 0.925 0.892 0.879 0.872 0.872 0.874 0.882 1 0.984 0.868 0.852 0.842 0.842 0.842 0.846 1 0.822 0.803 0.789 0.790 0.788 0.790 1 0.999 0.998 0.998 0.998 0.998 1 0.999 0.999 0.999 0.999 1 0.999 0.999 0.999 1 0.999 0.999 1 0.999 1 Table 5.18: Correlation Matrix Nominal T-notes 1988 - 2009 1 y 3 y 5 y 10 y 5.3 1 y 1 0.847 0.800 0.729 3 y 5 y 10 y 1 0.941 0.873 1 0.926 1 Simulation: Rebonato Since the simulation from the Principal Component Representation was mostly unplausible a test with the simulation conducted by Rebonato, Mahal, Joshi, Buchholz and Nyholm [17] was also implemented. I started out with their first approach; the naive. As described earlier it is an historical simulation method, where the future yields are simulated from the historical yield changes. The method is as follows: 1. Draw a random starting yield change curve from the historical yield curve changes. 2. Draw a random number between 5 and 50 for the window length of the subsimulation. 3. In each extraction of a yield curve change there is a 5 % probability of jumping out of the window and go back to step 1. 4. Continue with the subsimulations until the desired simulation length is obtained. 5.3.1 T-notes US Treasury yields with maturities of 3 months to 10 years between 2nd of January 1986 to 7th of May 2004 were studied to be able to relate to the study by Rebonato and Nyholm [18]. The difference from the other approach was that no averaging was done and percentage changes were used to keep the yields positive. One yield curve outcome is shown in Figure 5.26. The Principal Components before and after are shown in Figures 5.27 and 65 Table 5.19: Correlation Matrix Simulated Nominal T-notes 1988 - 2009 1y 3y 5y 10 y 1y 1 3 y 0.586 1 5 y 0.417 0.916 1 10 y 0.215 0.581 0.852 1 Figure 5.26: Simulated Nominal Surface T-yields 1962 - 2009 5.28. As can be seen the components are preserved. To clarify this the correlation matrices are shown in Table 5.20 and 5.21. As stated in Rebonato, Mahal, Joshi, Buchholz and Nyholm [17] this procedure asymptotically recovers the correlation matrix and the same holds for the eigenvalues and eigenvectors. The final result with this naive approach is not pleasing although that the correlation structure and the principal components are recovered. Therefore the author’s spring and mean reversion model was tried. It uses spring constants and mean-reversion levels to match the variance of the curvature of the simulation with the empirical variance of the curvature. It is not clear 66 Principal Component Analysis Treasury yields 1986−2004 0.8 1st PC 2nd PC 3rd PC 0.6 Change in yield 0.4 0.2 0 −0.2 −0.4 −0.6 0 1 2 3 4 5 6 7 8 9 10 Time to maturity Figure 5.27: Principal Components Treasury Yields 1982 - 2004 Principal Component Analysis Simulated yields 0.8 1st PC 2nd PC 3rd PC 0.6 Change in yield 0.4 0.2 0 −0.2 −0.4 −0.6 0 1 2 3 4 5 6 7 8 9 10 Time to maturity Figure 5.28: Principal Components Simulated Yields 67 Table 5.20: Correlation Matrix Nominal Treasury Yields 1982 - 2004 3 m 6 m 1 y 2 y 3 y 5 y 7 y 10 y 3 m 1 0.786 0.053 0.038 0.037 0.038 0.043 0.045 6 m 1 y 2 y 3 y 5 y 7 y 10 y 1 0.057 0.047 0.042 0.039 0.043 0.042 1 0.889 0.860 0.816 0.769 0.730 1 0.964 0.908 0.859 0.810 1 0.953 0.912 0.871 1 0.964 0.940 1 0.974 1 Table 5.21: Correlation Matrix Simulated Nominal Treasury Yields 3 m 6 m 1 y 2 y 3 y 5 y 7 y 10 y 3 m 1 0.790 0.054 0.041 0.040 0.043 0.049 0.051 6 m 1 y 2 y 3 y 5 y 7 y 10 y 1 0.070 0.057 0.047 0.044 0.050 0.047 1 0.888 0.858 0.811 0.766 0.721 1 0.962 0.899 0.853 0.802 1 0.951 0.909 0.866 1 0.961 0.936 1 0.973 1 from their paper on how to do this matching. In this study a non-linear least squares routine was tried on the data. Unfortunately, the results were still implausible, the final curvatures obtained translated distributions, either greater or less than zero. This yielded even more implausible results than the naive approach. Therefore this model was abandoned. 5.4 Simulation: Brute Force Model Since the two former models did not result in plausible future yield curves, a need for a new model arose. Another criteria was that the yield curves should be affected by the PURCS cycle. As stated in the article by Rebonato [17] - Brute-force simulation and analysis of the results is therefore often the only possible investigative route. Due to the failure of the former models, the focused shifted to brute force modeling. The idea was to add the cycle as a factor to the yield curve evolution. To keep the correlation structure smoothing splines were applied to the LIBOR and Swaps yields with a smoothing factor of 0.995, see Section 5.1.3. The relationship between the PURCS cycle and the smoothed yields was set up in the following way a (j) + log (y (i, j)) = log (1 + c (i) · d (j)) (5.7) where a (j) is a translation factor, y (i, j) is the yield at date i and maturity j, c (i) is the PURCS cycle at date i and d (j) is a scaling factor. 1 is added to the cycle since it has negative values. The reason that there is no scaling factor on the yields is that this is covered by the scaling on the PURCS cycle. If a scaling factor is introduced the optimization routine fails to find plausible solutions, it will just keep on rescaling. This relationship was studied for historical data for the LIBOR and Swaps 68 and the PURCS cycle. Expression (5.7) was re-expressed as F (i, j) = a (j) + log (y (i, j)) − log (1 + c (i) · d (j)) (5.8) To determine the functions a and d, for each maturity j a non-linear least squares optimization routine, lsqnonlin, in MATLAB was used. The results are shown in Figure 5.29 and 5.30 Optimized a 3.5 3.4 3.3 a 3.2 3.1 3 2.9 2.8 0 1 2 3 4 5 6 7 8 9 10 Term to Maturity (years) Figure 5.29: Least Squares optimized a As can be seen, a do not vary that much and is therefore set to 3. For d an exponential function (5.9) d (tj ) = m + pe−rtj was fitted with least squares to the function to describe the factor variation across maturities. The fitted function is shown in Figure 5.31 and its values in Table 5.22 and Goodness of Fit values in Table 5.23. Table 5.22: Estimated bounds parameter m p r parameters for function d with 95 % Confidence value 0.4152 5.835 0.2812 Lower Bound -0.211 5.245 0.2067 69 Upper Bound 1.041 6.426 0.3557 Optimized d 6 5 d 4 3 2 1 0 0 1 2 3 4 5 6 7 8 9 10 Term to Maturity (years) Figure 5.30: Least Squares optimized d d with fitted function d fit 6 5.5 5 d(TTM) 4.5 4 3.5 3 2.5 2 1.5 1 0 1 2 3 4 5 6 7 8 9 10 Term to Maturity (years) Figure 5.31: Least Squares fitted function to d 70 Table 5.23: Goodness of Fit for function d Measure Value Sum of Squared Errors 0.3051 R-Squared 0.9926 Adjusted R-Squared 0.9907 Having determined a and d, equation (5.8) could be rewritten as F (i, j) − a = log (y (i, j)) − log (1 + c (i) · d) eF (i,j)−a = elog(y(i,j))−log(1+c(i)·d) y (i, j) = (1 + c (i) · d) (5.10) (5.11) (5.12) in other words, eF (i,j)−a is the yields without PURCS cycle influence, and that is what should be simulated. From this it is clear that we only need the parameter d, since the value of a does not affect the expression. The last equation could now be re-expressed as z (i, j) = y (i, j) (1 + c (i) · d) (5.13) where z is referred to as the Cycle Neutral Yield at date i and maturity j. By using this ”logarithmic approach” the PURCS cycle is affecting the yields as a factor. The Cycle Neutral surface is shown in Figure 5.32. The next step was to apply Nelson-Siegel’s curve fitting technique to this surface. This was conducted by running an nonlinear least squares routine, lsqnonlin, in MATLAB. It was done with the same conditions as stated in Section 3.17. The time series of the four parameters are shown in Figure 5.33, and the error of the fit in Figure 5.34. The downward trend since 1988 to 2009 is evident in β0 . 71 Figure 5.32: Historical PURCS Cycle Neutral Yield Surface Time Series β0 Time Series β1 0.2 0 0.15 −0.02 0.1 −0.04 0.05 0 0 2000 4000 6000 8000 −0.06 0 Time Series β2 2000 4000 6000 8000 Time Series τ 0.05 2.002 2.001 0 2 1.999 −0.05 0 2000 4000 6000 8000 1.998 0 2000 4000 6000 8000 Figure 5.33: Nelson-Siegel estimated parameters from z 72 Nelson−Siegel SSE Error −4 1.2 x 10 1 SSE 0.8 0.6 0.4 0.2 0 0 1000 2000 3000 4000 5000 6000 7000 8000 Figure 5.34: Nelson-Siegel Sum of Squares fit error How to model these series? Since we are still in a world of low yields post the financial crisis, it is fairly reasonable to expect that the yields will rise with time. PK Air had also expressed a desire to have mean-reversion in the model and therefore Ornstein-Uhlenbeck processes are applied, see Section 3.16. By looking at the time series in Figure 5.33, it seemed as τ was a combination of β0 and β1 . If this is possible simulation time can be reduced. Therefore the following relationship was tried on historical data with a multiple linear regression. The obtained parameters are shown in Table 5.24 and the Goodness of Fit values in Table 5.25. τ = v + qβ0 + sβ1 (5.14) The following parameter estimates were obtained Table 5.24: Estimated Confidence bounds parameter v q s parameters for fitting β0 and β1 to τ with 95 % value 2.0021 -0.0365 -0.0367 Lower Bound 2.0021 -0.0370 -0.0374 73 Upper Bound 2.0022 -0.0360 -0.0360 Table 5.25: Goodness of Fit for regression Measure Value Sum of Squared Errors 8.7159e-004 R-Squared 0.7712 Adjusted R-Squared 0.7711 5.4.1 Calibration of Ornstein-Uhlenbeck processes To Calibrate the processes Maximum-Likelihood Estimation was applied, see Section 3.13 for a general description and Appendix B for a derivation of the estimates for the Ornstein-Uhlenbeck process. Since PK Air wants the simulation to be as fast as possible, the data set of estimated Nelson-Siegel parameters were chosen to be represented by 30 points instead, with equal distance in between themselves. Later when the simulation is conducted more frequent data sets can be obtained by linear interpolation. The Maximum Likelihood parameter estimation results are shown in Tables 5.26, 5.27, and 5.28. Table 5.26: Estimated Ornstein-Uhlenbeck parameters β0 parameter value µ 0.0526 λ 0.0994 σ 0.0053 Table 5.27: Estimated Ornstein-Uhlenbeck parameters β1 parameter value µ -0.0279 λ 0.7090 σ 0.0133 The condition in Nelson Siegel’s model that β0 + β1 > 0 had still to be governed. To cope with these negativity problems, a help function was introduced as soon as their sum was smaller than 0.01. If this study has been done 2 years ago, it would probably have been unnecessary with this ”pillow function”, but since the financial crisis and its aftermath is considered as an extreme event, these kind of helping functions is justified. If the sum the 74 Table 5.28: Estimated Ornstein-Uhlenbeck parameters β2 parameter value µ -0.0077 λ 1.7382 σ 0.0274 two parameters is defined as f (i) = β0 (i) + β1 (i) (5.15) Then the help function g (i) is ( f (i) ae a −1 , if f (i) < 0.01 g(i) = f (i) , if f (i) ≥ 0.01 a graphical illustration of the function is given in Figure 5.35. As can bee seen it is linear above the threshold of 0.01. The ”help” from the function is then portioned out in the following way Help function to avoid negative yields 0.06 0.05 g(f(i)) 0.04 0.03 0.02 0.01 0 −0.1 −0.08 −0.06 −0.04 −0.02 0 0.02 0.04 0.06 f(i) Figure 5.35: Help function with threshold 0.01 β0 (i) (β0 (i) + abs (β1 (i))) β1 (i) β1 (i) = (β0 (i) + abs (β1 (i))) β0 (i) = f (i) · ae a −1 − f (i) + β0 (i) (5.16) f (i) · ae a −1 − f (i) + β1 (i) (5.17) as can bee seen from the equations above the help is related to their respective proportions. After non-negativity is assured, the next step is to 75 generate the yields with the Nelson-Siegel function, see equation 3.44, the results are shown in Figures 5.36 and 5.37. Figure 5.36: Simulated PURCS Cycle Neutral Yield Surface The final step of the simulation is to add a simulated future PURCS Cycle influence, by rearranging equation (5.13) to y (i, j) = z (i, j) · (1 + c (i) · d) the final result is shown in Figures 5.38 and 5.39 76 (5.18) Simulated Nelson−Siegel Time Series 0.06 1m 2m 3m 6m 1y 2y 3y 4y 5y 7y 10y 0.05 yield 0.04 0.03 0.02 0.01 0 0 1000 2000 3000 4000 5000 6000 Simulated Days Figure 5.37: Simulated PURCS Cycle Neutral Yield Time Series Figure 5.38: Final Simulated Yield Surface 77 Final Simulated Time Series 0.07 1m 2m 3m 6m 1y 2y 3y 4y 5y 7y 10y 0.06 0.05 yield 0.04 0.03 0.02 0.01 0 0 1000 2000 3000 4000 5000 6000 Simulated Days Figure 5.39: Final Simulated Yield Time Series To produce statistics over the simulations, 10000 yield surfaces were generated and percentiles were calculated to visualize the results. The results for the 1 month, 1 year and 10 year term to maturities over the 30 time steps are shown in Figures 5.40, 5.41 and 5.42. The Cyclical behavior is as expected more evident in the shorter maturities. The same results are available for all term to maturities and can in turn be used for PK Air’s Swap Breakage Scenario calculation. 78 Percentiles 1 month yields 0.08 2.5 25 50 75 97.5 0.07 0.06 yield 0.05 0.04 0.03 0.02 0.01 0 0 5 10 15 20 25 30 Figure 5.40: Percentiles for 10 000 simulations of 1 month yield Percentiles 1 year yields 0.08 2.5 25 50 75 97.5 0.07 0.06 yield 0.05 0.04 0.03 0.02 0.01 0 0 5 10 15 20 25 30 Figure 5.41: Percentiles for 10 000 simulations of 1 year yield 79 Percentiles 10 year yields 0.08 2.5 25 50 75 97.5 0.07 yield 0.06 0.05 0.04 0.03 0.02 0 5 10 15 20 25 30 Figure 5.42: Percentiles for 10 000 simulations of 10 year yield 80 Chapter 6 Discussion 6.1 Conclusions With Principal Component Analysis, the results were analogue with precedent studies. The first three components are enough to describe the variation of the yield curve and the inclusion of shorter maturities decreases the explanatory power. Long horizon simulations of the yield curve from the Principal Component Representation did not yield plausible results. The term structure loses its shape and this historical simulation procedure exhibits the same trend as the historical data. This trend results in either negative yields or a ”zero carpet” depending on the valuation basis. The ”Rebonato approach” also did not yield plausible results. The problem with this approach was the implementation of the spring constants by matching the simulated variance of the curvature to the historical ones. It is not stated in their paper on how to do this, and efforts in this thesis failed. Because of the lack of long-term models in the Real World Measure the Brute Force Model was developed. The model manages to produce plausible future yield curve scenarios with cyclical influence, which was demanded by PK Air since it influences all other parts of their model. The model also follows the other criteria specified by PK Air: tractability, few parameters and fast to simulate. For instance, it is easy to edit the input parameters of the Ornstein-Uhlenbeck processes with PK Air’s view of the future. The help function to assure non-negativity is justified since the financial crisis is an extreme event - and must be treated thereafter. This author believes that this function is better than ordinary methods such as the absorption and reflection methods. 81 The model could be even better with some refinement and development. Maybe some people argue that the PURCS Cycle effect is to large and that it could be wise to investigate another way of linking the yields to the cycle than what has been done here. Some information is also lost by smoothing the yields, representing frequent data with fewer points, fitting functions and manipulating the data back and forth. This could all be added to a ”noise part” in the simulation that could be added in the end, but this needs further investigation. The Ornstein-Uhlenbeck process could probably be replaced by an even more sophisticated process, e.g. with varying volatility, mean-reversion level and speed of mean-reversion. The fitting techniques with least squares, regression and maximum-likelihood could also be inspected and developed. The same holds for the Nelson-Siegel model, there is an extension of the model - the Nelson-Siegel and Svensson model, which introduces another parameter and allows for a second ”hump” on the curve. But as always one is limited by time so this lies outside of the scope of this thesis. Correlation is often criticized of being a limited measure of dependence. It could be interesting to look at other dependence measures, e.g. cointegration and copulas. The data set used for the Brute Force Model is the US LIBOR and Swaps rates, it would be interesting to estimate the model from a longer set of data, e.g. the Treasury Yields, but they have the drawback of not having the shorter maturities for the longer period. Another topic for discussion is how valid the data from 1962 are? Has the financial markets changed since then in terms of globalization and size? This author believes that the US LIBOR and Swaps are more relevant, although the longer period Treasury Yields could serve as a reference. 6.2 Future Studies During the work with this thesis many ideas and questions arose, some of them have been stated in the previous section. Unfortunately these lie out of the scope of this thesis. Some of these are: • Investigate if the use of the Nelson-Siegel and Svensson model improves the model and also consider other curve fitting techniques. • Study the estimation techniques used and see if these could be replaced or refined by even better ones, e.g. Maximum Likelihood, Least 82 Squares and Regression. • Evaluate the link between the yield curve and the business cycle - are there other methods? • Develop or replace the Ornstein-Uhlenbeck processes. • Evaluate the forecasting ability of the model. • Estimate the model from other data sets, e.g. other lengths and countries. • Investigate other dependence measures than correlation, for instance cointegration and copulas. To conclude this model presents one approach for simulating yield curves under the influence of a business cycle. From the research by this author this is one of the first in this field. Therefore this author hopes it could serve as a foundation for further developments and refinements. 83 84 Appendix A PCA Results Here are the results of PCA on T-notes between 1962 and 2009 with 5 year windows. Principal Component Analysis T−notes 1962 − 1967 1 1st PC 2nd PC 3rd PC 0.8 Change in yield 0.6 0.4 0.2 0 −0.2 −0.4 −0.6 −0.8 1 2 3 4 5 6 7 8 9 Time to maturity Figure A.1: Principal Components T-notes 1962-1967 85 10 Principal Component Analysis T−notes 1967 − 1972 1 1st PC 2nd PC 3rd PC 0.8 Change in yield 0.6 0.4 0.2 0 −0.2 −0.4 −0.6 1 2 3 4 5 6 7 8 9 10 Time to maturity Figure A.2: Principal Components T-notes 1967-1972 Principal Component Analysis T−notes 1972 − 1977 0.8 1st PC 2nd PC 3rd PC 0.6 Change in yield 0.4 0.2 0 −0.2 −0.4 −0.6 −0.8 1 2 3 4 5 6 7 8 9 Time to maturity Figure A.3: Principal Components T-notes 1972-1977 86 10 Principal Component Analysis T−notes 1977 − 1982 1 1st PC 2nd PC 3rd PC 0.8 Change in yield 0.6 0.4 0.2 0 −0.2 −0.4 −0.6 −0.8 1 2 3 4 5 6 7 8 9 Time to maturity Figure A.4: Principal Components T-notes 1977-1982 87 10 Principal Component Analysis T−notes 1982 − 1987 0.8 1st PC 2nd PC 3rd PC 0.6 Change in yield 0.4 0.2 0 −0.2 −0.4 −0.6 1 2 3 4 5 6 7 8 9 Time to maturity Figure A.5: Principal Components T-notes 1982-1987 88 10 Principal Component Analysis T−notes 1987 − 1992 0.8 1st PC 2nd PC 3rd PC 0.6 Change in yield 0.4 0.2 0 −0.2 −0.4 −0.6 1 2 3 4 5 6 7 8 9 Time to maturity Figure A.6: Principal Components T-notes 1987-1992 89 10 Principal Component Analysis T−notes 1992 − 1997 0.8 1st PC 2nd PC 3rd PC 0.6 Change in yield 0.4 0.2 0 −0.2 −0.4 −0.6 1 2 3 4 5 6 7 8 9 Time to maturity Figure A.7: Principal Components T-notes 1992-1997 90 10 Principal Component Analysis T−notes 1997 − 2002 1 1st PC 2nd PC 3rd PC 0.8 Change in yield 0.6 0.4 0.2 0 −0.2 −0.4 −0.6 1 2 3 4 5 6 7 8 9 Time to maturity Figure A.8: Principal Components T-notes 1997-2002 91 10 Principal Component Analysis T−notes 2002 − 2007 1 1st PC 2nd PC 3rd PC 0.8 Change in yield 0.6 0.4 0.2 0 −0.2 −0.4 −0.6 −0.8 1 2 3 4 5 6 7 8 9 Time to maturity Figure A.9: Principal Components T-notes 2002-2007 92 10 Principal Component Analysis T−Notes 2007−2009 1 1st PC 2nd PC 3rd PC 0.8 Change in yield 0.6 0.4 0.2 0 −0.2 −0.4 −0.6 −0.8 1 2 3 4 5 6 7 8 9 Time to maturity Figure A.10: Principal Components T-notes 2007-2009 93 10 94 Appendix B Maximum Likelihood Estimation Ornstein-Uhlenbeck From the solution from Section 3.16 Xt = θ + e −κ(t−s) Z t (Xs − θ) + σ e−κ(t−u) dWu s the conditional mean and variance can be derived E [Xt | Xs ] = θ + e−κδ (Xs − θ) where δ has been introduced for the time step (t − s). Z t −κ(t−u) e dWu V ar [Xt | Xs ] = E σ s = (by Itô Isometry) Z t 2 −2κ(t−u) =E σ e du s σ2 = 1 − e−2κδ 2κ Therefore Xt is normally distributed with E [Xt | Xs ] = θ + e−κδ (Xs − θ) 2 and V ar [Xt | Xs ] = σ2κ 1 − e−2κδ . From this the conditional probability density function for Xi+1 given Xi with time step δ as " 2 # Xi − Xi−1 e−κδ − θ 1 − e−κδ 1 f (Xi+1 | Xi ; κ, θ, σ) = √ exp − 2σ̂ 2 2πσ̂ 2 95 where σ̂ 2 = σ 2 1 − e−2κδ 2κ The log-likelihood function for a sample X0 , X1 , ..., Xn is derived from the conditional density function [22] ` (κ, θ, σ̂) = n X logf (Xi | Xi−1 ; κ, θ, σ̂) i=1 n i2 1 Xh n Xi − Xi−1 e−κδ − θ 1 − e−κδ = − log (2π) − n · log (σ̂) − 2 2 2σ̂ i=1 To find the maximum of the log-likelihood surface the partial derivatives is set to zero and solved for n i d` (κ, θ, σ̂) 1 Xh Xi − Xi−1 e−κδ − θ 1 − e−κδ =0 = 2 dθ 2σ̂ i=1 Pn −κδ i=1 Xi − Xi−1 e ⇒θ= n (1 − e−κδ ) n i d` (κ, θ, σ̂) δe−κδ X h 2 −κδ =0 = − (X − θ) (X − θ) − e (X − θ) i i−1 i−1 dκ 2σ̂ 2 i=1 Pn (Xi − θ) (Xi−1 − θ) 1 ⇒ κ = − log i=1 Pn 2 δ i=1 (Xi−1 − θ) n i2 d` (κ, θ, σ̂) n 1 Xh =0 = − 3 Xi − Xi−1 e−κδ − θ 1 − e−κδ dσ̂ σ̂ σ̂ i=1 n i2 1 Xh 2 ⇒ σ̂ = Xi − Xi−1 e−κδ − θ 1 − e−κδ n i=1 The conditions depend on each other. Although θ and κ are independent of σ̂. Therefore θ can be found be substituting κ into the expression for θ. 96 To faciliate some notations are introduced Sx = n X Si−1 i=1 Sy = Sxx = n X i=1 n X Si 2 Si−1 i=1 Sxy = Syy = n X i=1 n X Si−1 Si Si2 i=1 results in Sy − e−κδ SX n (1 − e−κδ ) θ= 1 Sxy − θSx − θSy + nθ2 κ = − ln δ Sxx − 2θSx + nθ2 substituting κ into θ yields Sy − nθ = 1− Sxy −θSx −θSy +nθ2 Sxx −2θSx +nθ2 Sxy −θSx −θSy +nθ Sxx −2θSx +nθ2 SX 2 remove denominators Sy Sxx − 2θSx + nθ2 − Sxy − θSx − θSy + nθ2 Sx nθ = (Sxx − 2θSx + nθ2 ) − (Sxy − θSx − θSy + nθ2 ) collecting terms (Sy Sxx − Sx Sxy ) + θ Sx2 − Sx Sy + θ2 n (Sy − Sx ) nθ = (Sxx − Sxy ) + θ (Sy − Sx ) and this yields nθ (Sxx − Sxy ) − θ Sx2 − Sx Sy = (Sy Sxx − Sx Sxy ) and the final expressions are obtained as θ= (Sy Sxx − Sx Sxy ) n (Sxx − Sxy ) − (Sx2 − Sx Sy ) 97 1 Sxy − θSx − θSy + nθ2 κ = − ln δ Sxx − 2θSx + nθ2 σ̂ 2 = i 1h Syy − 2αSxy + α2 Sxx − 2θ (1 − α) (Sy − αSx ) + nθ2 (1 − α)2 n with σ 2 = σ̂ 2 2λ 1 − α2 and α = e−κδ 98 Bibliography [1] C. Alexander. Market Models - A Guide to Financial Data Analysis. John Wiley & Sons, Ltd, The Atrium, Southern Gate, Chichester, West Sussex PO19 8SQ, England, 2001. [2] British Bankers Association. Libor rate. http://www.bbalibor.com/ bba/jsp/polopoly.jsp?d=1625. [3] C. Bernadell, J. Coche, and K. Nyholm. Yield curve prediction for the strategic investor. Technical report, Risk Management, European Central Bank, 2005. [4] G. Blom, J. Enger, G. Englund, J. Grandell, and L. Holst. Sannolikhetsteori och statistikteori med tillämpningar. Studentlitteratur, Studentlitteratur, Lund, 2005. [5] D. Bolder and D. Stréliski. Yield curve modelling at the bank of canada. Technical report, Bank of Canada, 1999. [6] G. R. Duffee. Idiosyncratic variation of treasury bill yields. The Journal of Finance, 2:527–550, 1996. [7] R. Fiori and S. Iannotti. Scenario based principal component valueat-risk: An application ot italian banks’ interest rate risk exposure. Technical report, Banca d’Italia, 2006. [8] Financial Trend Forecaster. Us historical inflation http://inflationdata.com/Inflation/Inflation_Rate/ HistoricalInflation.aspx. data. [9] O. Hinton. Describing random sequences. http://www.staff.ncl.ac. uk/oliver.hinton/eee305/Chapter6.pdf. [10] J. Hull. Options, Futures and other Derivative Securities, 2nd Ed. Prentice Hall, Englewood Cliffs, New Jersey 07632, 1993. [11] R. Litterman and J. Scheinkman. Common factors affecting bond returns. The Journal of Fixed Income, pages 54–61, June 1991. 99 [12] D. G. Luenberger. Investment Science. Oxford University Press, Inc., 198 Madison Avenue, New York, NY 10016, USA, 1998. [13] Mathworks. Curve fitting toolbox user’s guide. Technical report, Mathworks, 2008. [14] C. R. Nelson and A.F. Siegel. Parsimonious modeling of the yield curves. Journal of Business, 60:473–489, 1987. [15] Federal Reserve Bank of St. Louis. Economic research. research.stlouisfed.org/fred2/categories/115. http:// [16] W. Phoa. Yield curve risk factors: Domestic and global contexts. Technical report, The Capital Group of Companies. [17] R. Rebonato, S. Mahal, M. Joshi, L-D. Buchholz, and K. Nyholm. Evolving yield curves in the real-world measures: a semi-parametric approach. The Journal of Risk, 7:29–61, 2005. [18] R. Rebonato and K. Nyholm. Long horizon yield curve forecasts: Comparison off semi-parametric and parametric approaches. Applied Financial Economics, 18:1597–1611, 2008. [19] A. Santomero and D. Babbel. Financial Markets, Instruments and Institutions, Irwin McGraw-Hill. Irwin McGraw-Hill, Boston, Massachusetts, 1997. [20] K. P. Scherer and M. Avellaneda. All for one... one for all?: A principal component analysis of latin american brady bond debt from 1994 to 2000. Technical report, Courant Institute of Mathematical Sciences, New York University, United States, 2000. [21] G. M. Soto. Using principal component analysis to explain term structure movements: Performance and stability. Progress in Economics Research, 8:203–225, 2004. [22] M.A. van den Berg. Ornstein-uhlenbeck process. http://www.sitmo. com/doc/Ornstein-Uhlenbeck_process. 100