Tinge exchange rate model using a new neural network training

advertisement
Tenge exchange rate model using new neural network training
method
Author1
Author affiliation
Author Email
Author2
Author affiliation
Author Email
Abstract: The paper reports results derived from new robust neural network (RNN) based
KZT/USD exchange rate model against AR (2) model. The characteristic of the RNN method
is explained followed by comparative study of quarterly KZT/USD models. Diagnostic
measure such as mean absolute deviations (MAD), R2 and tracking signals (TS) suggests
superiority of the RNN KZT/UDS model. The RNN training methods computes adoptive
directional search vectors such that the error function is gradually optimized. The method
automatically computes adjustable dynamic learning rates to direct search toward the
minimum valley of the error function. The algorithm identifies dynamic adjustable learning
rates to generate convergent sequence of the error function. As a result, the training is robust
to model exchange rate. The RNN model performs better in MAD, TS and R2 value.
Key-Words- KZT/USD Exchange rate, Robust Neural Networks, Adjustable training,
Forecast.
1. Introduction
Due to the exchange rate crisis in the CIS region during the fall of USSR era, many small
open economies have adopted flexible exchange rates in combination with some kind of
monetary or interest rate mechanism. Volatility of exchange rate in Kazakhstan is significant
issue taking account possibility of huge transaction loss induced by volatility exchange rate.
Kazakhstan economy highly depends on international trade and FDI (Foreign Direct
Investment), exchange rate volatility in Kazakhstan have influence in international trade and
FDI. Kazakhstan has consistently achieved moderate growth rate induced by oil export. But
international oil price is declining and under such market condition exchange rate could be
influenced by several possible macroeconomic factors. USD (dollar) exchange rate trends had
grown up to 2003 but dollar’s exchange rate to Tenge has fallen since 2003 (see Table 1).
There have been number of attempts to apply neural network (NN) to the task of financial
modeling (Cao et al. 2005, Jasic and Wood 2004, Kaastra and Boyd 1995, Lam 2004, Nygren
2004). When it comes to performing a predictive analysis, it is very difficult to build one
general model that will fit every market. Such models tend to be specific to markets and asset
classes and a general model may not be applicable across markets. Similarly, there may be
some temporal changes as well which mean that the models may need to be modified over
time in order to preserve their effectiveness (Zhang, et al, 1998). We develop KZT/USD
exchange rate forecast model using new RNN training method and brief results are reported
to show the superior performance of the model, against AR (2) model.
2 Rationales for Self-Adaptive Training Method
Consider a NN error function with m training weights and the learning rates are identified
m
automatically. For computational convenience, the higher dimensional error function in E
dimension is decomposed into several error functions in lower dimension. Such transformed
error function retains the true convex characteristics of the original error function. The
gradient information is evaluated and the training method updates all the network weight
parameters say wj, j=1,2,…..m, by a factor such that improvement in training is noticed. Each
epoch identifies m different learning rates along the training directions. Once the NN training
weights are updated, the error function is evaluated to notice the improvement in error
function and the rate of convergence. Various form of standard back propagation training
method and its variant are not self-adaptive. They are heuristic training method (Ahmed et al
2000a; Haykin, 1999; Weir, 1991; and Kamarthi et al., 1999). The training direction, d k of
the error function f (w) is computed using the gradient, ▽f (wk) information from a single
training pattern in standard on-line back propagation (Bishop, 1995, Ahmed, et al, 2000b).
The fixed value of a learning rate,  k does not always lead to a maximum local decrease in
function value. The learning rate depends on the shape of the error function (Jacobs, 1988) as
it trains from the current epoch k to the next epoch k+1, k+2,… and so on. The constant
value of learning rate, η k , can direct the search away from the local minimum during epoch k
and the directions dk generated from gradient ▽f (wk) are different for a single weight
component in standard back propagation training. Consider a training method, we call RNN,
where an iterative algorithm is applied to an error function with an initial arbitrary weight
vector wk, at the beginning of iteration k, the algorithm generates a sequence of vectors w k+1,
w k+2,.… during epoch k+1, k+2,….,.., and henceforth. The iterative algorithm is globally
convergent if the sequence of vectors converges to a solution set Ω. Consider the problem:
minimize f (w), subject to: w Є Em. Let, Ω Є Em be the solution set, and let, the application of
an algorithmic map, ℬ, generates the sequence w k+1, w k+2,.…, starting with weight vector wk
such that (w k, w k+1 , w k+2,.…) Є Ω, then the algorithm converges globally and the algorithmic
map is closed over Ω. The following properties are utilized to train the new RNN.
Property 1: Suppose that f: E m → E1 and the gradient of the error function, ▽f (w), is
defined then there is a directional vector d such that ▽f (w)T d < 0, and f(w +ηd) < f(w) : (η
Є (0,  ),  > 0), then the vector d is a descent direction of f (w), where  is assumed arbitrary
positive scalar.
Property 2: Let f: E m → E1 is a descent function. Consider any training weight w Є Em and
d Є Em : d ≠0. Then the directional derivative ▽f(w, d) of the error function f(w) in minimum
direction d always exists. The expression to update the RNN network weight, wk+1 is given by
w k+1 =w k + ηk dk. Here, η k is defined as a minimization problem of the type: η k = {min f(w +
ηd)}; subject to : ηk Є L, where, L = (η : η Є E1).
3. Results
The following table shows R2 as well the MAD value for the RNN model is better than the
AR (2) model. This confirms the superiority of the KZT/USD exchange model using the new
RNN method. The TS vale in both case are within acceptable limits, however, the plot shows
that the TS vale in RNN is asymmetrical on the zero level, hence the overall predictability is
good with RNN model.
Mode
l
RNN
R2
0.9
2
MA
D
3.11
TS
0.05
1
TS Plot
Fitted & Actual Data
TS= ∑E/MAD
6.00
100.00
7 9 11 13 15 17 19 21 23 25 27 29 31 33 35
50.00
-4.00
0.00
-6.00
1
-8.00
0.8
7
5.60
0.00
2
F
150.00
2.00
0.00
-2.00 1 3 5
AR(2
)
A
200.00
4.00
TS= ∑E/MAD
2.00
1.00
0.00
-1.00
1
3
5
7
9
11 13 15 17 19 21 23 25 27 29 31 33
-2.00
-3.00
-4.00
-5.00
-6.00
-7.00
-8.00
4
7
10 13 16 19 22 25 28 31 34
A
180
160
140
120
100
80
60
40
20
0
1
3
5
7
F
9 11 13 15 17 19 21 23 25 27 29 31 33
Table 1: Performance of RNN and AR (2) model (A=Actual, F=Fitted Quarterly KZT/USD Exchange
rate1999 till 2007)
4. Conclusions
This research examined and analyzed the use of a newly developed neural network
model (RNN) in foreign exchange forecasting with KZT against USD. The RNN
model gives the evidence that there is possibility of extracting information to forecast
exchange rate reliably. The evaluation of the model is based on the estimate of mean
absolute error, R2 and tracking signal.
5. References
1] Xxxx, uyyyy (year). Neural Network without Learning Parameter”, xyz 2nd International
conference, 2000, xyz Place.
2] vvvvv, cccccc. (Year). “Multi-directional training algorithm for Feed-Forward Neural Network”,
Journal xyz, 2000.
3] Bishop, Christopher M. (1995). “Neural Networks for Pattern Recognition”. Oxford, UK:
Clarendon.
4] Cao, Q., K.B. Leggio and M.J. Schniederjans (2005), ‘A Comparison Between Fama and French’s
Model and Artificial Neural Networks in Predicting the Chinese Stock Market’, Computers and
Operations Research, 32: 2499–2512.
5] Haykin, Simon, (1999). “Neural Networks: A Comprehensive Foundation”, Prentice Hall, NJ.
6] Jacobs, R.A. (1988). “Increased Rate of Convergence Through Learning Rate Adaptation”. Neural
Networks, 1, 295-307.
7] Jasic, T. and D. Wood (2004), ‘The Profitability of Daily Stock Market Indices Trades Based on
Neural Network Predictions: Case Study for the S&P 500, the DAX, the TOPIX and the FTSE in
the Period 1965–1999’, Applied Financial Economics, 14(4): 285–297.
8] Kaastra, L. and M. Boyd (1995), ‘Designing a Neural Network for Forecasting Financial and
Economic Time Series’, Neurocomputing, 10(3): 215–236.
9] Kamarthi, S.V., and Pittner, S., (1999). “Accelerating Neural Network Training Using Weight
Extrapolations”, Neural Networks, 12, 1285-1299.
10] Lam, M. (2004), ‘Neural Network Techniques for Financial Performance Prediction: Integrating
Fundamental and Technical Analysis’, Decision Support Systems, 37(4): 567–581.
11] Nygren, K. (2004), Stock Prediction—A Neural Network Approach, Master’s Thesis, Royal
Institute of Technology, KTH, Sweden.
12] Weir, M.K., (1991). “A Method for Self-Determination of Adaptive Learning Rates in Back
Propagation”. Neural Networks, 4, 371-379.
13] Zhang, G. and M.Y. Hu (1998), ‘Neural Network Forecasting of the British pound/US dollar
Exchange Rate’, International Journal of Management Science, 26(4): 495–506.
Download