2011 6th International Conference on Industrial and Information Systems, ICIIS 2011, Aug. 16-19, 2011, Sri Lanka
Financial Forecasting Based on Artificial Neural Networks: Promising Directions for Modeling W. D. S. Roshan, R. A. R. C. Gopura, Member, IEEE, A. G. B. P. Jayasekara, Member, IEEE. Abstract- Financial forecasting plays a critical role in present economic context where neural networks have become a good alternative technique over traditional methods. Vast ranges of neural models are developed to achieve better accuracy in forecasting. In addition, the ways to find out a good neural architecture is being explored by the research community. In the literature, main problems are figured out within the area of data preparing and neural network design. In this paper, the reasons that affect the performance of the models are discussed based on empirical and mathematical evidence. Finally, this paper presents the directions towards a more suitable neural model for financial forecasting by combining data preprocessing techniques, clustering techniques and support vector machine. Index Terms— Back propagation, Bias variance dilemma, Cover’s theorem, Self-organizing maps, Structural risk minimization, Support vector machine, Wavelet transform.
I.
INTRODUCTION
Forecasting exchange rates plays a major role in day-today financial markets, which becomes a difficult task due to high volatility and complexity. Even though it is easy to trade in financial markets, it is very difficult to make a profit as a result of highly unpredictability of exchange rates. During the past few years ample research has been carried out on understanding inter-relations within financial market data. Recently, there is an increasing trend on adopting artificial neural networks (ANN) and its variants, in order to explore non-linearites of the financial data [1]. Even with vast amounts of available financial data, identification of underlying patterns has become difficult due to economic, political, environmental, and even psychological factors that affect the fluctuation of exchange rates. Under these circumstances data-driven self-adaptive feature of ANN [2] and universal functional approximation capability are highly desirable in financial forecasting. Even though ANN has achieved good performance than the conventional forecasting methods there are certain limitations in designing ANN models [3], [4]. The common limitations that are highlighted in this work are within the area of data preparing and neural network design. In order to overcome those limitations, different existing data preprocessing techniques and neural network architectures are compared to highlight the strengths and weaknesses of them. Then, the directions towards a more suitable model for W. D. S. Roshan is with the Department of Mechanical Engineering, University of Moratuwa, Katubedda, Sri Lanka. (e-mail:
[email protected]). R. A. R. C. Gopura is with the Department of Mechanical Engineering, University of Moratuwa, Katubedda, Sri Lanka (corresponding author:+9411-2640472; fax: +94-11-2650622; e-mail:
[email protected]). A. G. B. P. Jayasekara is with the Department of Electrical Engineering, University of Moratuwa, Katubedda, Sri Lanka (e-mail:
[email protected]).
forecasting are explored. Section II and section III are based on empirical evidences. Input selection and input data pre-processing are discussed in section II under data preparing. Learning algorithm selection, neural architecture and performance measure selection are discussed under neural network design in section III. Mathematical background and empirical evidence for selecting particular choices for data preparing and network design are discussed in section IV. Section V concludes the paper. II.
DATA PREPARING
A. Input selection The selection of inputs for neural networks directly affect to its performance. In the field of financial market forecasting, most of the researchers have used time delay inputs and various technical indicators (moving average, stochastic oscillator etc.) as inputs for the neural network [4]. There are large numbers of technical indicators and they got their own strengths over others. However, use of technical indicators as much as possible does not necessarily lead to better forecasting results; rather it will weaken the forecast due to curse of dimensionality [5]. Hidden relations among inputs can lead to the bias effect of neural network. Therefore, inputs should be selected carefully. Selected inputs should not be correlated with other inputs and should have maximum correlation with itself to achieve better forecast. There is no general method to select input parameters of a neural network. In [6], four technical indicators have been used. In [7] five technical indicators have been used while [8] uses 53 financial indices. Furthermore [9] used 75 technical indicators. However, most researchers focused on finding optimal input variables among the available inputs. Rescaled range analysis [10], [11], [12] and Lyapunov exponent [13] are used as a measure of time series predictability and forecastable time span is selected using those analyses. Trial and error, stepwise-regression [14], auto-regression testing [15], [16], [17] genetic algorithm based selection techniques [18], [19] are used to find most influential input variables by removing redundant ones. About 45% articles that are reviewed in this study have used technical indicators while 35% used time lagged inputs. Details of the selected inputs are presented in Table I. B. Input data preprocessing Data preprocessing is essential for better neural network performance especially for financial time series forecasting. Financial time series has highly noisy, nonlinear and nonstationary characteristics which reduces the overall
978-1-61284-0035-4/11/$26.00 ©2011 IEEE 322
2011 6th International Conference on Industrial and Information Systems, ICIIS 2011, Aug. 16-19, 2011, Sri Lanka
TABLE I INPUT PARAMETERS USED AS INPUTS Indicator
Article
Input
[6]
Four inputs(TAIEX index futures prices and technical indicators) 53 financial indices 75 technical indicators Time delay inputs, indicators Reconstructed phase space variables, normalized daily closing price, 23 technical indicators 5 financial indices Technical indicators Daily opening, closing, highest and lowest price and daily trade volume Current fundamental asset price, strike price and time-to-maturity 27 technical indicators 4 to10 technical indicators
[8] [9] [10] [13] [19] [20] [21] [22] [23] [24] [7] [11] [25] [26] [27] [28] [29] [18] [30] [31] [14] [32] [33] [34] [15] [16] [17] [35] [36] [37] [38] [39] [40] [41] [42] [43] [12]
these models forecasts more accurately than that of back propagation (BP) according to [3], multilayer feed forward network with BP is most widely used network design and also lead to satisfactory results. However, BP algorithm has some undesirable disadvantages such as slow convergence speed, higher likeliness to get trapped in local minima, initial weight determination, and selecting transfer functions. Recurrent neural network (RNN), a variant of feed forward ANN, has effectively used in some models [13], [34]. Recently many researchers have shown that SVM is more suitable and has performed better than other algorithms in [19], [20], [29], [30], [42].
• • • 9 ~ • • 9 • •
10 to 20 technical indicators • Time lagged price values 9
•
Technical Indicators
9
Time lagged inputs
Fundamental indicators
~
Other
network performance. Almost all recently published articles have used advanced data preprocessing techniques such as self-organizing maps (SOM) [27], [28], [33], [35], [41] clustering techniques [9], [14] genetic algorithm (GA) based systems [18], [19] principal component analysis (PCA) [43] independent component analysis (ICA) [6] and wavelet transform (WT) [34], [41]. Among those techniques SOM, PCA, ICA, WTs and combinations of these techniques have given the promising results. When selecting a specific method from above promising approaches, PCA has drawback of missing small amount of time series characteristics and better to use with nonlinear extension to deal with nonlinear characteristics of time series. When selecting ICA, nonstationary effect can affect the performance of separation of independent sources. On the other hand, WT is superior alternative for process nonstationary and noisy time series. Its multilevel decomposition and orthogonal data projection capability have been used to remove the inter relations of inputs, noise and nonstationary characteristics of the financial time series. III.
NEURAL NETWORK DESIGN
A. Learning algorithm Learning algorithm is another key factor for neural network performance. When selecting learning algorithm, although [44] has shown that scaled conjugate gradient and Bayesian regularization based models show competitive results and
B. Architecture selection Generally, neural networks are fed with past data and get the future values as outputs. The network should be able to positively identify relationships within inputs. If the network architecture become more complex it will be difficult to recognize patterns and often lead to over fit. Too many hidden neurons and higher number of training epochs can also be caused neural network to over fit. On the other hand, if the network is designed with more simple architecture it will not be able to separate correct patterns from input data and often lead to under fitting. Recently more and more researchers are realizing that using a single neural network may lead to bias effect and combining multiple neural architectures can consistently outperform the single models [3]. There is an improving attention going on for applying hybrid models and ensemble models for better performance [9], [12], [20], [40], [43]. C. Performance measure The network performance is assessed by the performance measures. Mean squared error (MSE) [17], [22], normalized mean squared error (NMSE) [10], [27], root mean square error (RMSE) [24], [34], mean absolute error (MAE), directional symmetry (DS) [6], [14], and mean absolute percentage error (MAPE) [14], [23] are widely used indicators as performance measures. In addition, [35] used profit as the performance measure. Rather than using just one performance measure, combination of performance measures like error, directional and profit [21] gives better understanding of the performance of the model. IV.
MATHEMATICAL BACKGROUND AND EMPIRICAL EVIDENCE
A. Cover’s theorem Determining number of hidden neurons is critical for good accuracy. According to Cover’s theorem [45] “A complex pattern-classification problem cast in a high-dimensional space nonlinearly is more likely to be linearly separable than in low dimensional space”. If the input space is transformed into higher dimensional space according to Cover’s theorem, it is easy to linearly separate those patterns and advisable to choose higher number of hidden neurons than number of input nodes. Higher number of hidden nodes than the input
323
2011 6th International Conference on Industrial and Information Systems, ICIIS 2011, Aug. 16-19, 2011, Sri Lanka
nodes has been found in [15], [16], [17], [21], [22], [25] while [26] claimed best neural model as 9-3-1 architecture. Radial basis functions (RBF) [31] and SVM have more principal basis for construction and often create their architecture according to Cover’s theorem. B. Structural risk minimization Generalization error should be kept in minimum possible value for best performance. Therefore, reduction of upper bound of the generalization error is essential for every forecasting model. Vapnik–Chervonenkis (VC) dimension, first proposed by [46], is the representation of maximum number of examples that can be learned by a learning network without an error. VC dimension plays major role in finding upper bound of generalization error. They have defined the guaranteed risk (upper bound of the , , is as in (2), where generalization error) confidence interval and it depends on training sample size N, VC-dimension , and probability . , ,
1 2
, ,
(1)
1
1
, ,
(2)
Here, training error is denoted by and represents weight vector for the fixed training set. There is a value of h, which reduces the upper bound of generalization error to its minimum possible value. According to [47], VC dimension, h of purely linear network is proportional to number of free parameters , purely threshold neural network is proportional to , while Sigmoidal activation function based feed forward neural architectures has VC dimension proportional to W . Going through above-mentioned proofs, upper bound of generalization error can be varied by varying number of hidden neurons. Due to the fact most articles related to feed forward neural networks have used trial and error methods to find the architecture with best generalizing capability. Results have shown significant improvement of the generalization capability [26], [40]. Unlike feed forward neural networks and RBF networks, SVM has interesting property of structural risk minimization. SVM’s VC dimension is bounded from above as shown in (3). can be minimized by properly selecting the optimal margin of separation and it is independent of the dimensionality of the and thus immune to the curse of input space dimensionality. Let denote the diameter of smallest ball containing all the input vectors can be written as, ,
1
(3)
Empirical evidences [19], [29], [30], [38], [42] along with the structural risk minimization property have shown that SVM achieved good generalization performance than feed forward ANN and RBF networks in their benchmark performance comparisons. Significant comparison was not performed
within SVM and recurrent models. However [9] claimed that ensemble recurrent model outperform single SVM. Mathematical evidence for ensemble performance is discussed in next section. C. Neural network generalization and Bias variance dilemma Generalization of neural network is still a challenging problem. For a good generalization, the model should give low training error as well as low testing error. Cross validation is the most widely used generalization method for feed forward neural networks. Input data set is divided into three sets called training, validation and testing. Training set is used to train the neural network. Validation set contains the data that has never seen by the neural network. Validation set is used in training along with training set to calculate out of sample error. If the output of the sample error increases for consecutive iterations, it will be the stopping criteria for network training. However, there is a risk of early stopping the training. Even using optimized VC dimension to reduce the upper bound of generalization error as presented in section IV-B, once it reaches its minimum point, generalization error cannot be reduced further more for a single network. This is due to well-known bias variance dilemma. Average value of the estimation error between the | regression function and the approximating function , evaluated over entire training set can be written as follows. , And
,
(4)
are defined by , ,
|
(5) ,
(6)
Here, bias is the representation of inability to approximate the regression function by given network and variance is the representation of lack of information contained in the training sample. A single neural network achieves low bias result for higher variance and vice versa, unless training set is infinitely large. However, infinitely large training set cause slow convergence due to its size [48]. In a neural architecture design, bias/variance dilemma should be considered seriously. Therefore, both bias and variance should be kept in small values in order to achieve good generalization. According to principal of divide and concur, dividing a computationally complex task into computationally simple tasks and combining its output to take the solution is a better method for solving such task [49] .This method allows reduction in the variance by combining multiple over-trained models, which are trained to zero bias. It will reduce both bias and variance at once for a finite training set, thus remove the bias variance dilemma. Empirical evidence for reducing the bias variance dilemma be found in [14], [27], [28], [33], [35], [41]. Summary of all reviewed articles are presented in Table II.
324
2011 6th International Conference on Industrial and Information Systems, ICIIS 2011, Aug. 16-19, 2011, Sri Lanka
TABLE II SUMMARY OF MODELING TECHNIQUES Article [6]
Data preprocessing Normalize, ICA
Network type SVM
Training method Support vector regression(SVR)
[7]
Normalized
Dynamic Ridge Polynomial ANN
Constructive learning algorithm
[8] [9]
Grey Relational Analysis Euclidean distance based partitioning
BP Genetic algorithm (GA)
[10]
Rescaled range analysis
FFANN Recurrent ensemble model FFANN
BP
[11]
FFANN
BP
Ensemble
vary
Probabilistic neural network TDNN,RNN RBF
[15]
Rescaled range analysis+ Normalizing+ sensitive analysis Normalizing+ Hurst exponent Phase Diagrams, Correlation Dimension, Lyapunov exponent Step wise regression analysis +Normalizing+ Dynamic learning algorithm based clustering Auto-regression testing
FFANN
Temporal BP, Extended Kalman filter, Particle-swarm optimization (PSO)+ adaptive recursive least-squares Adaptive BP
[16]
Auto-regression testing
FFANN
[17]
Auto-regression testing
FFANN
[18] [19]
GA based instant selection GA
FFANN SVM
[20]
-
[21]
-
Stochastic Volatility (SV) model with jumps +SVM ANN
[22]
Normalize
FFANN
[23] [24]
Grey-GJR–GARCH Particle swarm optimization + normalize -
FFANN FFANN
[12] [13]
[14]
[26]
[27] [28] [29] [30] [31]
Normalize, Self-organizing feature maps Kohonen self-organizing map Z-score normalization method Normalize Extract indicators with better performance
FFANN
Benchmark models Random walk (RW), SVR
Multilayer perceptron (MLP), Functional Link Neural Network (FLNN), Pi-Sigma Neural Network (PSNN), Ridge Polynomial Neural Network (RPNN) ANN, ANN / Grey model (1, N ) Genetic programming prediction, SVM
Performance measures RMSE, NMSE MAD, DS ,correct down trend (CD) , correct up trend (CP) NMSE, Annualized Return
RMSE Frequency of correct prediction
Auto regression integrated moving average(ARIMA) ARIMA 3-Benchmark models for profit calculation
NMSE, Gradient
FFANN, K-nearest neighbor, Decision tree, Native Bayesian classifier Fisher linear classifier
Average error rate
Type 2 fuzzy time series model, Fuzzy time series model, Fuzzy dual-factor timeseries
MAD,MAPE,DS,CP, CD
Standard BP, LM-based learning, Extended Kalman filter (EKF), BP with optimal learning rates Batch learning, EKF-based learning, Levenberg-Marquardt(LM)based learning, Standard BPNN
NMSE, Directional change statistics (DS) NMSE, DS
Four different networks with modified back propagation for each
MSE Hit ratio Hit ratio
SVR
GA-ANN RW ,ARIMA , linear discriminant analysis, BPANN, SVM, GA-SVM ANN-SV, Garman–Kohlhagen model
GMDH algorithm
FFANN
BP+ Stochastic time effective function BP BP
-
RME, MAPE,DS, Profitability MSE
BP Learning Algorithm with Adaptive Forgetting Factors BP + Adaptive Smoothing+ Momentum Terms GA SVR
ANN models with different indicators as inputs BPANN
SVM
Improved bacterial chemotaxis optimization (IBCO)+BP SVR
FFANN
BP
BSE-30 SENSEX
SVM
Adaptive SVR
SVM RBF
SVR Artificial fish swarm algorithm +K means clustering
Three-layer BP neural network, Regularized RBF neural network BPANN, Case based reasoning RBF optimized by GA ,PSO, ARIMA,ANN,SVM
325
SVR
NMSE, Sign statistics
False alarms
MAE, MAPE
RMSE , MAE, MAPE MSE, RMSE, MAE MSE
NMSE,MAE, DS,WDS Not specified NMSE, DS, MAE Hit ratio Error ratio
2011 6th International Conference on Industrial and Information Systems, ICIIS 2011, Aug. 16-19, 2011, Sri Lanka
[32] [33]
Mean removal Normalize + Selforganizing feature maps Wavelet transform
SVM SVM
SVR SVR + grid search
5 SVM based feature selection methods Single SVR
DS MSE, MAE, MAPE
RNN FFANN FFANN
BP
BP-ANN, conventional ANN optimized by the ABC algorithm, two conventional fuzzy time-series models FFANN, Highly Granular Unsupervised time Lagged Network Adaptive exponential smoothing
RMSE, MAE, MAPE, Theil’s inequality coefficient Profit
[36]
Normalization + Kohonen SOM -
artificial bee colony algorithm(ABC) temporal BP
[37]
-
BP
BPNN, Exponential Smoothing Forecast model
[38]
Normalizing + Interval variation
Exponential Smoothing +FFANN Ensemble SVM
[39]
GARCH(1,1) Volatility
FFANN
Proposed multistage SVM-based nonlinear ensemble model Conjugate gradient based method
Ensemble models trained using simple averaging , Simple MSE, Stacked regression, Variance-based weighting , Artificial neural network BS model with historical volatility ,BS model with GARCH(1,1), SV, SVJ
[40]
Proposed interval sampling method
FFANN
PCA+ Meta-modeling
ARIMA,FNN,SVM, 4 meta-models
[41]
Recurrent Self-Organizing Map + Wavelet transform
-
ANN, SVMs, Generalized autoregressive conditional heteroskedasticity (GARCH)
RMSE
[42]
Chaos-based delay coordinate embedding Normalize
kernel partial least square regressions SVM
SVR
Pure SVR, Chaos-BPNN, BPNN
MSE, RMSE, MAE
GLAR+PCA + FFANNensemble
BP
Generalized auto regression (GLAR),ANN,Hybrid, Equal weights, Minimum error
NMSE, DS
[34] [35]
[43]
V.
CONCLUSION AND SUGGESTIONS
RMSE, Correct tendencies number RMSE, DS RMSE
Average absolute error, average squared errors NRMSE, DS
[2] L. Yu, S. Wang, and K. K. Lai, "Adaptive smooth neural network in foreign exchange rate forecasting," LNCS, pp. 523 – 530, 2005.
Financial forecasting is a challenging problem and still an open research area. This paper presents the empirical and mathematical evidences for working towards most suitable forecasting model. Data preparing and neural architecture design are identified as two main problems. ICA, PCA, WT are superior to available various data preprocessing techniques. However, any comparison between ICA, PCA and WT has not performed in previous approaches. The nonstationary effect can be removed by using WT and clustering input space into multiple clusters. Then the bias variance dilemma can be removed by combining individual networks which are trained for each cluster. SVM is proposed to be the most suitable model based on empirical and mathematical evidence. Structural risk minimization property raises SVM performs better than other algorithms. By going through above mentioned mathematical and empirical evidence, it can be concluded that wavelet decomposition or ICA and PCA with nonlinear and nonstationary extensions can be utilized to preprocess input data. Preprocessed input space clustered with clustering techniques like SOM or its variants, the use of SVM model for each cluster, and combining these models will lead to superior performance and will be a good area for future researches. REFERENCES [1] L. Yu, S. Y. Wang, and K. K. Lai, "Forecasting Foreign Exchange Rates via Artificial Neural Networks," Springer, New York, forthcoming., 2006.
[3] L Yu, S Wang, W Huang, and K K Lai, "Are foreign exchange rates predictable-a survay from artificial neural network perspective," Scientific Inquiry, vol. vol. 8, no. 2, pp. 207 – 228, 2007. [4] G. S. Atsalakis and K. P. Valavanis, "Surveying stock market forecasting techniques – Part II: Soft computing methods," Expert Systems with Applications, pp. 5932–5941, 2009. [5] S. Hayking, "Multilayer perceptrons," in Neural Networks: A comprehensive foundation, 2nd ed. NJ, USA: Prentice Hall, 1999, ch. 4, pp. 211-212. [6] C. J. Lu, T. S. Lee, and C. C. Chiu, "Financial time series forecasting using independent component analysis and support," Decision Support Systems, vol. 47, pp. 115–125, February 2009. [7] R. Ghazali, A. J. Hussain, and P. Liatsis, "Dynamic Ridge Polynomial Neural Network: Forecasting the univariate non-stationary and stationary trading signals," Expert Systems with Applications, vol. 38, pp. 3765–3776, 2011. [8] K. Y. Huang, T. C. Chang, and J. H. Lee, "A Hybrid Model to Improve the Capabilities of Forecasting Based on GRA and ANN Theories," in IEEE Int. Conf. on Grey Systems, Nanjing, 2009, pp. 1687-1693. [9] Y. K. Kwon and B. R. Moon, "A Hybrid Neurogenetic Approach for Stock Forecasting," IEEE Transactions On Neural Networks, Vol. 18, No. 3, May 2007, vol. 18, no. 3, pp. 851-864, may 2007. [10] J. Yao and C. L. Tan, "A case study on using neural networks to perform technical forecasting of forex," Neurocomputing, vol. 34, pp. 79-98, 2000. [11] J. Yao, C. L. Tan, and H. L. Poh, "Neural networks for technical analysis: a study on KLCI," Int. Journal of Theoretical and Applied Finance, vol. 2, no. 2, pp. 221-241, 1999. [12] B. Qian and K. Rasheed, "Foreign Exchange Market Prediction with Multiple Classifi ers," Journal of Forecasting, vol. 29, pp. 271–284, 2010. [13] E. W. Saad, D. V. Prokhorov, and D. C. Wunsch, "Comparative Study of Stock Trend Prediction Using Time Delay, Recurrent and
326
2011 6th International Conference on Industrial and Information Systems, ICIIS 2011, Aug. 16-19, 2011, Sri Lanka
Probabilistic Neural Networks," IEEE transactions on neural networks, vol. 9, no. 6, pp. 1456-1470, November 1998. [14] H. M. Feng and H. C. Chou, "Evolutional RBFNs prediction systems generation in the applications of financial time series data," Expert Systems with Applications, vol. 38, pp. 8285–8292, 2011. [15] L. Yu, S. Wang, and K. K. Lai, "Forecasting Foreign Exchange Rates Using an Adaptive Back-Propagation Algorithm with Optimal Learning Rates and Momentum Factors," in Foreign exchange rate forecasting with artificial neural networks, Fred Hillier , Ed. New York, United States of America: Springer Science+Business Media, 2007, ch. 4, pp. 65-84. [16] L. Yu, S. Wang, and K. K. Lai, "An Online BP Learning Algorithm with Adaptive Forgetting Factors for Foreign Exchange Rates Forecasting," in Foreign exchange rate forecasting with artificial neural networks, Frederick S Hillier, Ed. New York, United States of America: Springer Science+Business Media, 2007, ch. 5, pp. 87-100. [17] L. Yu, S. Wang, and Lai K. K. Kin, "An Improved BP Algorithm with Adaptive Smoothing Momentum Terms for Foreign Exchange Rates Prediction," in Foreign exchange rate forecasting with artificial neural networks, Frederick S Frederick, Ed. New York, United States of America: Springer Science+Business Media, 2007, ch. 6, pp. 101-118. [18] K. J. Kim, "Artificial neural networks with evolutionary instance selection for financial forecasting," Expert Systems with Applications, vol. 30, pp. 519–526, 2006. [19] L. Yu, S. Wang, and K. K. Lai, "A Hybrid GA-Based SVM Model for Foreign Exchange Market Tendency Exploration," in Foreignexchange-rate forecasting with artificial neural networks, Frederick S Hillier, Ed. New York, United States of America: Springer Science+Business Media, 2007, ch. 9, pp. 155-173. [20] P. Wang, "Pricing currency options with support vector regression and stochastic volatility model with jumps," Expert Systems with Applications, vol. 38, pp. 1-7, 2011. [21] M. Mehrara, A. Moeini, M. Ahrari, and A. Ghafari, "Using Technical Analysis with Neural Network for Forecasting Stock Price Index in Tehran Stock Exchange," Middle Eastern Finance and Economics, no. 6, pp. 51-61, 2010.
[31] W. Shen, X. Guo, C. Wu, and D. Wu, "Forecasting stock indices using radial basis function neural networks optimized by artificial fish swarm algorithm," Knowledge Based Systems, vol. 24, pp. 378–385, 2011. [32] L. P. Ni, Z. W. Ni, and Y. Z. Gao, "Stock trend prediction based on fractal feature selection and support vector machine," Expert Systems with Applications, 2010. [33] C. L. Huang and C. Y. Tsai, "A hybrid SOFM-SVR with a filter-based feature selection for stock market forecasting," Expert Systems with Applications, vol. 36, pp. 1529–1539, 2009. [34] T. J. Hsieh, H. F. Hsiao, and W. C. Yeh, "Forecasting stock markets using wavelet transforms and recurrent neural networks: An integrated system based on artificial bee colony algorithm," Applied Soft Computing, vol. 11, pp. 2510–2525, October 2010. [35] S. C. Hui, M. T. Yap, and P. Prakash, "A hybrid time lagged network for predicting stock prices," Int. Journal of the Computer, the Internet and Management, vol. 8, 2000. [36] E. L. de Faria, J. L. Gonzalez, J. T. P. Cavalcante, M. P. Albuquerque, and M. P. Albuquerque, "Predicting the Brazilian stock market through neural networks and adaptive exponential smoothing methods," Expert Systems with Applications, vol. 36, pp. 12506–12509, 2009. [37] L. Yu, S. Wang, and K. K. Lai, "Hybridizing BPNN and Exponential Smoothing for Foreign Exchange Rate Prediction," in Foreign exchange rate forecasting with artificial neural networks, Frederick S Hillier, Ed. New York, United States of America: Springer Science+Business Media, 2007, ch. 7, pp. 121-131. [38] L. Yu, S. Wang, and K. K. Lai, "Forecasting Foreign Exchange Rates with a Multistage Neural Network Ensemble Model," in Foreign exchange rate forecasting with artificial neural networks, Frederick S Hillier, Ed. New York, United States of America: Springer Science+Business Media, 2007, ch. 10, pp. 177-202. [39] R. Gençay and R. Gibson, "Model Risk for European Style Stock Index Options," IEEE transactions on neural networks, vol. 18, no. 1, pp. 193202, January 2007. [40] L. Yu, S. Wang, and K. K. Lai, "A neural-network-based nonlinear metamodeling approach to financial time series forecasting," Applied Soft Computing, vol. 9, pp. 563–574, August 2009.
[22] Z. Liao and J. Wang, "Forecasting model of global stock index by stochastic time effective neural network," Expert Systems with Applications, vol. 37, pp. 834–841, 2010.
[41] S. C. Huang and T. K. Wu, "Integrating recurrent SOM with waveletbased kernel partial least square regressions for financial forecasting," Expert Systems with Applications, vol. 37, pp. 5698–5705, 2010.
[23] Y. H. Wang, "Nonlinear neural network forecasting model for stock index option price: Hybrid GJR–GARCH approach," Expert Systems, vol. 36, pp. 564–570, 2009.
[42] S. C. Huang, P. J. Chuang, C. F. Wu, and H. J. Lai, "Chaos based support vector regressions for exchange rate forecasting," Expert Systems with Applications, vol. 37, pp. 8590–8598, 2010.
[24] J. F. Chang, C. W. Chang, and W. Y. Tzeng, "Forecasting Exchange Rates using Integration of Particle Swarm Optimization and Neural Networks," in 4th Int. Conf. on Innovative Computing, Information and Control, Kaohsiung, 2009, pp. 660-663.
[43] L. Yu, S. Wang, and K. K. Lai, "A novel nonlinear ensemble forecasting model incorporating GLAR and ANN for foreign exchange rates," Computers & Operations Research, vol. 32, no. 10, pp. 25232541, October 2005.
[25] C. J. Lu, "Integrating independent component analysis-based denoising scheme with neural network for stock price prediction," Expert Systems with Applications, vol. 37, pp. 7056–7064, 2010.
[44] J. Kamruzzaman and R. A. Sarker, "Forecasting of currency exchange rates using ANN: a case study," Neural Networks and Signal Processing, vol. 1, pp. 793 - 797, April 2004.
[26] Z. Yudong and W. Lenan, "Stock market prediction of S&P 500 via combination of improved BCO approach and BP neural network," Expert Systems with Applications, vol. 36, pp. 8849–8854, 2009.
[45] T. M. Cover, "Geometrical and statistical properties of systems of linear inequalities with application in pattern recognition," IEEE Transactions on Electronic Computers, vol. EC-14, pp. 326-334, 1965.
[27] S. H. Hsu, J. J. P. A. Hsieh, T. C. Chih, and K. C. Hsu, "A two-stage architecture for stock price forecasting by integrating self-organizing map and support vector regression," Expert Systems with Applications, vol. 36, pp. 7947–7951, 2009.
[46] V. N. Vapnic and A. Y. Chervonenkis, "On the Uniform Convergence of Relative Frequencies of Events to Their Probabilities," Theory of Probability and its Applications, vol. 16, no. 2, pp. 264-280, 1971.
[28] A. U. Khan, T. K. Bandopadhyaya, and S. Sharma, "SOM and Technical Indicators Based Hybrid Model gives better returns on Investments as Compared to BSE-30 Index," in 3rd Int. Conf. on Knowledge Discovery and Data Mining, Phuket, 2010, pp. 544-547. [29] L. J. Cao and F. E. H. Tay, "Support Vector Machine With Adaptive Parameters in Financial Time Series Forecasting," IEEE transactions on neural networks, vol. 14, no. 6, pp. 1506-1519, November 2003.
[47] S. Hayking, "Learning processes," in Neural Networks: A comprehensive foundation, 2nd ed. NJ, USA: Prentice Hall, 1999, ch. 2, pp. 97-98. [48] S. Geman, E. Bienenstock, and R. Doursat, "Neural networks and the bias/variance dilemma," Neural Computation, vol. 4, pp. 1-58, 1992. [49] S. Hayking, "Committee Machines," in Neural Networks: A comprehensive foundation, 2nd ed. NJ, USA: Prentice Hall, 1999, ch. 2.
[30] K. J. Kim, "Financial time series forecasting using support vector machines," Neurocomputing, vol. 55, pp. 307 – 319, March 2003.
327