Research Article A Hybrid Least Square Support Vector ...Research Article A Hybrid Least Square...
Transcript of Research Article A Hybrid Least Square Support Vector ...Research Article A Hybrid Least Square...
-
Research ArticleA Hybrid Least Square Support Vector Machine Model withParameters Optimization for Stock Forecasting
Jian Chai,1 Jiangze Du,2 Kin Keung Lai,1,2 and Yan Pui Lee3
1 International Business School, Shaanxi Normal University, Xian 710062, China2Department of Management Sciences, City University of Hong Kong, Hong Kong3School of Business, Tung Wah College, Hong Kong
Correspondence should be addressed to Kin Keung Lai; [email protected]
Received 30 May 2014; Accepted 20 August 2014
Academic Editor: Shifei Ding
Copyright Β© 2015 Jian Chai et al.This is an open access article distributed under the Creative CommonsAttribution License, whichpermits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
This paper proposes an EMD-LSSVM (empirical mode decomposition least squares support vector machine) model to analyzethe CSI 300 index. A WD-LSSVM (wavelet denoising least squares support machine) is also proposed as a benchmark tocompare with the performance of EMD-LSSVM. Since parameters selection is vital to the performance of the model, differentoptimization methods are used, including simplex, GS (grid search), PSO (particle swarm optimization), and GA (geneticalgorithm). Experimental results show that the EMD-LSSVMmodel with GS algorithm outperforms other methods in predictingstock market movement direction.
1. Introduction
Stockmarket is one of themost sophisticated and challengingfinancial markets since many factors affect its movement,including government policy, global economic situation,investorsβ expectations, and even correlations with othermarkets [1]. References [2, 3] described financial time seriesas essentially noisy, dynamic, and deterministically chaoticdata sequences. Hence, a precise prediction of stock indexmovement can help investors make decisions to take or shedpositions in the stock market at the right time and makeprofits. Many works have been published by researchers tomaximize investment profits and minimize risk. Therefore,predicting stock market is quite important and significant.
Neural networks have been successfully applied in fore-casting of financial time series during the past two decades[4β6]. Neural networks are general function approximationswhich can approximate many nonlinear functions regardlessof the properties of time series data [7]. Besides, neuralnetworks are able to learn dynamic systems which makethem a more powerful tool for studying financial time seriescompared with traditional models [8β10]. However, thereare a couple of weaknesses when neural networks are used
in forecasting financial time series. For instance, when thetypical back-propagation neural network is applied, a hugenumber of parameters are required to be controlled for.This makes the solution unstable and causes overfitting. Theoverfitting problem results in poor performance and becomesa critical issue for researchers.
Accordingly, [11] proposed a support vector machine(SVM)model. According to [12β14], there are two advantagesof using SVM rather than neural networks. One is thatSVM has a better performance in terms of generalization.Unlike the empirical risk minimization principle in tradi-tional neural networks, SVM reduces generalization errorbounds based on the structural risk minimization principle.SVM seeks to achieve an optimal structure through findingout a balance between generalization errors and Vapnik-Chervonenkis (VC) confidence interval. Another advantageis that SVM prevents the model from getting stuck into localminima.
Since the introduction of SVM, it has been developedrapidly in the real world. There are mainly two ways forapplying SVM: one is classification and the other is regres-sion. For classification, [15] constructed a SVM based modelto accurately evaluate the consumersβ credit score and solve
Hindawi Publishing CorporationMathematical Problems in EngineeringVolume 2015, Article ID 231394, 7 pageshttp://dx.doi.org/10.1155/2015/231394
-
2 Mathematical Problems in Engineering
classification problems. Also, SVM is widely used in thearea of forecasting. Reference [16] used SVM to predict thedirection of daily stock price in the Korea composite stockprice index (KOSPI). More recently, [17] applies the SupportVector Regression to forecast the Nikki 225 opening indexand TAIEX closing index after detecting and removing thenoise by independent component analysis (ICA).
However, the performance of SVM mainly depends onthe input data and is sensitive to parameters. Recent empiricalstudies have demonstrated that properties of the modelperformance are influenced by two aspects: low level of signalto noise ratio (SNR) and instability of model specificationduring the estimation process. For example, [18] investigatesthe hyperparameters selection for support vector machinewith different noise distributions to compare the modelperformance. Moreover, [19] applied wavelet to denoise thebearing vibration signals by improving the SNR and thenfigure out the best model according to the performances ofANN and SVM.
To improve the classification and forecasting accuracy,several researchers including [20, 21] have proved that thecombined classifying and forecasting models perform betterthan any individual model. Also, [22] showed that theensemble empirical model decomposition (EEMD) can beintegrated with extreme learning machine (ELM) to an effec-tive forecasting model for computer products sales. In thispaper, we propose a hybrid EMD-LSSVM (empirical modedecomposition least squares support vector machine) withdifferent parameters optimization algorithms. The experi-mental results prove that the EMD-LSSVM model has abetter performance than the WD-LSSVM (wavelet denois-ing least squares support vector machine) model. Firstly,we use the empirical mode decomposition and waveletdenoising algorithm to deal with the original input data.Secondly, parameters of SVM are optimized by differentmethods, including simplex, grid search (GS), particle swarmoptimization (PSO), and genetic algorithm (GA). Resultsfrom empirical studies show that the hybrid model EMD-LSSVM with GS parameter optimization outperforms theother model.
2. EMD-LSSVM Model and WD-LSSVM
2.1. Empirical Mode Decomposition (EMD). References [23,24] proposed empirical mode decomposition (EMD) whichdecomposes data series into a number of intrinsic modefunctions (IMFs). It was designed for nonstationary andnonlinear data sets. In order to apply EMD, time series dataset must satisfy the following two conditions.
(1) The sum of local maxima and local minima mustequate to the total number of zero crossings or thedifference between them is 1. In other words, for everylocal maxima and local minima, there must be onezero crossing following up.
(2) The local average is zero, which means that meanvalue of the upper envelope (defined by localmaxima)and lower envelope (defined by localminima)must bezero.
Thus, if a function is an IMF, it represents a signalsymmetric to local mean zero. An IMF is a simple oscillatorymode which is more general than the simple harmonicfunction and the frequency and amplitude of the IMF canbe variable. Then, data series π₯(π‘) (π‘ = 1, 2, . . . , π) can bedecomposed by the following sifting procedure.
(1) Find all local maxima and minima in π₯(π‘). Then usethe cubic spline line to connect all local maxima togenerate upper envelope π₯up(π‘) and connect all localminima to generate lower envelop π₯low(π‘).
(2) According to the upper and lower envelopes obtainedin Step (1), calculate the envelope meanπ
1(π‘):
π1(π‘) =
(π₯up (π‘) + π₯low (π‘))
2
.(1)
(3) Data series π₯(π‘)minus envelope meanπ1(π‘) gives the
first component π1(π‘):
π1(π‘) = π₯ (π‘) β π
1(π‘) . (2)
(4) Check if π1(π‘) satisfies the IMF requirements; if
π1(π‘) does not satisfy them, go back to Step (1) and
replace π₯(π‘) with π1(π‘) to conduct the second sifting
procedure; that is, π2(π‘) = π
1(π‘) β π
2(π‘). Repeat the
sifting procedure π times ππ(π‘) = π
πβ1(π‘)βπ
π(π‘) until
the following stop criterion is satisfied:
π
β
π‘ = 1
[ππ(π‘) β π
π+1(π‘)]
2
π2
π(π‘)
< SC, (3)
where SC is the stopping condition. Normally, it isset between 0.2 and 0.3. Then, we get the first IMFcomponent; that is, π
1(π‘) = π
π(π‘).
(5) Subtract first IMF component π1(π‘) fromdata sets π₯(π‘)
and get the residual π1(π‘) = π₯(π‘) β π
1(π‘).
(6) Treat π1(π‘) as the new data series and repeat Steps (1)
to (5).Then get the new residual π2(π‘). In thisway, after
repeating π times, we get
π2(π‘) = π
1(π‘) β π
2(π‘) ,
π3(π‘) = π
2(π‘) β π
3(π‘) ,
.
.
.
ππ(π‘) = π
πβ1(π‘) β π
π(π‘) .
(4)
When the residual ππ(π‘) becomes a monotonic function,
the data sets cannot be decomposed anymore. The wholeEMD is completed. The original date series can be describedas the combination of π IMF components and a mean trendππ(π‘); that is,
π₯ (π‘) =
π
β
π = 1
ππ(π‘) + π
π(π‘) . (5)
-
Mathematical Problems in Engineering 3
In this way, the original data series π₯(π‘) can be decom-posed into π IMFs and a mean trend function. Then, we usethe IMFs for instantaneous frequency analysis.
The traditional Fourier transform decomposes a dataseries into a number of sine or cosine waves for the analysis.However, the EMD technique decomposes the data seriesinto several sinusoid-like signals with variable frequenciesand amean trend function.The EMD has several advantages.First, this method is relatively easy to understand and isalso widely applied since it avoids complex mathematicalalgorithms. Secondly, EMD is suitable to deal with nonlinearand nonstationary data series. Thirdly, EMD is more suitablefor analysing data series with trends such as weather andeconomic data. Finally, EMD is able to find the residual whichreveals the data series trends [25β27].
2.2. Wavelet Denoising Algorithm. While the traditionalFourier analysis can only remove noise of certain patternsover the entire time horizon, wavelet analysis can deal withmultiscales and more detailed data and is more suitablefor financial time series. Wavelets are continuous functionswhich satisfy the unit energy and admissibility condition in
πΆπ= β«
β
0
π (π)
π
ππ < β, β«
β
ββ
π (π‘)
2
ππ‘ = 1, (6)
where π is the Fourier transform of frequencyπ. π is thewavelet transform.
The continuous wavelet function can orthogonally trans-form the original data into subdata series in the waveletdomain. Consider
π(π’, π ) = β«
β
ββ
π₯ (π‘)
1
βπ
π(
π‘ β π’
π
) ππ‘, (7)
where u is the dilation parameter and s is the translationparameter.
The wavelet synthesis rebuilds the original data series,guaranteed by the properties of orthogonal transformation in
π₯ (π‘) =
1
πΆπ
β«
β
0
β«
β
ββ
π(π’, π ) ππ’,π (π‘) ππ’
ππ
π 2. (8)
In wavelet analysis, the denoising technique separatesthe data and noise from the original data sets by selectinga threshold. The raw data series are first decomposed intosome data subsets. Then, based on a certain strategy ofselecting the threshold, the boundary between noises anddata is set. Depending on the boundary, smaller data pointsare eliminated and the remaining data are handled by settingcertain thresholds. Finally, these denoised data sets are rebuiltfrom the decomposed data points [28].
2.3. LSSVM in Function Estimation. This section shows thebasic theory of the least squares support vector machine.Thesupport vector methodology has been used mainly in twoareas, that is, classification and function estimation. Consid-ering regression in the set of function π(π₯) = πππ(π₯) + πwith given training data inputs π₯
πβ π π and outputs π¦
πβ π ,
we apply π(π₯) to map π₯πfrom π π to π πβ . Notice that π(π₯)
can be of infinite dimensional and is defined only implicitly.Also, vector π can also be infinite dimensional. Thus, theoptimization problem becomes
min π½π(π, π, π
β) =
1
2
πππ + π
π
β
π = 1
(π + πβ) ,
s.t. π¦πβ πππ (π₯π) β π β€ π + π
π, π = 1, . . . , π,
πππ (π₯π) + π β π¦
πβ€ π + π
β
π, π = 1, . . . , π,
ππ, πβ
πβ₯ 0, π = 1, . . . , π.
(9)
The constant π > 0 defines the tolerance of deviationsfrom the desired π accuracy. It defines the weight of theregularization term empirical risk. The larger the c is, themore important it is for the empirical risk to grow, comparedwith the regularization term. π is called the tube size andrepresents the accuracy required in training data points.
By introducing Lagrange multipliers πΌ, πΌβ, π, πβ β₯ 0, weobtain the Lagrangian for this problem. Consider
πΏ (π, π, ππ, πβ
π; πΌ, πΌβ, π, πβ)
=
1
2
πππ + π
π
β
π = 1
(π + πβ)
β
π
β
π = 1
πΌπ(π + π
πβ π¦π+ πππ (π₯π) + π)
β
π
β
π = 1
πΌβ
π(π + π
β
π+ π¦πβ πππ (π₯π) β π)
β
π
β
π = 1
(ππππ+ πβ
ππβ
π) .
(10)
The reason of introducing another LagrangemultiplierπΌβπ
is that there are other slack variables ππ, πβ
π. Bymaximizing the
Lagrangian
maxπΌ,πΌβ,π,πβminπ,π,ππ,π
β
π
πΏ (π, π, ππ, πβ
π; πΌ, πΌβ, π, πβ) , (11)
we obtain
ππΏ
ππ
= 0 β π =
π
β
π = 1
(πΌπβ πΌβ
π) π (π₯π) ,
ππΏ
ππ
= 0 β
π
β
π = 1
(πΌπβ πΌβ
π) = 0,
ππΏ
πππ
= 0 β π β πΌπβ ππ= 0,
ππΏ
ππβ
π
= 0 β π β πΌβ
πβ πβ
π= 0.
(12)
-
4 Mathematical Problems in Engineering
Table 1: Input and output variables.
Category Input variables selection Outputvariable SampleData
number
Training and validationtesting
CSI 300, USDX, SHIBOR, REPO, CDS, PE,M2/mkt cap, short-mid not/mkt cap, New
Loan/mkt capCSI 300 05/01/2009β23/08/201124/08/2011β20/01/2012
643100
Then we obtain the following dual problem:
maxπΌ,πΌβ
π½π·(πΌ, πΌβ) = β
1
2
π
β
π = 1
(πΌπβ πΌβ
π) (πΌπβ πΌβ
π)πΎ (π₯
π, π₯π)
β π
π
β
π = 1
(πΌπ+ πΌβ
π) +
π
β
π = 1
π¦π(πΌπβ πΌβ
π) ,
s.t.π
β
π = 1
(πΌπβ πΌβ
π) = 0, πΌ
π, πΌβ
πβ [0, π] .
(13)
Here we use the kernel function πΎ(π₯π, π₯π) = π(π₯
π)ππ(π₯π)
for π, π = 1, . . . , π. Then the function estimation becomes
π (π₯) =
π
β
π = 1
(πΌπβ πΌβ
π)πΎ (π₯
π, π₯π) + π, (14)
where πΌπ, πΌβ
πare solutions of the above quadratic program-
ming problem and π is obtained from the complementarityof KKT conditions. It is obvious that the decision functionis determined by the support vectors in which coefficients(πΌπβ πΌβ
π) are not zero. In practice, a larger π results in a
smaller number of support vectors and thus the sparser of thesolution. Also, the larger the π is, the worse the accuracy oftraining points will be. Hence, π can be applied to control thebalance between closeness to training data and sparseness ofthe solution.
Kernel function can be obtained by seeking the functionwhich satisfies Mercerβs condition. Here are some popularkernel functions [14, 29, 30]:
linear: πΎ(π₯, π₯π) = π₯ππ₯π;
polynomial: πΎ(π₯, π₯π) = (π₯
ππ₯π+ 1)π, where π is the
degree of the polynomial kernel;RBF kernel:πΎ(π₯, π₯
π) = exp(ββπ₯ β π₯
πβ2/π2), whereπ2
is the bandwidth of the Gaussian kernel.
Parameters of the kernel function define the structure ofthe high dimensional feature space π(π₯) and also control theaccuracy of the final solution. Thus, they should be selectedcarefully.
3. Empirical Study
3.1. Data Description. The CSI 300 is chosen for empiricalanalysis and to examine the performance of the proposedmodel. This index comprises 179 stocks from Shanghai stock
exchange and 121 stocks from Shenzhen stock exchange andis managed by the China Securities Index Company Ltd.
Most researchers have chosen international indices in thepast, including S & P 500, NIKKEI 225, NASDAQ, DAX,and gold price as input variables. They have examined thecross relationship between stock market index and macroe-conomic variables. The potential input variables that canbe used for forecasting model mainly consist of the grossdomestic product (GDP), gross national product (GNP),short-term interest rate (ST), long-term interest rate (LT), andterm structure of interest rate (TS) [1, 31, 32].
Although China has overtaken Japan to become theworldβs second largest economy and theChinese stockmarkethas developed into one of the most important markets in theglobal economy, Chinese consumption capacity is limited inthe domestic market. The movement of the stock market hasa close relationship with the money available of the investors,which is determined by the money supply and the interestrate. Considering that the Chinese stock market is affectedby the global economic situation as well as the domesticeconomic development, we choose US Dollar Index (USDX),Shanghai Interbank Offered Rate (SHIBOR), P/E ratio (PE),money supply (M2), repurchase agreement (REPO), ChinaCNY Monthly New Loan, market capitalization of the 300publicly traded companies (mkt cap), Peopleβs Bank 5-yearCDS, and short-mid note as input variables.
The lag of input variables is 3 days. We use the daily datato predict the CSI 300 index by nonlinear SVM regression.Since M2, short-mid note, and New Loan are published oncea month, we transform these variables into daily variables bydividing them by a daily variable. We divided all data setsinto two sections and used the first section as the trainingpart to find the optimal parameters for the LSSVM and avoidoverfitting by training and validating the model. The othersection is used for testing. As shown in Table 1, we choosenine variables as the input variables and one variable as theoutput variable including 643 daily data fromMay 1, 2009, toAugust 23, 2011, to train the parameters in the model. Oncewe obtain these parameters, we use the same input and outputvariables from August 24, 2011, to January 20, 2012, including100 daily data to examine the performance of different modelin the testing part.
In the hybrid wavelet denoising least squares supportvector machine model (WD-LSSVM), we first denoise theCSI 300 index with wavelet denoising technique. As shownin Figure 1, the original data, which is depicted in the upperpart of the figure, is packed with irrelevant noise. Then thewavelet denoising algorithm is applied to reduce the noise inthe upper figure of Figure 1. The denoised data is depicted inthe lower part of Figure 1 and it is clear that the denoised data
-
Mathematical Problems in Engineering 5
05/05/2009 28/12/2009 10/07/2010 21/01/2011 23/08/2010150020002500300035004000
Date (dd/mm/yy)
CSI 3
00 in
dex
Original data
05/05/2009 28/12/2009 10/07/2010 21/01/2011 23/08/2010150020002500300035004000
Date (dd/mm/yy)
CSI 3
00 in
dex
Denoised data
Figure 1: The original and denoised daily CSI 300 index.
Table 2: Parameters setting for simplex, GS, GA, and PSO.
Method Parameter settingSimplex Chi: 2, Gamma: 0.5, Rho: 1, Sigma: 0.5GS TolX: 0.001, maxFunEvals: 70, grain: 7, zoomfactor: 5
GA Sizepop: 20, maxgen: 200, πmin: 0, πmax: 100, πmin: 0,πmax: 1000
PSO Sizepop: 20, maxgen: 200, πmin: 0, πmax: 100, πmin: 0,πmax: 1000, π: 0.6
can better reveal the trend of the index. Also, in both EMD-LSSVM and WD-LSSVM models, we preprocess the inputdata by scaling to the range of [0, 1] to prevent small numbersin the data sets from being overshadowed by large numbers,resulting in loss of information.
3.2. Optimization Methods and Parameters Setting. In bothEMD-LSSVM models and WD-LSSVM, we try four kindsof search methods, that is, simplex, GS, GA, and PSO. Inthe simplex method, we define the parameters of expanding(Chi), contracting (Gamma), reflecting (Rho), and shrinking(Sigma) and get the optimal parameters for SVM throughiteration until the stopping criteria is satisfied. Also, bycalculating the objective function, we can get all points in thegrids, which are related to the range and the unit grid searchsize.
The optimal parameters can be obtained from the pointwhich has the lowest cost. Another effective method tosolve optimization problem is the genetic algorithm. Thefirst step of this method is to randomly select parents fromthe population.Then, parents produce children continuously.Step by step, the population eventually develops and optimalsolution can be obtained when the stopping criteria are met.The PSO algorithm works by moving the candidate solution(particles) within the given search range. These particles aremoved by the best known positions of particles and the entire
Table 3: Performance metrics and their calculations.
Metrics Calculation
NMSENMSE =
βπ
π=1(ππβ ππ)2
πΏ2π
πΏ2=
βπ
π=1(ππβ π)2
π β 1
MAPE MAPE =1
π
π
β
π=1
ππβ ππ
ππ
Γ 100%
HRHR =
βπ
π=1ππ
π
ππ=
{
{
{
1 if (ππβ ππβ1) (ππβ ππβ1) β₯ 0
0 otherwise
Table 4: Results of eight different forecasting models.
Model NMSE MAPE HREMD-LSSVM (simplex) 0.0253 0.79222% 77.7778%EMD-LSSVM (GS) 0.0245 0.78834% 79.798%EMD-LSSVM (GA) 2.5749 9.0641% 42.4242%EMD-LSSVM (PSO) 9.4471 18.2733% 40.404%WD-LSSVM (simplex) 0.0521 1.1772% 65.6566%WD-LSSVM (GS) 0.0609 1.1357% 61.6162%WD-LSSVM (GA) 0.0657 1.2997% 62.6263%WD-LSSVM (PSO) 0.0910 1.5072% 63.6364%
swarm in the search space. When the particles arrive at abetter position, they guide the swarm tomove.The procedureis repeated until the stopping criteria are satisfied. In ourexperiment, Table 2 shows the setting of each optimizationmethod.
3.3. Performance Criteria. We evaluate the performance ofthese models using three measurement methods, that is,normalizedmean squared error (NMSE), mean absolute per-centage error (MAPE), and the hitting ratio (HR) (Table 3).NMSE and MAPE are designed to measure the deviationof predicted value from the actual value; smaller values ofNMSE and MAPE indicate better performance of the model.In the stock market, smaller values of MAPE and NMSE areable to control investment risk. We also introduce hittingrate to evaluate the model since the HR reveals accuracy ofprediction of theCSI 300, which is valuable for individual andinstitutional traders.
3.4. Experiment Results. The experiments explore fourparameter selectionmethods in both EMD-LSSVMandWD-LSSVM. Results of the experiments are as in Table 4. Fromthe results, we can see that the hybrid model EMD-LSSVMwith GS parameter optimization method not only has thesmallest NMSE and MAPE but also gets the best hitting rate,which means it outperforms the other model with differentparameter search methods.
From the experiment results, we can draw three conclu-sions.
-
6 Mathematical Problems in Engineering
(1) For overall accuracy, the EMD-LSSVM (GS) is thebest approach, followed by EMD-LSSVM (simplex),WD-LSSVM (simplex), WD-LSSVM (PSO), WD-LSSVM (GA), and WD-LSSVM (GS). Hitting ratesof the other approaches are below 60%. Predictionaccuracy of all methods is also related to the chosensample. So it is difficult to identify which model is thebest and performs the best. However, tests based onthe same samplemay help us identify which is the bestmodel.
(2) According to the experiments, the PSO and GAneed more computational time to obtain the bestparameters for themodel compared with simplex andGS optimizationmethods. Although the PSO andGAalgorithm are relatively more complex than the othertwomethods, they do not perform better thanGS andsimplex.
(3) Another interesting finding is that thresholds of thedenoising algorithm also influence the performanceof the model. When the threshold is too large, usefulinformation in the data gets damaged. Besides, a smallthreshold makes the denoising process insignificantfor handling noise. Therefore, we argue that theperformance of the wavelet denoising algorithm issensitive to the estimation method of the thresholdlevel.
4. Conclusion
We have examined the use of the hybrid EMD-LSSVM andWD-LSSVM models to predict financial time series by fourdifferent parameters selection methods in this paper. Thestudy shows that the hybrid EMD-LSSVM model providesa better way to forecast financial time series compared withWD-LSSVM. The key findings contain two aspects. First,empirical mode decomposition can serve as a potential toolfor removing noise from original data during the modelingprocess and improving the prediction accuracy. Second, wecompare four kinds of search methods for parameters in theexperiments. The results show that the EMD-LSSVM withGS parameter optimization method provides the best per-formance. Use of the GS algorithm reduces the computationtime and improves the prediction accuracy of the model forforecasting financial time series.
Future research in this direction mainly includes gainingbetter understanding of the relationship between optimalloss function, noise distribution, and the number of trainingsamples. In this paper, we only consider applying differentalgorithm to denoise the original data without consideringthe distribution of the noise. The research on the densityof noise which will be reduced for the SVM model willattract the effort of us.Moreover, another interesting researchdirection is to figure out the minimum number of samplesbased on which a theoretically optimal loss function willindeed have superior generalization performance.
Conflict of Interests
The authors declare that there is no conflict of interestsregarding the publication of this paper.
References
[1] W. Huang, Y. Nakamori, and S.-Y. Wang, βForecasting stockmarket movement direction with support vector machine,βComputers and Operations Research, vol. 32, no. 10, pp. 2513β2522, 2005.
[2] J. W. Hall, βAdaptive selection of U.S. stocks with neural nets,βin Trading on the Edge: Neural, Genetic, and Fuzzy Systems forChaotic Financial Markets, G. J. Deboeck, Ed., John Wiley &Sons, New York, NY, USA, 1994.
[3] Y. S. Abu-Mostafa and A. F. Atiya, βIntroduction to financialforecasting,βApplied Intelligence, vol. 6, no. 3, pp. 205β213, 1996.
[4] W. Cheng, L. Wagner, and C.-H. Lin, βForecasting the 30-yearUS treasury bond with a system of neural networks,β Journal ofComputational Intelligence in Finance, vol. 4, pp. 10β16, 1996.
[5] R. Sharda and R. B. Patil, βA connectionist approach to timeseries prediction: an empirical test,β in Neural Networks inFinance and Investing: Using Artificial Intelligence to ImproveRealWorld Performance, R. R. Trippi and E. Turban, Eds., IrwinProfessional Publishing, Chicago, Ill, USA, 1996.
[6] J. R. van Eyden, The Application of Neural Networks in theForecasting of Share Prices, Finance and Technology Publishing,Haymarket, Va, USA, 1996.
[7] I. Kaastra and M. S. Boyd, βForecasting futures trading volumeusing neural networks,β Journal of Futures Markets, vol. 15, pp.853β970, 1995.
[8] G. Zhang and M. Y. Hu, βNeural network forecasting of theBritish pound/US dollar exchange rate,β Omega, vol. 26, no. 4,pp. 495β506, 1998.
[9] W.-C. Chiang, T. L. Urban, and G. W. Baldridge, βA neuralnetwork approach to mutual fund net asset value forecasting,βOmega, vol. 24, no. 2, pp. 205β215, 1996.
[10] F. E. H. Tay and L. Cao, βApplication of support vectormachinesin financial time series forecasting,β Omega, vol. 29, no. 4, pp.309β317, 2001.
[11] V.N. Vapnik,TheNature of Statistical LearningTheory, Springer,New York, NY, USA, 2nd edition, 2000.
[12] K.-R. Muller, A. J. Smola, G. Ratsch, B. Scholkopf, J. Kohlmor-gen, and V. N. Vapnik, βPredicting time series with supportvector machines,β in Proceedings of the International Conferenceon Artificial Neural Networks, pp. 999β1004, Lausanne, Switzer-land, 1997.
[13] S. Mukherjee, E. Osuna, and F. Girosi, βNonlinear predictionof chaotic time series using support vector machines,β inProceedings of the IEEEWorkshop on Neural Networks for SignalProcessing (NNSP β97), pp. 511β520, Amelia Island, Fla, USA,September 1997.
[14] V. N. Vapnik, S. E. Golowich, and A. Smola, βSupport vectormethod for function approximation, regression estimation, andsignal processing,β Advances in Neural Information ProcessingSystems, vol. 9, pp. 281β287, 1996.
[15] C.-L. Huang, M.-C. Chen, and C.-J. Wang, βCredit scoring witha data mining approach based on support vector machines,βExpert Systems with Applications, vol. 33, no. 4, pp. 847β856,2007.
-
Mathematical Problems in Engineering 7
[16] K.-J. Kim, βFinancial time series forecasting using supportvector machines,β Neurocomputing, vol. 55, no. 1-2, pp. 307β319,2003.
[17] C.-J. Lu, T.-S. Lee, and C.-C. Chiu, βFinancial time seriesforecasting using independent component analysis and supportvector regression,β Decision Support Systems, vol. 47, no. 2, pp.115β125, 2009.
[18] V. Cherkassky and Y. Ma, βPractical selection of SVM parame-ters and noise estimation for SVM regression,βNeural Networks,vol. 17, no. 1, pp. 113β126, 2004.
[19] G. S. Vijay, H. S. Kumar, P. P. Srinivasa, N. S. Sriram, andR. B. K. N. Rao, βEvaluation of effectiveness of wavelet baseddenoising schemes using ANN and SVM for bearing conditionclassification,βComputational Intelligence andNeuroscience, vol.2012, Article ID 582453, 12 pages, 2012.
[20] L. Zhou, K. K. Lai, and L. Yu, βLeast squares support vectormachines ensemble models for credit scoring,β Expert Systemswith Applications, vol. 37, no. 1, pp. 127β133, 2010.
[21] Y. Bao, X. Zhang, L. Yu, K. K. Lai, and S. Wang, βAn integratedmodel using wavelet decomposition and least squares supportvector machines for monthly crude oil prices forecasting,β NewMathematics andNatural Computation, vol. 7, no. 2, pp. 299β311,2011.
[22] C.-J. Lu and Y. E. Shao, βForecasting computer products salesby integrating ensemble empirical mode decomposition andextreme learningmachine,βMathematical Problems in Engineer-ing, vol. 2012, Article ID 831201, 15 pages, 2012.
[23] N. E. Huang, Z. Shen, S. R. Long et al., βThe empiricalmode decomposition and the Hilbert spectrum for nonlinearand non-stationary time series analysis,β The Royal Society ofLondon. Proceedings A: Mathematical, Physical and EngineeringSciences, vol. 454, no. 1971, pp. 903β995, 1998.
[24] N. E. Huang, Z. Shen, and S. R. Long, βA new view of nonlinearwater waves: the Hilbert spectrum,β Annual Review of FluidMechanics, vol. 31, pp. 417β457, 1999.
[25] L. Yu, S. Wang, and K. K. Lai, βAn EMD-based neural networkensemble learningmodel forworld crude oil spot price forecast-ing,β in Soft Computing Applications in Business, B. Prasad, Ed.,vol. 230 of Studies in Fuzziness and Soft Computing, pp. 261β271,Springer, 2008.
[26] L. Yu, K. K. Lai, S. Wang, and K. He, βOil price forecasting withan EMD-based multiscale neural network learning paradigm,βin International Conference on Computational Science, pp. 925β932, 2007.
[27] S. Zhou and K. K. Lai, βAn improved EMD online learning-based model for gold market forecasting,β in Proceedings of the3rd International Conference on Intelligent Decision Technolo-gies, pp. 75β84, 2011.
[28] K. He, C. Xie, and K. K. Lai, βEstimating real estate value-at-risk using wavelet denoising and time series model,β inComputational ScienceβICCS 2008, vol. 5102 of Lecture Notesin Computer Science, pp. 494β503, Springer, Berlin, Germany,2008.
[29] S. Zhou, K. K. Lai, and J. Yen, βA dynamic meta-learning rate-based model for gold market forecasting,β Expert Systems withApplications, vol. 39, no. 6, pp. 6168β6173, 2012.
[30] C.-W. Hsu, C.-C. Chang, and C.-J. Lin, A Practical Guide toSupport Vector Classification, Department of Computer Scienceand Information Engineering, University of National Taiwan,Taipei, Taiwan, 2003.
[31] J. Lakonishok, A. Shleifer, and R. W. Vishny, βContrarianinvestment, extrapolation, and risk,β Journal of Finance, vol. 49,pp. 1541β1578, 1994.
[32] M. T. Leung, H. Daouk, and A.-S. Chen, βForecasting stockindices: a comparison of classification and level estimationmodels,β International Journal of Forecasting, vol. 16, no. 2, pp.173β190, 2000.
-
Submit your manuscripts athttp://www.hindawi.com
Hindawi Publishing Corporationhttp://www.hindawi.com Volume 2014
MathematicsJournal of
Hindawi Publishing Corporationhttp://www.hindawi.com Volume 2014
Mathematical Problems in Engineering
Hindawi Publishing Corporationhttp://www.hindawi.com
Differential EquationsInternational Journal of
Volume 2014
Applied MathematicsJournal of
Hindawi Publishing Corporationhttp://www.hindawi.com Volume 2014
Probability and StatisticsHindawi Publishing Corporationhttp://www.hindawi.com Volume 2014
Journal of
Hindawi Publishing Corporationhttp://www.hindawi.com Volume 2014
Mathematical PhysicsAdvances in
Complex AnalysisJournal of
Hindawi Publishing Corporationhttp://www.hindawi.com Volume 2014
OptimizationJournal of
Hindawi Publishing Corporationhttp://www.hindawi.com Volume 2014
CombinatoricsHindawi Publishing Corporationhttp://www.hindawi.com Volume 2014
International Journal of
Hindawi Publishing Corporationhttp://www.hindawi.com Volume 2014
Operations ResearchAdvances in
Journal of
Hindawi Publishing Corporationhttp://www.hindawi.com Volume 2014
Function Spaces
Abstract and Applied AnalysisHindawi Publishing Corporationhttp://www.hindawi.com Volume 2014
International Journal of Mathematics and Mathematical Sciences
Hindawi Publishing Corporationhttp://www.hindawi.com Volume 2014
The Scientific World JournalHindawi Publishing Corporation http://www.hindawi.com Volume 2014
Hindawi Publishing Corporationhttp://www.hindawi.com Volume 2014
Algebra
Discrete Dynamics in Nature and Society
Hindawi Publishing Corporationhttp://www.hindawi.com Volume 2014
Hindawi Publishing Corporationhttp://www.hindawi.com Volume 2014
Decision SciencesAdvances in
Discrete MathematicsJournal of
Hindawi Publishing Corporationhttp://www.hindawi.com
Volume 2014 Hindawi Publishing Corporationhttp://www.hindawi.com Volume 2014
Stochastic AnalysisInternational Journal of