• No results found

BayesianInterpretationofLeastSquaresSupportVector o1-~B

N/A
N/A
Protected

Academic year: 2021

Share "BayesianInterpretationofLeastSquaresSupportVector o1-~B"

Copied!
5
0
0

Bezig met laden.... (Bekijk nu de volledige tekst)

Hele tekst

(1)

o1-~B

Proc. ofthe 5th World Multieonferenee on Systemies, Cybernetics and Informaties (SCI 2001).

Orlando, Florida, Jul. 2001, pg. 254-259

Bayesian Interpretation

of Least Squares Support Vector

Machines

for Financial

Time Series Prediction

Tony Van Gesteil, Johan A.K. SuykensI; Gert LanckrietI, Annemie

Lambrechts\

Dirk-Emma Baestaens2, Bart De Moorl and Joos Vandewaliel

1 Dept. Electrical Engineering ESAT-SISTA, Katholieke Universiteit Leuven Kasteelpark Arenberg 10, B-300l Leuven, Belgium

2 Financial Markets Research, Fortis Bank Brussels Warandeberg 3, B-lOOO Brussels, Belgium

ABSTRACT

For financial time series, the generation of error bars on the point prediction is important in order to estimate the corresponding risk. In Least Squares Support Vector Ma-chines (LS-SYMs) for nonlinear function estimation, the training problem is presented and formulated so as to ob-tain a set of linear equations in the dual space, while the training problem of Support Vector Machines involves a (convex) Quadratic Programming (QP) problem in the dual space. In this paper, a Bayesian interpretation is related to the LS-SYM regression formulation within the evidence framework, in a similar way as it has been done for multilayer perceptrons. The LS-SYM formulation aI-lows to derive analytic expressions in the feature space and practical expressions are obtained in the dual space by replacing inner product with the related positive definite kernel function using Mercer's theorem. We illustrate the method on the one step ahead prediction of the Standard & Poor's 500 Financials stock index (S&PTFIN). Signifi-cant out-of-sample sign predictions are obtain with respect to the Pesaran- Timmerman test for directional accuracy, while trading on the Sharpe Ratio allows to improve the risk adjusted return.

Keywords: Financial Time Series Prediction, Least Squares Support Vector Machines, Bayesian Inference

1.INTRODUCTION

Motivated by the universal approximation property of multilayer perceptrons (MLPs), neural networks have been applied to model and predict financial time series [1, 6, 10, 15]. The focus of many nonlinear forecasting methods [2, 9, 19] is on predicting of next points of a time series, while the error bars on the prediction are typically less important. In financial time series the noise is often larger than the underlying deterministic signal and one also wants to know the error bars on the prediction. Com-bining the prediction and corresponding uncertainty, the return/risk ratio of Sharpe Ratio (SR) allows to compare different investments per unit of risk.

In [12, 13], the Bayesian evidence framework was suc-cessfully applied to MLPs so as to model nonlinear

rela-'Corresponding author. E-mail:

{tony.vangestel.jo-han.suykens }@esat.kuleuven.ac.be

tions and to infer output probabilities on the correspond-ing predictions. However, there are drawbacks to the practical design of MLPs like the non-convex optimization problem and the choice of the number of hidden units. In Support Vector Machines (SYMs), the regression prob-lem is formulated and represented as a convex Quadratic Programming (QP) problem [5, 18, 24, 26]. Basically, the SYM regressor maps the inputs into a higher dimensional feature space in which a linear regressor is constructed by minimizing an appropriate cost function. Using Mercer's theorem, the regressor is obtained by solving a finite di-mensional QP problem in the dual space avoiding explicit knowledge of the high dimensional mapping and using only the related positive-definite kernel function.

In Least Squares Support Vector Machines (LS-SYMs) [20, 21] for nonlinear classification and regression, the use of the least squares cost function with equality constraints allows to obtain a linear Karush-Kuhn-Tucker system in the dual space. This formulation can also be related to regularization networks [7, 10]. When no bias term is used in the LS-SYM formulation, as proposed in kernel ridge regression [16], the expressions in the dual space correspond to Gaussian Processes [27]. However, the ad-ditional insight of using the feature space has been used in kernel PCA [17], while the use of equality constraints and the primal-dual interpretations of LS-SYMs have al-lowed to make extensions towards recurrent neural net-works [22] and nonlinear optimal control [23].. In this pa-per, a Bayesian interpretation is related to LS-SYM re-gression [25]in order to infer the point prediction and the corresponding error bars. While error bars are obtained for MLPs using alocal quadratic approximation to the non-convex cost function, no approximation has to be made for LS-SYMs since aquadratic cost function is used. The resulting return/risk or Sharpe Ratio is then used to trade on the S&P 500 Financials stock index.

This paper is organized as follows. The LS-SYM re-gression formulation is reviewed in Section 2. In Section 3 a probabilistic framework is related to the LS-SYM re-gression formulation by applying Bayes' rule in the feature space. Expressions in the dual space for the probabilistic interpretation of the prediction are derived in Section 4. The daily one step ahead prediction of the S&P 500 Fi-nancials stock index is discussed in Section 5.

(2)

--2. LEAST SQUARES SUPPORT VECTOR

MACHINES

In Support Vector Machines [5, 16, 18, 21, 24, 26] for non-linear regression, the nonnon-linear function Yi

=

f(x;) + ei,

disturbed by additive noise ei, is assumed to be of the following form

Yi

=

wT rp(x;) + b + ei- (1) For financial time series, the output Yi E R is typieally a retum of an asset or exchange rate at the time index i. The input vector Xi E Rn may consists of lagged re-turns, volatility measures and macro-economie explana-tory variables. The mapping rp(.) : Rn -+ RnI is a non-linear function that maps the input vector X into a higher (possibly infinite) dimensional feature space Rnl. How-ever, the weight vector w E RnI and the function rp(.) are never calculated explicitly. Instead, Mercer's condi-tion rp(Xi)Trp(x)

= K(Xi,X) is applied to relate the

func-tion rp(.) with the positive definite kemel funcfunc-tion K. For K(Xi,X) one typically has the following choiees:

. K(Xi, x) = xT x (linear SVM)j

. K(Xi, x) = (xT x+ l)d (polynomial SVM of degree d)j

· K(Xi,X) = exp(-lIx - xi1lV0-2)(SVMwith

RBF-kemel), where 0- is a tuning parameter.

In the sequel of this paper, we will focus on the use of an RBF-kemel.

In Least Squares Support Vector Machines (LS-SVMs) [16, 20, 21], the model parameters w and b of the non-linear function f(x) are inferred from given data D =

{(Xi, Yi)}f:l by solving the following least squares mini-mization problem with ridge regression:

N

.

)

J.t T I"" 2

mm.71(w,e = _

2

w w+ _

2 L.J(iei, w,e i=l (2)

subject to the equality constraints

Yi

=

wTrp(Xi)+ b + ei,

i= 1,...,N.

(3)

The meaning of the hyperparameters J.t and (i (i

=

1, . . . , N) will become clear in the next Sections. When

(i

=

( for i

= 1, . .., N, the minimization problem (2)-(3)

corresponds to the standard LS-SVM formulation [20, 21]. For non-constant (i, the error term corresponds to a weighted least squares cost function. In the probabilistie interpretation of Section 3, non-constant (i are related to the time varying volatility 1/..;(; of the time series [3, 4]. Substitution of (3) into (2) yields the following weighted least squares cost function with ridge regression:

min.71(w,b) = J.tEw+ 2::;:'1(iEv,i

w,b

(4)

with

Ew

=

EV,i

twTw,

tei = t(Yi - wTrp(Xi)- b)2.

(5)

(6)

This formulation will be used in the next Section. Notiee that the solution w, bof (2)-(3) only depends on the ratio's "ti =

~

(i

= 1,..., N),

as introduced in [20, 21].

To solve the minimization problem (2)-(3), one con-structs the Lagrangian .cl(w,b,eja) = .71(w,e)-Z:;:l Qi(WT<p(Xi)+ b + ei

-

Yi), where Qi E JRare the La-grange multipliers (also called support values). The con-ditions for optimality are given by:

{

~

=

0 -+ w

=

2::::1 airp(Xi)

'*" =

0 -+ 2::::1 ai

=

0 (7) ~(Je.

=

0 -+ ai

= "tiei, i = 1,...,N

~

.

=

0 -+ b

=

Yi

-

wT rp(Xi)

-

ei,

i = 1,..., N.

As in standard SVMs, w and rp(x;) are never calculated and by elimination of wand e the following linear system is obtained [20, 21, 25]:

[~n:b~l

] [+] = [f],

(8)

withl Y = [Ylj...j YN], Iv = [Ij ...j 1], e = [elj ...j eN], a = [alj...jaN], D~ = diag(!'Yl,...,"tN]) and where Mercer's condition [5, 18, 26] is applied within the n matrix

nij

=

rp(Xi)T cp(Xj)

= K(Xi,Xj).

(9)

In the optimum we have w

=

2::;:'1

airp(x;) and the LS-SVM regressor is obtained by applying the Mercer condi-tion:

f(x)

=

wT rp(x) + b

=

2::::1

aiK(x, Xi)

+ b. (10)

Efficient algorithms exist in numerieal linear algebra for solving large scale systems. By reformulating the linear system (8) into two linear systems with positive definite data matriees as in [21], iterative methods can be ap-plied such as, e.g., the Hestenes-Stiefel conjugate gradi-ent algorithm. The sparseness property of standard SVMs [7, 24, 26] is lost by the introduction of the 2-norm. How-ever, sparseness can be obtained by sequentially pruning the support value spectrum [21].

3. BAYESIAN INTERPRETATION

IN THE

FEATURE SPACE

Given the data points D = {(Xi, Yi)}f:l and the hyperpa-rameters J.tand (l:N = [(1, ..., (N] ofthe model1l (LS-SVM with kemel function K), we obtain the model parameters by maximizing the posterior P(w, biD, log J.t,log (i,l:N, 1l). Applieation of Bayes' rule at the fust level of inference [2, 13] gives:

P(w, biD, 10gJ.t,log (l:N, 1l)

P(Dlw, b, log J.t,log (l:N, 1l)P( W, bllog J.t,log (l:N, 1l)

= P(Dllog J.t,log (l:N, 1l) ,

(11)

where the evidence P(Dllog J.t,log (i,l:N, 1l) is a normaliz-ing constant such that integration over all possible w and b parameters will give probability 1.

lThe matlab notation [Xl; X2] is used, where [Xl; X2] = [Xl' XnT. The diagonal matrix Da

=

diag(a) E jRNxN has diagonal elements Da(i, i) = a(i), i = 1,..., N, witha E jRN.

(3)

We take the prior P( w, bIlog p" log (I:N, 1l) independent ofthe hyperparameters (i, i.e., P(w, bllogp" log (I:N, 1l) = P(w,bjlogp,1l). Both w and b are assumed to be inde-pendent. The weight parameters w are assumed to be Gaussian distributed with zero mean: P( wllog p" 1l)

=

(f,r) ;t. exp(- ~wT w). This means that a priori we do not

expect a functional relation between the feature vector

'f'(x) and the observation y. Before the data are

avail-able, the most likely model has zero weights Wk

=

0 (k

= 1,... ,nf), corresponding to the EfficientMarket

Hy-pothesis. A uniform distribution for the prior on b is taken, which can also be approximated as a Gaussian distribution P(bilog Ub,1l) =

b

exp(

-

~), with Ub

-+

00. The

V 21r<Tb Uh

prior P(w, bllogp" log (I:N, 1l) in (11) then becomes:

P( w, bilog p" 1l) ex:

':..L

(~)

2 exp(-~wT w). (12)

The negative logarithm of the prior (12) corresponds to the regularization term p,Ew in (5).

We take the likelihood of the observed data D = {(Xi, Yi)}f::l independent of the hyperparameter p, and as-sume that all data points (Xi, Yi) are independent:

P(Dlw, b, log (I:N, 1l) = n;:'1 P(Xi, y;jw, b, log (i, 1l) = n;:'1 P(y;jXi, w, b, log (i, 1l)P(x;jw, b, log (i, 1l) ex:n;:'1 P(y;jXi, w, b, log (i, 1l), (13)

where the probability P(xilw, b, log (i, 1l) is independent from the model1l, i.e., P(xilw, b, log (i, 1l) = P(Xi). The probability P(x;) is assumed to be constant. Assuming that the additive noise ei is drawn from a:Gaussian distri-bution with variance l/(i, we have:

P(Yilxi, W, b, log (i, 1l) =

!fi

exp(-~en

=

!fi

exp(-~(Yi - wT'f'(Xi) - b?).

(14)

Other distributions with heavier tails like, e.g., the student-t distribution, are sometimes assumed in the liter-aturej a Gaussian distribution with time-varying variance (;1 is used here [3].

Substituting (12) and (14) into (11) and neglecting all constants, Bayes' rule at the fust level of inference gives:

P( w, biD, log p" log (I:N, 1l)

ex:exp(-~wT w - 2:;:'1 ~e;)

= exp(-.7i(w,b»,

(15)

with the LS-SVM cost function .11(w, b) defined in (4). Hence, the maximum a posteriori estimates WMP and bMP are obtained by minimizing the negative logarithm of (15), corresponding to the minimization problem (2)-(3). As already explained in the previous Section, the least squares problem (2)-(3) is not explicitly solved in wand b. Instead, the linear system (8) in Q and b is solved in the dual space.

We discuss now an alternative representation of the cost function (4) and the likelihood (15), which will be used in the next Subsection. By observing that the cost function (4) is a quadratic cost function, the posterior

..

P(w, biD, logp"log (i,I:N, 1l) can also be written as the Gaussian distribution:

P(w, biD, logp" log (I:N, 1l)

= 1 exp(_~gTQ-lg), (16)

with 9 = [w - WMPj b - bMP] and Q = covar(w,b) = &(gTg), where the expectation is taken with respect to w and b. The covariance matrix Q is related to the Hessian Hof the LS-SVM cost function (4):

[ lP:1

~

]

-1 Q_ H-1 - 82w2' 8~8b - - 8:1,8b8w 8{J8b . (17)

3. MODERATED

OUTPUT OF THE

LS-SVM IN THE DUAL SPACE

The uncertainty on the estimated model parameters re-sults into an additional uncertainty for the one step ahead

prediction YMP,N+l

=

wLp'f'(x) + bMP, where the

in-put vector X E ]Rn may be composed of lagged returns YN, YN-l,... and of other explanatory variables available at the time index N. By marginalizing over the nuisance parameters w and b [12, 13, 25] one obtains that the pre-diction YN+l is Gaussian distributed with mean

YMP,N+l

=

ZMP

=

wLp'f'(x) + bMP (18)

and variance

UZN+1

=

(N~1+ u;.

(19)

The variance is thus composed of two terms: the fust term (N~1 corresponds to the volatility of the noise eN+! in the next time step N + 1. Different volatility models can be constructed. In this paper, we use a moving average approach based on the last 20 business days. The second term u~ is due to the Gaussian uncertainty on the estimated model parameters w and b in the linear

transform Z

=

wT 'f'(x) +b. We will now derive expressions

for ZMP and u~ in the dual space.

Expression for ZMP

Now, we further discuss the expressions for the mean

ZMP

=

&(z) and the variance u~

=

&[(z

-

ZMP?],

tak-ing the expectation with respect to the Gaussian distri-bution over the model parameters w and b. The mean

ZMP is obtained as: zMP

=

&{z}

=

wLp'f'(x) + bMP,

with w = 2:;:'1 Qi'f'(Xi) [20, 21, 25]. Applying the Mercer condition, we obtain N ZMP = LQ;K(x,x;) + bMP i=1 (20) Expression for u~

Since Z is a linear transformation of the Gaussian dis-tributed model parameters w and b, the variance u~ in the feature space is given by

u;

=

&{(z

-

ZMP )2}

=

&([(wT 'f'(x) + b)

-

(wLp'f'(x) + bMP )]2}

(4)

--

~...-with

tjJ(x) = [<p(x)j1]. The computation for O'~can be obtained without explicit knowledge of the mapping <p('), Using matrix algebra and replacing inner products by the related kemel function, the expression for Ij~ in the dual space is derived in [25]:

Ij~ =8(x)TUoQDUÓ8(X)UT

+ ~1~ Dç~W~lv

p.s

,

- s28(x)TUoQDUÓQDç1v +

}

p-18(x)T Dç1v

I

(22)

+

f

(

+ ~l~DçQUoQDUóQDç\v + .!.K(x,x)

s( p.

with QD = (p.INe" + DO)-l - p-1INe" and the scalar

=

2:::1 (i. The vector 8(x) E JRN and the ma-trices Uo E JRNxNe" and Do E JRNe//xNe// are de-fined as follows: 8i(X)

= K(x, Xi), i = 1,..., Nj Uo(:

,i) = (VO,iQvO,;)!VO,i,i= 1,...,Nell::::; N-1

and

Do = diag(["o,l,...,"O,Ne//]), where VO,i and "O,i are the solution to the eigenvalue problem [25]:

(Dç

-

fçDç1v1~ Dç}QVO,i =

"O,WO,i,

(23)

i = 1, ..., Neii::::; N -1. The number of non-zero eigenval-ues is denoted by Nel I < N. The matrix Dç = diag([(l, ..., (N]) E ]RNxN is a diagonal matrix with diagonal elements Dç(i, i) = (i. Further details can be found in [25].

5. PREDICTING

THE S&P 500

FINANClALS

INDEX

We apply the Bayesian interpretation of the LS-SVM to the prediction of the Standard and Poor's 500 Financials (S&PTFIN) index from till . The fust 600 points are used for training, while the 200 subsequent data points are used for validation. These 200 points were used to select the inputs or explanatory variables of the inputs using the p-value of the Pesaran- Timmerman test for directional accu-racy [14]. We selected 'Y

=

800 and 0'

=

110 corresponding to a Pesaran-Timmerman statistic (PTstat) of 2.23 and corresponding p-value of 0.025. The following inputs were selected:

. S&P 500 Financials Index: lags -1, -2, -3, ..., -9 and moving average over 20 and 40 days

Table 1: Out-of-sample test set performances obtained on the one step ahead prediction of S&PTFIN with dif-ferent modeis: RBF-LS-SVM, RBF-LS-SVMw, ARX and ARXw. The Directional Accuracy is assessed by means of the Percentage of Correct Sign Predictions (PCSP), the Pesaran-Timmerman test statistic (PTstat) and the cor-responding p-value.

. Dow Jones Industrial Index: lags -1, -2, ..., -5 and

moving average over 10 days

. US$-Yen Exchange Rate: lags -2, -3, ..., -9.

All data were retrieved from Datastream. All inputs were normalized to zero mean and unit variance [2], while the output was normalized to unit variance for convenience.

The inputs and hyperparameters 'Y and Ij were then kept fixed and the LS-SVM regressor is validated on the out-of-sample dataset consisting of 1493 data points cov-ering the period from 22/03/1993 till 09/12/1998 (using the dd/mm/yy convention), which includes the Asian cri-sis in 1998. To model possible time varying relations of the financial markets, the predictions were made using the rolling approach, re-estimating the model parameters w and b after 200 data points. The corresponding Percent-age of Correct Sign Predictions (PCSP) and the Pesaran-Timmerman test statistic with corresponding p-values are reported in Table 1. The performance of the LS-SVM with RBF-kernel (RBF-LS-SVM) is here compared with lin-ear regressor autoregressive model with exogenous inputs (ARX), which is estimated using Ordinary Least Squares using exactly the same inputs. While both models yield a significant result with respect to the Directional Accuracy test, the nonlinear model gives a better performance than the linear ARX model. Relating the sign predictions to a classifier performance, the McNemar test [8] can be ap-plied to compare the RBF-LS-SVM model with the ARX model. We obtained a test statistic of 2.34, which corre-sponds to a p-value of 0.0627.

We also illustrate the use of taking different noise lev-els and corresponding weightings into account (RBF-LS-SVMw) and ARX-model (ARXw), as has also been applied in [24]. Different algorithms exists for modeling the volatil-ity of financial time series, related to the variance of the additive noise [3, 4]. Here we use a simple approach tak-ing the 10 days movtak-ing average estimate of volatility [4]. Since the point prediction of the LS-SVM only depends

on the ratio 'Y

=

(/ p (with 'Y

= 800),

we fust estimated

( from the validation set in order to obtain p

=

(h.

We

Table 2: Annualized Returns (Re), Risk (Ri) and cor-responding Sharpe Ratio (SR) of Investment Strategy 1 and Investment Strategy 2 (transaction cost of 0.1%) on the test set comparing the RBF-LS-SVM, RBF-LS-SVMw, ARX and ARXw with a Buy and Hold strategy.

Directional Accuracy PCSP PTstat p-value

RBF-LS-SVM 55.94% 3.69 0.00022

RBF-LS-SVMw 56.43% 3.92 0.00009

ARX 54.97% 3.02 0.00252

ARXw 55.18% 3.08 0.00203

Investnuent Strategy 1 SRl Rel' Rh 1. RBF-LS-SVM 1.1296 17.36 15.35 2. RBF-LS-SVMw 1.3757 20.79 15.36

3. ARX 0.9437 14.25 15.11

4. ARXw 1.0755 15.74 14.63

5. Buy&Hold 0.9498 18.24 19.21

Investnuent Strategy 2 SR2 Re2 Rb

6. RBF-LS-SVMw 1.6354 14.56 8.91

(5)

involves the solution of a QP-problem, the LS-SVM regres-sor is obtained from a linear Karush-Kuhn-Tucker system. The least squares cost function also allows to obtain an-alytical expressions for the variance, while both in SVMs and MLPs use alocal quadratic approximation to the cost function in order to estimate the Hessian. We applied the nonlinear LS-SVM to the daily one step ahead pre-diction of the S&P 500 Financials stock index. Signifi-cant out-of-sample test set predictions were obtained for the Directional Accuracy ofthe Pesaran-Timmerman test. The corresponding uncertainty of the predictions is used in trading on the corresponding risk-adjusted return.

ACKNOWLEDGMENTS

400 350 300 250 200 150 >:<",.~,..z:::y:.~:~:::.,..';:<;:._~.;~._"'" 100 50 Q Q 500 1000

time index (test set)

1500

Figure 1: Cumulative returns on the test set using

Investment Strategy 1 (ISl) assuming homo-scedastic

noise: (1) RBF-LS-SVM (ISl, fuIl Hne, no marker),

(3) ARX (ISl,

dash-dotted

Hne, no marker),

(5)

Buy&Hold (dotted Hne, na marker).

then replaced the constant ( by time varying (i using the corresponding moving average volatility estimates in order to estimate the non-linear RBF-LS-SVMw from (8). We compared the results with a linear ARXw using General-ized Least Squares with diàgonal weighting matrix corre-sponding to the same volatility estimates. The results for Directional Accuracy are summarized in Table 1, which in-dicate that taking into account the time varying volatility of the markets yields better modeis. The McNemar test statistic [8] for a comparison between the RBF-LS-SVMw and ARXw "classifier" performance was 3.4839, with a p-value of 0.0307.

The model was also validated for trading assuming a transaction cost of 0.1% (10 bps [15]). Investment Strat-egy 1 implements a naive allocation of 100% equities or 100% cash, depending on the sign of the prediction. The corresponding annualized return/risk ratios (Sharpe Ra-tio when neglecting the risk free return) for the RBF-LSSVM, RBF-LS-SVMw, ARX, ARXw and the Buy-and-Hold strategy are summarized in Table 2. In order to im-prove the annualized Sharpe Ratio SRl, we define a trad-ing signal based on YMP,NH/UYMP.N+l' with UYMP,N+l from (19). When this trading signal goes above or below a threshold, one changes the position (100% equities

-

0%

cash and vice versa) in Investment Strategy 2. This strat-egy is implemeIited for RBF-LS-SVMw and ARXw, while the thresholds were obtained from the validation set so as to obtain the annualized Sharpe Ratio. This allows to im-prove the Sharpe Ratio on the test set as we can see from Table 2. The cumulative returns of the investment strate-gies with the different models depicted in Figures 1 and 2. The Asian crisis can be recognized by the drop in the cumulative return of, e.g., the Buy&Hold strategy.

In order to illustrate the influence on the model

uncer-+~;'"+... "' +~ , , 1 1C"' r1nT'O.;,.+nr1 +1,n ,, "'... h<101''"CO

[11] T. Kailath, Linear Systems, Prentice Hall, Engle-wood Cliffs, 1980.

[12] D.J.C. MacKay, "Bayesian Interpolation", Neural Computation, vol. 4, pp. 415-447, 1992.

[13] D.J.C. MacKay, "Probable Networks and Plausibie Predictions

-

A Review of Practical Bayesian ~ethods for Supervised Neural Networks", Network: Computation in Neural Systems, vol. 6, pp. 469-505, 1995.

[14] M.H. Pesaran and A. Timmerman, "A simple non-parametric test of predictive performance", Journal of Business and Economic Statistics, vol. 10, pp. 461-465, 1992.

[15] A.N. Refenes, A.N. Burgess and Y. Bentz, "Neural Networks in Financial Engineering: A Study in

Methodol-. .,..,,,...,.~ ~~~ .~.;.~ 400 350 300 250 200 150 ;;.t!~~~"'" . ' 5 50 Q Q 500 1000

time index (test set)

1500

Figure 2: Cumulative returns on the test set using

In-vestment Strategy 1 (IS1) and InIn-vestment Strategy

2 (IS2) assuming hetero-scedastic noise:

(2)

RBF-LS-SVMw (ISl,

fuIl Hne, marker +),

(4) ARXw

(151, dash-dotted Hne, marker +), (6) RBF-LS-SVMw

(IS2, fuIl Hne, marker

0

)and (7) ARXw (IS2,

dash-dotted Hne, marker 0). The cumulative return of the

Buy&Hold strategy is denoted by the dotted Hne (5).

3.0

"0

500 1000 1500

tJme Index (test set)

Figure 3: Error bars of the RBF-LS-SVMw model.

(A) Total standard deviation UYMP,N+l

=

((N~l+u~)!

from (19) (fuIl Hne) and (B) standard deviation (jz

from (22) due to the uncertainty on the model

param-eters (dashed-dotted Hne). 0 bserve the uncertainty

on the model parameters during the Asian crisis.

(19) and the standard deviation Uz from (22) due to the uncertainty on the model parameters are depicted. It can be observed that the model uncertainty becomes very high during the out break of the Asian crisis in the last part of the graph.

6.CONCLUSIONS'

A probabilistic interpretation has been related to the Least ~n"<1o,..ftC' <::.."."1"'\",,,+"0."'+,,_ "JI'"3 1-.1T'100IT .CL~'T"A'\ fnrm111~tinn

Referenties

GERELATEERDE DOCUMENTEN

Especially in terms of the performance on recursive simulation, the  and  models perform significantly better than the  model as the

Suykens is with the Department of Electrical Engineering-ESAT, STADIUS Center for Dynamical Systems, Signal Processing and Data Analytics, and iMinds Future Health Department,

After a brief review of the LS-SVM classifier and the Bayesian evidence framework, we will show the scheme for input variable selection and the way to compute the posterior

The aim of this study is to develop the Bayesian Least Squares Support Vector Machine (LS-SVM) classifiers, for preoperatively predicting the malignancy of ovarian tumors.. We

We start in section 2 with a brief review of the LS-SVM classifier and its integration within the Bayesian evidence framework; then we introduce a way to compute the posterior

In order to compare the PL-LSSVM model with traditional techniques, Ordinary Least Squares (OLS) regression using all the variables (in linear form) is implemented, as well as

Furthermore, it is possible to compute a sparse approximation by using only a subsample of selected Support Vectors from the dataset in order to estimate a large-scale

For the case when there is prior knowledge about the model structure in such a way that it is known that the nonlinearity only affects some of the inputs (and other inputs enter