High Frequency Trading Portfolio Optimisation - Semantic Scholar

3 downloads 270 Views 589KB Size Report
High Frequency Trading Portfolio Optimisation: Integration of financial and human factors. Å arunas Raudys and Aistis Raudys. Vilnius University, Faculty of ...
High Frequency Trading Portfolio Optimisation: Integration of financial and human factors Šarunas Raudys and Aistis Raudys Vilnius University, Faculty of Mathematics and Informatics Naugarduko st. 24, LT-03225 Vilnius, Lithuania [email protected]; [email protected] Abstract— To use human factors together with financial ones in portfolio management task we analyze lengthy series of successes and losses of numerous automated high frequency trading systems that buy and sell assets. We found that in spite of sparse, bimodal non-Gaussian time series, modern Markowitz solutions can be applied to weigh up contributions of diverse trading strategies. Training history should be rather short in situations where technological, social, financial, economic and political situations are changing swiftly. The Markowitz portfolio coefficients finding algorithm can be improved by careful application of the regularization and matrix structurization methods. Keywords— high frequency trading; Markowitz; portfolio optimisation; regularization; sample size

I.

INTRODUCTION



T

T

X = [ X1 , X1

T

, ... , X n

T

],

where i = [xi1, xi2, … , xip] are p-dimensional data vectorsrows and superscript “T” denotes a transpose. Portfolio management approaches based on the assets and based on the successes of trading systems are two different lines of attack. Important characteristic feature of latter one is a fact that the data’s larger part (72% in this case) consists of zeros. The distribution of the data becomes bimodal. If the number of trading strategies, p, is large, small sample problems arise in PM. Sparse data matrix requires that the aggregation (fusion) rule should be as simple as possible. Thus, we will use linear rule, i.e. weights W=(w1, w2, …, wp)T will be availed to assess p first level decisions

~ 

PNLi = w1×xi1+ w2×xi2 + … + wp×xip,

(ws > 0)

(1)

 

Sample based average and variance, P N L and varPNL, of T and W StrainW, where returns PNLi are T  1 1 = ¦i i and Strain= ¦i i i are training data based n n  mean vector and covariance matrix, i = i - are centred p-dimensional data vectors. Finding the weight vector W can





696





c 978-1-4577-1675-1 2011 IEEE

A. Profit and loss (PNL) data matrix Many aspects of investment analysis are said to be psychological in nature [6]. In order to manage portfolio “optimally” one needs to take into account both the financial and human factors. One may expect that diverse strategies are employing similar profit and loss data, however, strategies incorporate into the decision making algorithm their designers’ human factors. Therefore, one of the ways to take into account both types of factors is to consider long lasting history of the performance of a large number of the high frequency trading (HFT) strategies.. To integrate both kinds of information, one needs to simulate the large number of trading strategies (say, p ones) during a long time period (say n time moments) and calculate virtual profits obtained if each one of the p strategies would be applied. In such a way one obtains a n×p dimensional profit and loss data matrix



Portfolio management (PM) task is one of very important research topics in financial markets (see e.g. classical Markowitz paper [1, 2] and reviews of thousands of papers aimed to improve the Markowitz’s solution [3, 4, 5]). During the last decade the financial markets turned towards the high frequency trading [9, 10]. It potentially promises good returns with relatively small risk [11], as positions are rarely held overnight. Positions typically last from a few seconds to a few hours or a day. This line of attack minimises the risk of big losses. A gain is possible due to very short term market inefficiencies. Historically, majority of works aim at creating portfolio for a set of asset – an investment universe. We have rather different approach here. In real life we are facing hundreds and thousands of diverse trading strategies that acting together are influencing the financial markets. Diverse agents (market participants) are using individual portfolio managing strategies, decision making algorithms differing in trading logic, assumptions, concrete compositions of the multidimensional time series, exactness and timing of the empirical data, and methods used to estimate their statistical characteristics. Development of technologies and the society, information explosion, political and psychological aspects (a length of investment horizon, mispricing, investor credulity, preference to take risk, a desirable gain, a time of expected profits, etc., - we refer to them as human factors [6, 7, 8]), economic and financial indexes are changing swiftly. For that reason, we analyse portfolio construction methods based on a large number of trading strategies. What distinguishes the new PM approach from previous ones is that here one does not create the portfolio based on assets. One creates the portfolio based on success and losses of

automated trading systems that trade assets. On top of that, the underlying trading systems are short term / high frequency, so generated returns are vaguely correlated to variations of the underlying instruments.

be performed as calculations of an efficient frontier for a T number of parameter qj = WX values, as is done in the standard portfolio management where proportions of various assets are determined [2]. For a set of a priori selected qr values, q1, q2, …, qR, the efficient frontier is found by T minimizing W StrainW. Minimization can be done by using a Lagrange multiplier. Here we maximize a probability that return (1) is smaller as any a priori given value Pgiven, provided a distribution of the return (1) is Gaussian. The procedure actually maximizes the Sharpe ratio (sample average/standard deviation)

(w1, w2, … , wp)T in p = 142 dimensional space. We used artificially generated data of 142 trading strategies based on 8 years financial data time series history and the standard Markowitz algorithm “frontcon” realized in Matlab Financial Toolbox. Underneath it uses proprietarily algorithm QPLCPROG - positive-semi-definite quadratic programming based on linear complementary programming. In our approach, larger part of the data consists of zeros. In order to avoid the singularity of covariance matrix, we added constant h = 0.1 to diagonal elements of the matrix. We will call this algorithm as “standard Markowitz”. 8.6

T

Sh = PNL / stdevPNL, = XW (W StrainW)

-1/2

8.4

where one is obliged to satisfy following two constraints:

8.2

ws > 0 (s = 1, 2, … , p), and ¦s ws = 1.

Sharpe ratio

, (2)

(3)

One more way to estimate W = (w1, w2, … , wp) is straightforward: one generates very large number, say m random vectors Wj, (j= 1, 2, … , m) , weighs up their quality and selects the best. If distribution of inputs is a mixture of multivariate distributions, instead of the Sharpe ratio (2) one would need to employ more complex non-linear objective functions [12,13]. Similar situation we face in our approach where majority of data points are “zeros”. Complex objective functions, however, require more training data [14]. So, we use simple linear profit and loss model and apply traditional Sharpe criterion B. Sample size issues in HFT portfolio creation The world is changing in a chaotic manner. Financial markets are affected by modern data storing, processing and transmission technologies. Objective and subjective active human interferences and intelligent mathematically based methods are changing even faster. Speaking in general, a lot of popular indicator-based trading rules turn out to be lossmaking when applied individually in a systematic manner. Successful traders tend to adapt to market conditions by ‘dropping’ trading rules in situations where they become loss-making or when more profitable rules are found [16, 17]. This aspect forces the decision making algorithms adapt to changes as rapidly as possible. In dynamic financial time series prediction, a decision rules based on short data sequences result more accurate predictions as using lengthy historical data. In [18] for a random drift model of the changes and multidimensional time series data model it was shown that the generalization error decreases at first with an increase in the length of training history, reaches minimum and starts increasing afterwards. This conclusion was confirmed in empirical analysis of prediction of commodity prices. The dilemma with optimal training set size cannot be avoided in the portfolio management too. In situations where time series are changing in time, genetic algorithms [16], evolvable multiagent systems [19] can be used. In Figure 1 we present dependence of the Sharpe ratio on the length of historic data used to find weight vector W =

8 7.8 7.6 7.4 7.2 200

400 600 800 Training history (in days)

1000

Figure 1. Dependence of the Sharpe ratio on the length of training data. This example illustrates a necessity to shorten the length of training history if the data are changing very often. If the size of training data is finite and dimensionality, p, is high, problems arise with estimation of the covariance matrix. Especially this question is important in our problem setting, where the data is sparse. In the eight years 142 dimensional data, 72% of positions were filled with zeros - prudent trading strategies refuse from active transactions very often. Possible ways to mitigate small sample size/dimensionality problems in machine learning are: 1) to use regularization, 2) apply diverse constraints (simplified theoretical models of the covariance matrix), 3) add a noise to training data [20]. II.

FUSION METHODS TO CONSTRUCT HFT PORTFOLIO

A. Sequential forward selection method – Comgen Modern portfolio theory in an inexplicit way assumes that PNL (or assets returns) are Gaussian random variables, variances and correlations between inputs (assets) do not change. In order to avoid the assumptions just mentioned in one of the fusion procedures investigated we applied rule with integer weights. Here sequential forward selection procedure was used to select the best subset of inputs (the HFT strategies). In training phase, the quality of the portfolio was evaluated on the particular training data set. Usage of this method was suggested by our time series provider – proprietary trading firm and was named “Comgen”, the portfolio Combination Generator. In this rule initially we start with the best trading system, Ta, with the highest Sharpe ratio. In the second step we

2011 11th International Conference on Intelligent Systems Design and Applications

697

consider p pairs, Ta+Tj, (j = 1, 2, … , p), and select the pair with the highest Sharpe, say TB = Ta+Tr (r is the index of the best pair). And again, we select the best pair TC = TB+Tj, (j = 1, 2, … , p), where TB contains previously selected systems. This process continues by adding the single system that appears to be the best when combined with previously selected trading systems. In each step we add value 1 to a selected position in the weight vector. As a result we get integer weights starting from 0. We repeat the procedure until we reach p - the maximum number of trading systems in our portfolio (p =142 in this particular task). The benefits of this system is in its simplicity and granularity, it creates integer portfolio weights that can be traded straight away. In principle, other heuristic data mining variable selection procedures such as backward selection, genetic algorithms or diverse variants of random search can be applied (see, e.g. [15] or [20], Section 6.3.2). B. Standard weakly regularized Markowitz algorithm for aggregation of PM strategies In our research we use two modifications of the Markowitz algorithm. The first one was the standard Markowitz algorithm (“frontcon” function in Matlab). As inputs served sample training data based mean vector X and weakly regularized estimate of covariance matrix, Strain® = Strain + hI, where X and Strain were defined above, h is very small regularization constant (we used h=0.1, when largest diagonal elements of matrix S exceeded 105) to avoid the singularity of matrix Strain, I is p×p-dimensional identity matrix. As the output we obtain weight vector W = (w1, w2, … , wp). Optimisation procedure used in the Matlab’s Markowitz PM algorithm brings into play tolerances. The tolerances actually serve as means of weak regularization. If the regularization parameter, h, in sample estimate matrix Strain® is set to zero, the algorithm stops working in situations where training data set has too much of “zeros”. Therefore, usage of h>0 was obligatory. C. Three times regularized Markowitz PM algorithm There exist a number of regularization techniques [21, 22, 23, 24] that can be used in the PM task. Standard techniques require certain parameters to be estimated from learning data (variances of the input variables, eigenvalues, essential factors, etc.). Often they appear to be too complex to be used in nearly singular data situations. Thus, in similar circumstances we tried to use regularization means as simple as possible. Instead of using the Lagrange multiplier we insert two additional “penalty terms” into loss function and minimized modified expression T

(

T

SθMj = W Strain®W + Λ WX −

) +Λ (1W 2

T

)

2

− 1 , (4)

where 1=[1 1 … 1] stands for a vector row composed of p ones, Λ controls a tolerance for constraint (3) violations. Parameter Λ is a sufficiently large positive constant to ensure

(

T

both supplementary terms, Λ WX −

) and Λ (1W 2

T

(4) is found by equating a derivative to zero and solving an equation obtained:

∂S qȁ ∂W

(

T

= 2W Strain®+2Λ W X −

(

)

+2Λ 1W T − 1 1 =

)X 0, and T

T

W = (qj X +1) (Strain® /Λ + X X +1 1) −1.

n

p(x1, x2, ..., xn) = ∏

j =1

p ( xj⏐xm ) j

(0≤ mj ≤ n),

(7)

where a sequence m2, ..., mn constitutes a graph of (an unknown permutation of the integers 1, connections, 2, ... , n), and p(xi⏐x0), by definition, is equal to p(xi); for details see [26] or [20], Appendix A2. In this model one estimates only p variances and p-1 correlations. An inverse of estimate S has only 2p-1 nonzero elements. In a general case, the matrix S, however, is composed of p2 nonzero elements. If assumptions concerning the first order dependence structure are correct, we can obtain significant improvements in accuracy in finite sample size/high dimensionality situations. In our weight W determining algorithm we incorporated both regularization means, the regularization with h=30,000 and the first order dependence structurization. Use of the penalty terms in Eq. (4) tolerates the inaccuracies arising in the weight vector estimation. In such a way, the penalty terms serve as the third additional “regularizer” in the weight vector seeking algorithm. Therefore, we will refer to the novel modification as the “Three times regularized Markowitz algorithm”. Both Markowitz PM based algorithms create fractional weights. One has to scale and round them to get tradable weights (one cannot trade fractional part of the trading system). The downside is that the rounding reduces portfolio accuracy and the Sharpe ratio (∼8-10%).

2

converge to zeros, Λ=108). Given qj and Λ, the minimum of

698

(6)

To satisfy requirements (3) we brought negative weights to naught, if wj ≤ 0 and normalized W to meet with ¦s ws = 1. In the case of multimodal non-Gaussian data the complex methods can be less efficient as simple ones. Therefore, we used the training data based regularized matrix Strain® =Strain + hI with large value of parameter h in our weight seeking algorithm. This regularization technique actually reduces the correlations. For that reason, we introduced a way to take into account “principal dependencies”: we started to use the first order dependence tree [25, 26, 20] to improve estimate of the covariance matrix. The first order dependence tree model assumes each component of the input vector depends directly upon only one other component. The probability density function of random vector X = (x1, x2, ..., xn) having the first order tree type can be written in the very simple form:

)

−1

(5)

2011 11th International Conference on Intelligent Systems Design and Applications

III.

SIMULATION STUDY

A. An objective of the study The objective is to compare efficacies of three novel schemes of PM rules with benchmark one: a) equally weighted - when all HFT schemes were evaluated with the same weights (w1 = w2 = … = wp = 1). This is the original weighting as proprietary trading house was using (the benchmark method); b) Comgen, our heuristic algorithm with integer weights obtained in sequential forward selection; c) standard Markowitz algorithm - linear weighting of p PM trading strategies realized in Matlab Financial Toolbox; d) three times regularized Markowitz algorithm - linear weighting of p HFT trading strategies with the novel modification of the Markowitz algorithm (Eq. (6)), simple regularization, Strain® =Strain + hI, and the first order dependency tree structure (Eq. (7)) of the covariance matrix. B. Data used From the high frequency trading firm we received a set of 142 profit and loss series representing simulated track records of real life trading systems. The firm was trading them equally weighted and the aim was to create a portfolio maximizing risk-reward ratio – the Sharpe ratio. These systems were selected qualitatively by the analyst looking at their performance and selecting less correlated ones. The firm is trading US and European futures, for example: stock index, energies, metals, commodities, interest rate products, FX products and etc. In the portfolio we had high frequency and short term systems. Majority of the systems exits in 24h after position was taken but some can hold positions longer. Most of the time systems tend not to trade but wait for the right opportunity. As a result we have very sparse data mainly filled with zeros - where systems were not trading. This is very interesting challenge for us as it makes it difficult to compute valid covariance matrix. Data range: 01-Jan-2003, 17-Dec-2010 (8 years) Number of trading systems: 142 Number of days in the history: 2078 Total number of values: 295076 Out of them zeros: 212004 (71.84%)

~ 

C. The experiment design All experiments were performed in out-of-sample (oos) regime using walk forward methodology. One trains the model with N p-dimensional vectors of time series and tests on subsequent M vectors. Next, one walks M days forward: we shift the training and validation sets’ windows by M days. We used 10 walk forward iterations with M = 100 day period. So, for each size of training set, we trained the fusion algorithms ten times. As a result we got total of 10×100=1000 days out of sample trading history of values  i (i = 1, 2, … , 10), to estimate the Sharpe ratio. To have fair comparison, in both Markowitz PM algorithms the training set information and R = 50 values were used to calculate the efficient frontiers. The portfolios

were designed on training samples of length N = 200, 300, … , 1000 and tested on following M days. In the first series of the experiments we considered p=142 HFT strategies (trading robots). Results obtained from single data set study with 2100 days history are not satisfactorily conclusive. For that reason, to increase the validity of experimental results we formed K=25 randomly chosen sets of HFT strategies composed of p=100 trading robots each:

• •

• •

From 142 set select 100 HFT strategies. Perform walk forward portfolio optimisation of all four methods using: o N=200, 300, … , 1000 days training set and M=100 days forward periods, o Compute the Sharpe ratios on all four systems in out of sample data. Repeat the experiments K=25 times. Results are averaged form 25 experiments

In such experiment design instead of single estimates we were able to evaluate averages and standard deviations of the Sharpe ratio. D. Results The experiments performed with (c) and (d) aggregation schemes showed that both Markowitz’s solution based fusion schemes require regularization Strain® =Strain + hI. After short test, for all future experiments we selected h=1000 for standard Markowitz algorithm and h=30,000 for the new modification. In Fig. 2a we present graphs of dependence of the Sharpe ratio on training sample size in the single series of experiments in experiments with p=142 trading schemes. In Fig. 2b we present average values obtained in K=25 series of experiments with p=100 randomly selected collections of trading systems. In Fig 3a we provide more details of the experiments with the 25 runs. We show mean values, m, for both variants of the Markowitz algorithm calculated for 9 diverse sample sizes. We also demonstrate deviations from m by two standard deviations of the means. In Fig. 3b we present a scatter diagram of Sharpe ratio. The graphs and the diagram indicate that the improvement of the Three time regularized Markowitz rule is significant. E. Evaluation of the results Experiments show that the portfolio based on the weighted evaluation of success of automated trading systems allows increasing the Sharpe ratio of the portfolio. In comparison, trading the assets is notably less successful as one has to absorb not only up trends but down trends as well. It means that the information acquired from analysis of multiplicity of the HFT strategies was useful. The benchmark (“equally weighted”) algorithm was notably outperformed by the Comgen. The standard Markowitz resulted in similar accuracy (Fig.2). In Fig.2a we fused p=142 single HFT schemes and Fig.2b we see averages of 25 additional experiments with 100 randomly selected trading schemes. We can see that additional regularization (h=1000) is very useful to increase the profit.

2011 11th International Conference on Intelligent Systems Design and Applications

699

b

a

Traingin sample size vs quality (142 systems)

Traingin sample size vs quality (average of 25 experiments)

9

7.5

8.5

7

Sharpe on OOS

8

Sharpe on OOS

7.5 7

6.5

6

5.5

6.5

5

6 Standard Markowitz Comgen Original Three Times Regularized Markowitz

5.5 5 200

300

400

500 600 700 Training sample size

800

900

Standard Markowitz Comgen Original Three Times Regularized Markowitz

4.5

4 200

1000

300

400

500 600 700 Training sample size

800

900

1000

Figure 2. Dependence of the Sharpe ratio on the length of training data: a) – p=142 trading systems; b) – p=100 trading systems (average values in K=25 series of the experiments)

Sharpe ratio, Three Times Regularized Markowitz

a

8

Regularized Markowitz 7.5

sharpe ratio, oos

7 6.5

Standard Markowitz

6 5.5 5 4.5 200

400

600 800 training sample size

1000

b

9 8.5 8 7.5 7 6.5 6 5.5 5

5

6 7 8 Sharpe ratio, Standard Markowitz

9

Figure 3. Accuracy of comparison or standard and 3 times regularized Markowitz algorithms in K=25 series of the experiments: a) mean values and means ± 2×standard deviations/ K , b) scatter diagram of Sharpe ratios (N=500) The regularization and covariance matrix structurization are necessary if the fusion weights are found from Eq. (6). Value of the optimal regularization is not very crucial parameter. Differences between mean values of the Sharpe ratio of the novel and old methods are larger than two standard deviations/ K for all values in training sample size. IV.

CONCLUDING REMARKS

To take into account human factors acting together with financial and economy ones we started analyzing long lasting histories of successes of employment of diverse HFT strategies in the portfolio management tasks. Instead of creating the portfolio based on assets we create the portfolio based on successes and losses of multiple automated trading systems that trade the assets. The first difficulty we faced was sparse, bimodal time series data. We found that in spite of the non-Gaussian distribution of the training data, the Markowitz solution can be successfully applied to weight contributions of diverse

700

HFT strategies in automatic portfolio design. We become aware of that the training data history should be rather short in situations where the technological, financial, economic and political situations are changing swiftly. The tolerances used in optimisation moderately fix sample base estimates of the covariance matrix in the Matlab Financial toolbox. The Markowitz PM algorithm can be additionally improved by application of the regularization and covariance matrix structurization methods. The simulation experiments showed that regularization techniques reduce the decline in the accuracy and the Sharpe ratio when the sample size is decreasing or increasing too much. This fact gives a hope that the novel methodology could be applied in the future if the environmental situations will change more abruptly and unexpectedly. We also state that sample size is important in portfolio management task. From our experiments we recommend at least 400 days of history to include in PM task. In the case of regularized covariance matrix this does not seem to significantly change further increasing training size. That of

2011 11th International Conference on Intelligent Systems Design and Applications

course is subjective and may depend on the number of trading systems, types of the systems and a frequency and power of the changes. The algorithmic trading is affected by the outbreak of use of modern computers and telecommunications means and counts its history a decade of years. The HFT activities are affecting the markets. For that reason, the data histories useful for comparison of different trading methods are not sufficiently lengthy. Like in pattern recognition, the sample size/complexity considerations [14] are very important in portfolio management research in rapidly changing environments. We considered only a small number of means that are useful in this task, however, the number of covariance matrix simplification tools is much larger and constitutes a subject for future wide scale investigations. The Efficient-Market Hypothesis (EMH) asserts that the market is efficient and whenever new information comes up, the market absorbs it by correcting itself [27, 28, 29]. Simultaneous analysis of numerous long lasting success histories of diverse trading robots brings in additional information and generates the model where it is assumed that the future asset prices depend on the predictions of all participants of the market. It provides some theoretical basis for explanation of EMH and can be useful for analyzing simulation results obtained by exploration of prediction models of the financial time series when asset prices depend on predictions. Theoretical considerations and preliminary experiments showed that proposed regularization methodology can also be used for liquidity analysis and transaction cost management. ACKNOWLEDGMENT This research was funded in part by a grant (No. MIP20/2010) from the Research Council of Lithuania. The authors also want to express their appreciation to Vilnius University. REFERENCES [1] H.M. Markowitz, H.M. (1952). Portfolio selection, The Journal of Finance 7 (1): 77–91, 1952. [2] H. M. Markowitz. Portfolio Selection. Efficient diversification of investments, pp. 344. John Wiley & Sons: New York; Chapman & Hall: London, 1959. [3] R.C. Grinold and R. N. Kahn. Active portfolio management: A quantitative approach for producing superior returns and selecting superior returns and controlling risk, McGraw-Hill, 2nd ed., 1999. [4] R. J. Elliott and P. E. Kopp. Mathematics of Financial Markets, Springer, NY, 2005. [5] P.D. McNelis. Neural Networks in Finance: Gaining Predictive Edge in the Market. Elsevier Academic Press, 2005. [6] P. Slovic. Psychological study of human judgment: Implications for investment decision making under uncertainty, Journal of Finance, vol. 27(4), pp. 779-799, 1972. [7] D. Hirshleifer. Investor psychology and asset pricing. Journal of Finance, vol. 56(4), pp. 1533–1597, 2001. [8] K. Daniel, D. Hirshleifer and S.H. Teohc. Investor psychology in capital markets: evidence and policy implications, J. of Monetary Economics, vol. 49, pp. 139–209, 2002.

[9] E.P. Chan. Quantitative trading: how to build your own algorithmic trading business. Hoboken, N.J., Wiley; Chichester: John Wiley [distributor], 2008. [10] I. Aldridge. High-frequency trading: a practical guide to algorithmic strategies and trading systems. Hoboken, N.J., Wiley ; Chichester : John Wiley [distributor], 2010. [11] R. Bookstaber. Risk from high frequency and algorithmic trading not as big as many think. From: http://seekingalpha.com/article/158962risk-from-high-frequency-and-algorithmic-trading-not-as-big-asmany-think, 2009. [12] A. GraÀund, B. Nilsson, Dynamic portfolio selection: The relevance of switching regimes and investment horizon, European Financial Management, vol. 9(2), pp. 179–200, 2003. [13] I. Buckley, D. Saunders and L. Seco. Portfolio optimization when asset returns have the Gaussian mixture distribution. European Journal of Operational Research, vol. 185, pp. 1434–1461, 2008. [14] S. Raudys and V. Pikelis. On dimensionality, sample size, classification error and complexity of classification agorithm in pattern recognition. IEEE Trans. on Pattern Analysis and Machine Intelligence, vol. PAMI-2 (3), pp. 242-252, 1980. [15] I. Guyon and A. Elisseeff. An introduction to variable and feature selection. J. Machine Learning Research, vol. 3, pp. 1157–1182, 2003. [16] M.A.H. Dempster and C.M. Jones A real-time adaptive trading system using genetic programming, Quantitative Finance vol. 1, pp. 397-413, 2001. [17] A. Raudys. Accuracy of MLP based data visualization used in oil prices forecasting task Lect. Notes in Computer Science, Springer, vol. 3617, pp. 761 – 769, 2005. [18] S. Raudys, I. Zliobaite. Prediction of commodity prices in rapidly changing environments Lecture Notes in Computer Science, SpringerVerlag, Berlin Heidelberg, vol. 3686, pp. 154 – 163, 2005. [19] S. Raudys, A. Mitasiunas. Multi-agent system approach to react to sudden environmental changes, Lect. Notes in Artificial Intelligence, Springer-Verlag, Berlin Heidelberg, vol. 4571, pp. 810–823, 2007. [20] S. Raudys, Statistical and Neural Classifiers: An integrated approach to design. Springer, NY, 2001. [21] O. Ledoit and M. Wolf. Improved estimation of the covariance matrix of stock returns with an application to portfolio selection. J. Empir. Finance, pp. 10, pp. 603–621, 2003. [22] O. Ledoit and M. Wolf. A well-conditioned estimator for largedimensional covariance matrices. J. Multivariate Analysis, vol. 88 (2), pp. 365–411, 2004. [23] S. Raudys and D.Young. Results in statistical discriminant analysis: A review of the former Soviet Union literature. J. Multivariate Analysis, vol. 89, pp. 1-35, 2004. [24] Z. Bai, H. Liu and W.K. Wong. On the Markowitz mean–variance analysis of self-¿nancing portfolios. Risk and Decision Analysis, vol. 1, pp. 35–42, 2009. [25] C.K. Chow and C.N. Liu. Approximating discrete probability distributions with dependence trees, IEEE Trans. Information Theory, vol. 14, pp. 462-467, 1968. [26] S. Raudys and A. Saudargiene. First-order tree-type dependence between variables and classification performance. IEEE Trans. on Pattern Analysis and Machine Intelligence, vol. 23: pp. 1324-1328, 2001. [27] E. Fama. Effcient capital markets: A review of theory and empirical work. Journal of Finance. vol. 25(2), pp. 383–417, 1970. [28] B.G. Malkiel. The efficient market hypothesis and its critics. J. of Economi Perspectives, vol. 17 (1), pp. 59-82, 2003. [29] J. Mockus and A. Raudys. On the effcient-market hypothesis and stock exchange game model. Expert Systems with Appl., vol. 37(8), pp. 5673-5681, 2010

2011 11th International Conference on Intelligent Systems Design and Applications

701