Document not found! Please try again

172 ON GLOBAL OPTIMIZATION IN LEAST SQUARES NONLINEAR ...

14 downloads 1632 Views 146KB Size Report
The experiments have been performed on HP Compaq 6710b notebook computer with Intel Centrino processor, 2GB memory and Ubuntu LINUX. The results of ...
ISBN 978-9955-28-598-4 R. Kasımbeyli, C. Dinçer, S. Özpeynirci and L. Sakalauskas (Eds.): MEC EurOPT 2010 Selected papers. Vilnius, 2010, pp. 172–176

International Conference

24th Mini EURO Conference

“Continuous Optimization and Information-Based Technologies in the Financial Sector” (MEC EurOPT 2010) June 23–26, 2010, Izmir, TURKEY

© Izmir University of Economics, Turkey, 2010 © Vilnius Gediminas Technical University, Lithuania, 2010

ON GLOBAL OPTIMIZATION IN LEAST SQUARES NONLINEAR REGRESSION BY MEANS OF INTERVAL METHODS A. Žilinskas, J. Žilinskas

Institute of Mathematics and Informatics Akademijos 4, LT-08663 Vilnius, Lithuania Abstract:The optimization problems occurring in nonlinear regression normally cannot be proven unimodal. In the present paper the applicability to this problem of global optimization algorithms based on the interval arithmetic is investigated. Keywords: global optimization, interval arithmetic, nonlinear regression, nonlinear least squares.

1.

Introduction

2.

General properties of the optimization problems of nonlinear LSR

Among the statistical techniques used in development of economic and financial models least squares regression (LSR) is among the most popular. LSR algorithms are comprised of two main constituents: a minimization algorithm and a procedure of statistical analysis. In this paper we focus on optimization problems of nonlinear LSR. The contemporary statistical packages contain algorithms of nonlinear LSR including efficient subroutines for local minimization of sums of squared residuals. However, in non linear practical problems unimodality of objective functions hardly can be proven, and frequently they are indeed multimodal; see (Winkler, 2001). Therefore investigation of the efficiency of the available global optimization algorithms in nonlinear LSR is an urgent problem. For a general discussion on different global optimization methods and favourable conditions for their applications we refer to (Törn and Žilinskas, 1988). In the next section properties of objective functions of some typical LSR problems are analyzed. Although the general properties of such objective functions seem favourable for the application of algorithms based on interval arithmetic, the latter in LSR are almost not used. In the present paper we discuss the reasons of inefficiency of application of interval methods in nonlinear LSR. Let µ( X , Z ) denotes a regression function where X = (x1,…,xn) is vector of parameters and Z = (z1,…,zp) is vector of variables. The results of observations (measurements) at the points Z1,…,Zm are denoted by yi, i = 1,…,m. The optimization problem in nonlinear LSR is formulated as follows m

(

)

min ∑ yk − µ ( X , Z k ) 2 = min f ( X ),

X ∈A k =1

X ∈A

(1)

where A ∈ R n is a feasible region of parameters. The choice of an appropriate optimization algorithm depends on the properties of f (X) and A. Most frequently µ( X , Z ) is smooth, and A is simple, e.g. A={X: X ≥0}. The minimization problem (1) seems favourable for the application of classical nonlinear programming techniques: normally the number of variables (equal to the number of model parameters to be estimated) is small, and the objective function is smooth. Indeed, many well developed nonlinear programming techniques can be applied to find a local minimizer of (1). A practical problem could be solved easily using algorithms from many available packages if a starting point for local descent would be known in the region of attraction of the global minimizer. However, such a starting point frequently is not known, and (1) should be considered as a global optimization problem. This general theoretical argumentation is supported by the experimental results 172

ON GLOBAL OPTIMIZATION IN LEAST SQUARES NONLINEAR REGRESSION BY MEANS OF INTERVAL METHODS

presented in (Křivý, Tvrdík and Krepec, 2000) where 14 regression function have been considered. It has been shown in (Křivý, Tvrdík, and Krepec 2000) that the standard algorithms from the statistical packages NCSS, SYSTAT, S-PLUS, SPSS for the large percentage of random starting points failed to find the solution. For a general discussion on suitability of different global optimization methods for the minimization of squared residuals in nonlinear LSR we refer to (Žilinskas and Žilinskas, 2010b). Various heuristics can be applied here to find an appropriate solution, however without of justified estimate of precision. If an approximation of the global minimum with guaranteed accuracy is important then the interval arithmetic based global optimization methods seem perspective since they correspond to the mentioned above properties of (1). To test the suitability of the interval arithmetic based global optimization algorithms in nonlinear LSR in (Žilinskas and Žilinskas, 2010a) fourteen difficult test functions were minimized by the well established interval algorithm from the CToolBox described in (Hammer et al., 1995). The data related to these test problems are presented in (Tvrdík, 2010). The global optimization algorithm available in CToolbox is based on branch and bound approach and is implemented in C++. The branching of the search tree is performed by bisecting the promising sub-regions (multidimensional interval boxes) into two through the middle points of the longest sides. Bounds of the function values are estimated using interval arithmetic in standard and cantered forms. Monotonicity and concavity tests, and interval Newton method (see e.g. in (Hansen and Walster, 2004)) are used to discard non-promising sub-regions, where the objective function is monotone or concave, or where there are no stationary points. Intervals of first and second derivatives are evaluated using automatic differentiation. The stopping condition defining tolerance was set equal to 10-8. However, the algorithm stopped according to this stopping condition only in three cases of fourteen; in the other cases the computation was interrupted while exceeding the time limit equal to one hour. These results show that either interval methods or at least their implementation in CToolbox are not appropriate for the difficult cases of nonlinear LSR. In the present paper the results of the similar investigation are presented where the implementation of interval methods by Knüppel (1999) have been applied. 3.

Challenges of application of interval methods in nonlinear LSR

In this paper we consider fourteen problems which have been proven difficult by Křivý, Tvrdík and Krepec (2000); to facilitate the references, the formulas of the regression functions are presented in Table 1. Further references concerning these problems can be found in (Křivý, Tvrdík and Krepec, 2000).

Table 1. Test data and minimum points No.

Regression function

1. 2. 3. 4. 5. 6.

x1x3z1/(1 + x1z1 + x2z2) x3(exp(−x1z1) + exp(−x2z2)) x3(exp(−x1z1) + exp(−x2z2)) x1 + x2 exp(x3z) x1 exp(x2/(x3 + z)) exp(x1z) + exp(x2z)

11.

x1 exp(x3z) + x2 exp(x4z)

12. 13.

x1z (x2) + x3 (x2 /z ) x1 + x2z1(x3) + x4z2(x5) + x6z3(x7) x1 ln(x2 + x3z)

7. 8. 9. 10.

14.

x1 exp(x3z) + x2 exp(x4z) x1z x3 + x2z x4 exp(x1z) + exp(x2z) x1 + x2 exp((x3 + x4z)x5)

Feasible region

[0,1000], [0,1000], [0,1000] [0,1000], [0,1000],[0,10000] [0,2000], [0,100],[0,1000] [0,1000], [0,10], [0,2] [0,10000], [0,50000], [0,10000] [0,1], [0,1]

[0,1e+8], [0,1e+8], [−2,0], [−5,0] [0,1], [1,8], [1,5], [0,1] [−50,50], [−50,50] [0,100], [0,100], [0,5]([1,6]), [0,1], [0,1] [−10000,10000], [−10000,10000], [−1,1], [−1,1] [0,5], [0,5], [0,5]([1,6]) [−5,5], [−5,5], [0,10], [−5,5], [0,10], [−5,5], [0,10] [0,100], [0,100]([1,101]), [0,100]

173

Minimizer

3.1315, 15.159, 0.7801 13.241, 1.5007, 20.100 814.97, 1.5076, 19.920 15.673, 0.9994, 0.02222 0.00561, 6181.4, 345.22 0.2578, 0.2578

1655.2, 3.404e+7, −0.6740, −1.8160 0.004141, 3.8018, 2.0609, 0.2229 0.2807, 0.4064 9.3593, 2.0292, 1.3366, 0.4108, 0.3551 47.971, 102.05, −0.2466, −0.4965

0.05589, 3.5489, 1.4822 1.9295, 2.5784, 0.8017, −1.2987, 0.8990, 0.01915, 3.0184 2.0484, 18.601, 1.8021

A. Žilinskas, J. Žilinskas

We are not going to discuss these functions with respect to their suitability for general testing of global optimization algorithms. The requirements for such functions are discussed, e.g. in (Törn and Žilinskas, 1989; Mathar, Žilinskas, 1994). A representative set of multimodal nonlinear LSR is of great interest, but to our best knowledge such a set has not been yet proposed in the available publications. A representative set of tests should include the applications related problems, e.g. such as considered by Jerrell (1997); and Žilinskas and Bogle (2006). Although the set of problems proposed by Křivý, Tvrdík and Krepec (2000) is not necessary representative, it suits to the purpose of this research since it comprises difficult problems, and was used by Žilinskas and Žilinskas (2010a) to test the other implementation of interval methods. First and most important challenge which is obvious from the formula (1) is dependence of variables. Every summand in (1) contains the expression including the minimization variables (parameters to estimate) and data; in such a case the new occurrence of the same interval variable in interval operations is count as a new one severely increasing the interval of the result. The other challenge is specific for the problems of interest: for the majority of the considered problems the feasible region is huge. Seemingly the problems are essentially unconstrained, and the interval constraints are added only to enable testing of methods not applicable for unconstrained problems. 4.

Experimental results

For this investigation the implementation of interval global optimization algorithms by Knüppel (1999) was used. This implementation of the interval global optimization method is combined with searches implemented in real number arithmetic. The algorithm involving a local search, branch-andbound technique and interval arithmetic, and not requiring derivatives, was originally proposed in (Jansson and Knüppel, 1995). Numerical results for well-known problems and comparisons with other methods are given in (Jansson and Knüppel, 1995). The algorithm uses interval library PROFIL/BIAS by Knüppel (1994). In cases of arithmetic exceptions (e.g. division by the interval containing zero) this library allows to set the result as [−∞,+∞] without the program termination. However, extended interval arithmetic is not implemented in this library, and the algorithm hangs if the infinite interval is attempted to use in further computations. Therefore as in the case of CToolbox algorithm, the feasible regions of the 10th, 12th and 14th test problems had to be changed to the shown in brackets in Table 1. Objective functions of 10th and 12th test problems include power function which is not available in PROFIL/BIAS, and was implemented using the exponential and logarithmic functions. The real and interval functions values should be returned by a subroutine of the objective function. The optimization algorithm counts real (NFE) and interval (NIFE) function calls; it was slightly supplemented to measure the CPU time in seconds. The experiments have been performed on HP Compaq 6710b notebook computer with Intel Centrino processor, 2GB memory and Ubuntu LINUX. The results of optimization of the test problems using PROFIL/BIAS interval global optimization algorithm is shown in Table 2 where also the similar results for CToolbox (from Zilinskas and Zilinskas, 2010a) are presented for comparison. The termination of the algorithm by reaching the time limit does not guarantee that the global minimum is found. For the 2nd test problem algorithm produces strange results founding minimum 0 at point where values of parameters of the model are ‘Not a Numbers’. However the smallest found values of the objective functions and the corresponding minimizers are found for more than half of the test problems. Comparison of the minimization results by CToolbox and PROFIL/BIAS can be summarized as follows: the correct guaranteed answers are found by the former algorithm only in three cases, and by the latter algorithm in nine cases but without guarantee. As shown by Žilinskas and Žilinskas (2010a) CToolbox can successfully found the solution of several reformulated problem where linear parameters in (1) are replaced by their estimates (minimizers of (1) with respect to these variables); in this way the problems 1, 2, and 12 were successfully solved. However, reformulation of the problems has not helped to solve more problems by PROFIL/BIAS. Interval methods can be useful to verify the solutions found by other methods. Since the solutions (global minima) of all test problems are available, both algorithms have been tested in verifying these solutions. However, both algorithms have successfully verified the solutions only for the same problems which have been solved successfully.

174

ON GLOBAL OPTIMIZATION IN LEAST SQUARES NONLINEAR REGRESSION BY MEANS OF INTERVAL METHODS

Table 2. Comparison of testing results of CToolbox and PROFIL/BIAS Test No. 1.

Minimization results by Ctoolbox NFE

674258

Minimum

NFE

NIFE

Minimum

Not found

1468396

219526289

4.35527e−05

Not found

67523

10662377

1.25189

2.

1412003

Not found

4.

1150906

Not found

624496

65068262

0.0059862

124.362

603541

37755042

124.362

3. 5.

1038694 721521

6.

661

588071

Not found

9.

13570

623123

604961

10.

540084

12.

17454

11. 13. 14.

569687

1535228

83596

Not found

7. 8.

5.

Minimization results by PROFIL/BIAS

464100

2358916

55556193

0

Not found

86881

22652296

Not found

0.0088963

621164

40997835

0.0088963

Not found

219852

35410339

Not found

Not found

63620

Not found

23727

28371929

26893387

Not found

Not found

0.00437528

119219

19959291

0.00437528

Not found

922365

58388644

7.14651e−05

Not found

4306

7229973

0.0169413

Conclusion

The experiments with the hybrid algorithm PROFIL/BIAS gave more optimistic results than the recent experiments with CToolbox: more problems have been solved successfully. However, PROFIL/BIAS is a hybrid algorithm voiding the most important advantage of interval methods – the guaranteed accuracy. While the results of experiments with CToolbox were disappointing, the experiments with PROFIL/BIAS show that the application of interval methods for the problems of nonlinear LSR is not hopeless. Since the possibilities to reduce dependence of variables are very limited the development of pure interval methods oriented to nonlinear LSR does not seem promising. On the other hand, waiving the ambitious aim to find guaranteed solutions, the construction of a hybrid algorithm favourably competing with simple heuristics seems possible. The results of the experiments presented above suggest combining interval methods, branch and bound approach, and LSR problem oriented local search methods. Acknowledgments

The research is partially supported by the Agency for International Science and Technology Development Programmes in Lithuania through COST programme and The Research Council of Lithuania. References

Hammer, R.; Hocks, M.; Kulish, U.; Ratz, D. 1995. C++ Toolbox for Verified Computing: Basic Numerical Problems. Springer. Hansen, E.; Walster, G. W. 2004. Global Optimization Using Interval Analysis. 2nd ed. Dekker, New York. Jansson, C.; Knüppel, O. 1995. A branch-and-bound algorithm for bound constrained optimization problems without derivatives, Journal of Global Optimization 7(3): 297–331. doi:10.1007/BF01279453 Jerrell, M. 1997. Automatic differentiation and interval arithmetic for estimation of disequilibrium models, Computational Economics 10: 295–316. doi:10.1023/A:1008633613243 Knüppel, O. 1994. PROFIL/BIAS – A fast interval library, Computing 53(3–4): 277–287. doi:10.1007/BF02307379 Knüppel, O. 1999. PROFIL/BIAS V 2.0. Report 99.1, Technische Universität Hamburg-Harburg. Křivý, I. Tvrdík, J.; Krepec, R. 2000. Stochastic algorithms in nonlinear regression, Computational Statistics and Data Analysis 33: 277–290. doi:10.1016/S0167-9473(99)00059-6 Mathar, R.; Žilinskas, A. 1994. A class of test functions for global optimization, Journal of Global Optimization 5: 195–199. doi:10.1007/BF01100693

175

A. Žilinskas, J. Žilinskas

Törn, A.; Žilinskas, A. 1989. Global optimization, Lecture Notes in Computer Science 350: 1–255. Winkler, P. 2001. Optimization Heuristics in Econometrics, John Wiley & Sons, Chichester. Tvrdík, J. 2010. Global Optimization, Evolutionary Algorithms and Their Application to Computational Statistics [last visited 2010-02-04]. Available from Internet: . Žilinskas, J.; Bogle, I. D. L. 2006. Balanced random interval arithmetic in market model estimation, European Journal of Operational Research 175(3): 1367–1378. doi:10.1016/j.ejor.2005.02.013 Žilinskas, A.; Žilinskas, J. 2010a. Interval arithmetic based optimization in nonlinear regression, Informatica 21(1): 149–158. Žilinskas, A.; Žilinskas, J. 2010b. On probabilistic bounds inspired by interval arithmetic, Control and Cybernetics 39, (in print).

176

Suggest Documents