Numerical experiments on dual matrix algorithms for function minimization
- 37 Downloads
In Ref. 2, four algorithms of dual matrices for function minimization were introduced. These algorithms are characterized by the simultaneous use of two matrices and by the property that the one-dimensional search for the optimal stepsize is not needed for convergence. For a quadratic function, these algorithms lead to the solution in at mostn+1 iterations, wheren is the number of variables in the function. Since the one-dimensional search is not needed, the total number of gradient evaluations for convergence is at mostn+2.
In this paper, the above-mentioned algorithms are tested numerically by using five nonquadratic functions. In order to investigate the effects of the stepsize on the performances of these algorithms, four schemes for the stepsize factor are employed, two corresponding to small-step processes and two corresponding to large-step processes. The numerical results show that, in spite of the wide range employed in the choice of the stepsize factor, all algorithms exhibit satisfactory convergence properties and compare favorably with the corresponding quadratically convergent algorithms using one-dimensional searches for optimal stepsizes.
Key WordsMathematical programming function minimization method of dual matrices computing methods numerical methods
Unable to display preview. Download preview PDF.
- 1.Huang, H. Y., andChambliss, J. P.,Numerical Experiments on Dual Matrix Algorithms for Function Minimization, Rice University, Aero-Astronautics Report No. 89, 1972.Google Scholar
- 2.Huang, H. Y.,Method of Dual Matrices for Function Minimization, Rice University, Aero-Astronautics Report No. 88, 1972.Google Scholar
- 3.Hestenes, M. R., andStiefel, E.,Method of Conjugate Gradients for Solving Linear Systems, Journal of Research of the National Bureau of Standards, Vol. 40, No. 6, 1952.Google Scholar
- 4.Davidon, W. C.,Variable Metric Methods for Minimization, Argonne National Laboratory, Report No. ANL-5990, 1959.Google Scholar
- 5.Fletcher, R., andPowell, M. J. D.,A Rapidly Convergent Descent Method for Minimization, Computer Journal, Vol. 6, No. 2, 1963.Google Scholar
- 6.Fletcher, R., andReeves, C. M.,Function Minimization by Conjugate Gradients, Computer Journal, Vol. 7, No. 2, 1964.Google Scholar
- 7.Broyden, C. G.,Quasi-Newton Methods and Their Application to Function Minimization, Mathematics of Computation, Vol. 21, No. 99, 1967.Google Scholar
- 8.Pearson, J. D.,On Variable Metric Methods of Minimization, Research Analysis Corporation, Technical Report No. RAC-TP-302, 1968.Google Scholar
- 9.Huang, H. Y.,Unified Approach to Quadratically Convergent Algorithms for Function Minimization, Journal of Optimization Theory and Applications, Vol. 5, No. 6, 1970.Google Scholar
- 10.Huang, H. Y., andLevy, A. V.,Numerical Experiments on Quadratically Convergent Algorithms for Function Minimization, Journal of Optimization Theory and Applications, Vol. 6, No. 3, 1970.Google Scholar
- 11.Rosenbrock, H. H.,An Automatic Method for Finding the Greatest or the Least Value of a Function, Computer Journal, Vol. 3, No. 3, 1960.Google Scholar
- 12.Colville, A. R.,A Comparative Study of Nonlinear Programming Codes, IBM New York Scientific Center, TR No. 320–2949, 1968.Google Scholar
- 13.Huang, H. Y., andChambliss, J. P.,Quadratically Convergent Algorithms and One-Dimensional Search Schemes, Journal of Optimization Theory and Applications, Vol. 11, No. 2, 1973.Google Scholar
- 14.Cragg, E. E., andLevy, A. V.,Study on a Supermemory Gradient Method for the Minimization of Functions, Journal of Optimization Theory and Applications, Vol. 4, No. 3, 1969.Google Scholar