A Direct Search Optimization Method That Models the Objective and Constraint Functions by Linear Interpolation

Part of the Mathematics and Its Applications book series (MAIA, volume 275)


An iterative algorithm is proposed for nonlinearly constrained optimization calculations when there are no derivatives. Each iteration forms linear approximations to the objective and constraint functions by interpolation at the vertices of a simplex and a trust region bound restricts each change to the variables. Thus a new vector of variables is calculated, which may replace one of the current vertices, either to improve the shape of the simplex or because it is the best vector that has been found so far, according to a merit function that gives attention to the greatest constraint violation. The trust region radius ρ is never increased, and it is reduced when the approximations of a well-conditioned simplex fail to yield an improvement to the variables, until ρ reaches a prescribed value that controls the final accuracy. Some convergence properties and several numerical results are given, but there are no more than 9 variables in these calculations because linear approximations can be highly inefficient. Nevertheless, the algorithm is easy to use for small numbers of variables.

Key words

Direct search Linear interpolation Nonlinear constraints Optimization without derivatives 


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.


  1. R. Fletcher (1987), Practical Methods of Optimization,John Wiley and Sons (Chichester).Google Scholar
  2. R.E. Griffith and R.A. Stewart (1961), “A nonlinear programming technique for the optimization of continuous processing systems”, Management Sci., Vol. 7, pp. 379–392.MathSciNetzbMATHCrossRefGoogle Scholar
  3. D.M. Himmelblau (1972), Applied Nonlinear Programming, McGraw-Hill (NewYork).Google Scholar
  4. W. Hock and K. Schittkowski (1980), Test Examples for Nonlinear Programming Codes, Lecture Notes in Economics and Mathematical Systems 187,Springer-Verlag (Berlin).Google Scholar
  5. J.A. Nelder and R. Mead (1965), “A simplex method for function minimization”, Comput. J., Vol. 7, pp. 308–313.zbMATHCrossRefGoogle Scholar
  6. M.J.D. Powell (1978), “A fast algorithm for nonlinearly constrained optimization calculations”, in Numerical Analysis, Dundee 1977, Lecture Notes in Mathematics 630, ed. G.A. Watson, Springer-Verlag (Berlin), pp. 144–157.Google Scholar
  7. H.H. Rosenbrock (1960), “An automatic method for finding the greatest or least value of a function”, Comput. J., Vol. 3, pp. 175 – 184.MathSciNetCrossRefGoogle Scholar
  8. W. Spendley, G.R. Hext and F.R. Himsworth (1962), “Sequential application of simplex designs in optimisation and evolutionary operation”, Technometrics, Vol. 4, pp. 441—461.MathSciNetzbMATHCrossRefGoogle Scholar
  9. M.B. Subrahmanyam (1989), “An extension of the simplex method to constrained optimization”, J. Optim. Theory Appl., Vol. 62, pp. 311–319.MathSciNetzbMATHCrossRefGoogle Scholar

Copyright information

© Springer Science+Business Media Dordrecht 1994

Authors and Affiliations

  1. 1.Department of Applied Mathematics and Theoretical PhysicsUniversity of CambridgeCambridgeEngland

Personalised recommendations