The Powell method in its basic form can be viewed as a gradient-free minimization algorithm. It requires repeated line search minimizations, which may be carried out using univariate gradient free , or gradient based procedures . It was introduced by M.J.D. Powell [1]. The procedure is described in the algorithm steps below.
The minimization problem considered is:
1. Initialization
Select an accuracy ε > 0, and a starting point x (0). Set the initial search directions s (i) to be the unit vectors along each coordinate axis, for i = 1,..., n. Set the main iteration counter to k = 0, and the cycle counter i = 1. Initialize z (1) = x (0). Set counter j = 0 for the case where step 2.2a is used.
2. Directional univariate minimization
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Powell, M. J.D.: ‘An efficient method for finding the minimum of a function of several variables without calculating derivatives’, Computer J.7 (1964), 155–162.
Rao, S. S.: Optimization theory and applications, Wiley Eastern second, 1984.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2001 Kluwer Academic Publishers
About this entry
Cite this entry
Vassiliadis, V.S., Conejeros, R. (2001). Powell Method . In: Floudas, C.A., Pardalos, P.M. (eds) Encyclopedia of Optimization. Springer, Boston, MA. https://doi.org/10.1007/0-306-48332-7_393
Download citation
DOI: https://doi.org/10.1007/0-306-48332-7_393
Publisher Name: Springer, Boston, MA
Print ISBN: 978-0-7923-6932-5
Online ISBN: 978-0-306-48332-5
eBook Packages: Springer Book Archive