Abstract
The MM and EM algorithms are hardly the only methods of optimization. Newton’s method is better known and more widely applied. Despite its defects, Newton’s method is the gold standard for speed of convergence and forms the basis of most modern optimization algorithms. Its many variants seek to retain its fast convergence while taming its defects. They all revolve around the core idea of locally approximating the objective function by a strictly convex quadratic function. At each iteration the quadratic approximation is optimized. Safeguards are introduced to keep the iterates from veering toward irrelevant stationary points.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2004 Springer Science+Business Media New York
About this chapter
Cite this chapter
Lange, K. (2004). Newton’s Method. In: Optimization. Springer Texts in Statistics. Springer, New York, NY. https://doi.org/10.1007/978-1-4757-4182-7_8
Download citation
DOI: https://doi.org/10.1007/978-1-4757-4182-7_8
Publisher Name: Springer, New York, NY
Print ISBN: 978-1-4419-1910-6
Online ISBN: 978-1-4757-4182-7
eBook Packages: Springer Book Archive