Abstract.
In the Newton/log-barrier method, Newton steps are taken for the log-barrier function for a fixed value of the barrier parameter until a certain convergence criterion is satisfied. The barrier parameter is then decreased and the Newton process is repeated. A naive analysis indicates that Newton’s method does not exhibit superlinear convergence to the minimizer of each instance of the log-barrier function until it reaches a very small neighborhood, namely within O(μ2) of the minimizer, where μ is the barrier parameter. By analyzing the structure of the barrier Hessian and gradient in terms of the subspace of active constraint gradients and the associated null space, we show that this neighborhood is in fact much larger –O(μσ) for any σ∈(1,2] – thus explaining why reasonably fast local convergence can be attained in practice. Moreover, we show that the overall convergence rate of the Newton/log-barrier algorithm is superlinear in the number of function/derivative evaluations, provided that the nonlinear program is formulated with a linear objective and that the schedule for decreasing the barrier parameter is related in a certain way to the step length and convergence criteria for each Newton process.
Similar content being viewed by others
Author information
Authors and Affiliations
Additional information
Received: October 10, 1997 / Accepted: September 10, 2000¶Published online February 22, 2001
Rights and permissions
About this article
Cite this article
Wright, S. On the convergence of the Newton/log-barrier method. Math. Program. 90, 71–100 (2001). https://doi.org/10.1007/PL00011421
Issue Date:
DOI: https://doi.org/10.1007/PL00011421