Skip to main content

Automatic Differentiation: Calculation of Newton Steps

  • Reference work entry
Encyclopedia of Optimization
  • 218 Accesses

Many algorithms for solving optimization problems require the minimization of a merit function, which may be the original objective function, or the solution to sets of simultaneous nonlinear equations which may involve the constraints in the problem. To obtain second order convergence near the solution algorithms to solve both rely on the calculation of Newton steps.

When solving a set of nonlinear equations

the Newton step d at x (0), x ∈ R n, is obtained by solving the linear set of equations

where both the derivatives ∂s j /∂x i and the vector function s j are evaluated at a point, which we will denote by x (0).

For convenience we introduce the Jacobian matrix J and write the equation as

When minimizing a function f(x) the Newton equation becomes

where all the derivatives are calculated at a point, again denoted by x (0).

In terms of the Hessian, H, and the gradient, g, this can be written

Automatic differentiationcan be used to calculate the gradient, Hessian and Jacobian, but...

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 1,699.00
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Berz, M., Bischof, Ch., Corliss, G., and Griewank, A. (eds.): Computational differentiation: techniques, applications, and tools, SIAM, 1996.

    Google Scholar 

  2. Christianson, D.B., and Dixon, L.C.W.: ‘Reverse accumulation of Jacobians in optimal control’, Techn. Report Numer. Optim. Centre, School Inform. Sci. Univ. Hertfordshire267 (1992).

    Google Scholar 

  3. Coleman, T.F., and Cai, J.Y.: ‘The cyclic coloring problem and estimation of sparse Hessian matrices’, SIAM J. Alg. Discrete Meth.7 (1986), 221–235.

    MathSciNet  MATH  Google Scholar 

  4. Coleman, T.F., and Verma, A.: ‘Structure and efficient Jacobian calculation’, in M. Berz, Ch. Bischof, G. Corliss, and A. Griewank (eds.): Computational Differentiation: Techniques, Applications, and Tools, SIAM, 1996, pp. 149–159.

    Google Scholar 

  5. Curtis, A.R., Powell, M.J.D., and Reid, J.K.: ‘On the estimation of sparse Jacobian matrices’, J. Inst. Math. Appl.13 (1974), 117–119.

    MATH  Google Scholar 

  6. Dixon, L.C.W.: ‘Use of automatic differentiation for calculating Hessians and Newton steps’, in A. Griewank and G.F. Corliss (eds.): Automatic Differentiation of Algorithms: Theory, Implementation, and Application, SIAM, 1991, pp. 114–125.

    Google Scholar 

  7. Dixon, L.C.W.: ‘On automatic differentiation and continuous optimization’, in E. Spedicato (ed.): Algorithms for continuous optimisation: the state of the art, ASI series, NATO, 1993, pp. 501–512.

    Google Scholar 

  8. Dixon, L.C.W., and Maany, Z.: ‘The echelon method for the solution of sparse sets of linear equations’, Techn. Report Numer. Optim. Centre, Hatfield PolytechnicNOC TR177 (Feb. 1988).

    Google Scholar 

  9. Duff, I.S., Anoli, N.I., Gould, N.I.M., and Reid, J.K.: ‘The practical use of the Hellerman–Ranck P4 algorithm and the P5 algorithm of Erisman and others’, Techn. Report AERE HarwellCSS213 (1987).

    Google Scholar 

  10. Fletcher, R., and Hall, J.A.J.: ‘Ordering algorithms for irreducible sparse linear systems’, Techn. Report Dundee Univ.NA/131 (1991).

    Google Scholar 

  11. Geitner, U., Utke, J., and Griewank, A.: ‘Automatic computation of sparse Jacobians by applying the method of Newsam and Ramsdell’, in A. Griewank and G.F. Corliss (eds.): Automatic Differentiation of Algorithms: Theory, Implementation, and Application, SIAM, 1991, pp. 161–172.

    Google Scholar 

  12. Griewank, A.: ‘Direct calculation of Newton steps without accumulating Jacobians’, in A. Griewank and G.F. Corliss (eds.): Automatic Differentiation of Algorithms: Theory, Implementation, and Application, SIAM, 1991, pp. 126–137.

    Google Scholar 

  13. Griewank, A., and Corliss, G.F. (eds.): Automatic differentiation of algorithms: theory, implementation, and application, SIAM, 1991.

    Google Scholar 

  14. Griewank, A., and Reese, S.: ‘On the calculation of Jacobian matrices by the Markowitz rule’, in A. Griewank and G.F. Corliss (eds.): Automatic Differentiation of Algorithms: Theory, Implementation, and Application, SIAM, 1991, pp. 126–135.

    Google Scholar 

  15. Parkhurst, S.C.: ‘The evaluation of exact numerical Jacobians using automatic differentiation’, Techn. Report Numer. Optim. Centre, Hatfield Polytechnic, NOC TR224 (Dec. 1990).

    Google Scholar 

  16. Utke, J.: ‘Efficient Newton steps without Jacobians’, in M. Berz, Ch. Bischof, G. Corliss, and A. Griewank (eds.): Computational Differentiation: Techniques, Applications, and Tools, SIAM, 1996, pp. 253–264.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2001 Kluwer Academic Publishers

About this entry

Cite this entry

Dixon, L. (2001). Automatic Differentiation: Calculation of Newton Steps . In: Floudas, C.A., Pardalos, P.M. (eds) Encyclopedia of Optimization. Springer, Boston, MA. https://doi.org/10.1007/0-306-48332-7_16

Download citation

  • DOI: https://doi.org/10.1007/0-306-48332-7_16

  • Publisher Name: Springer, Boston, MA

  • Print ISBN: 978-0-7923-6932-5

  • Online ISBN: 978-0-306-48332-5

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics