Abstract
As already seen, the conjugate gradient algorithms presented so far use some principles based on: hybridization or modifications of the standard schemes, the memoryless or the scaled memoryless BFGS preconditioned or the three-term concept. The corresponding conjugate gradient algorithms are defined by the descent condition, the “pure” conjugacy or the Dai–Liao conjugacy conditions or by the minimization of the quadratic approximation with one or two parameters of the objective function. There are a number of convergence results, mainly based on the Zoutendijk and on the Nocedal conditions under the Wolfe line search (Dai, 2011). These algorithms have good numerical performances, being able to solve large-scale unconstrained optimization problems and applications. However, in the frame of conjugate gradient methods, which is a very active area of research, some other computational schemes were introduced in order to improve their numerical performances. They are too numerous to be presented in this study. However, a short description of some of them is as follows.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this chapter
Cite this chapter
Andrei, N. (2020). Other Conjugate Gradient Methods. In: Nonlinear Conjugate Gradient Methods for Unconstrained Optimization. Springer Optimization and Its Applications, vol 158. Springer, Cham. https://doi.org/10.1007/978-3-030-42950-8_11
Download citation
DOI: https://doi.org/10.1007/978-3-030-42950-8_11
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-42949-2
Online ISBN: 978-3-030-42950-8
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)