Abstract
In this paper, a small and necessary revision on an assumption condition of Aminifard and Babaie-Kafaki (Calcolo, 2019. https://doi.org/10.1007/s10092-019-0312-9) is made. By a little modification, a new conjugate gradient method is proposed, in which the search directions satisfy the sufficient descent condition with the strong Wolfe line search. The main difference between two algorithms is that the proposed method is globally convergent without boundedness assumption on the steplength. Comparative numerical results demonstrating efficiency of the proposed method are reported.
Similar content being viewed by others
Notes
If Powell’s restart criterion \(g_{k}^Tg_{k-1}\ge 0.2 ||g_k||^2\) is satisfied, then the search direction is calculated as \(d_k=-g_k\).
References
Fletcher, R., Reeves, C.: Function minimization by conjugate gradients. Comput. J. 7, 149–154 (1964)
Hestenes, M.R., Stiefel, E.: Methods of conjugate gradients for solving linear systems. J. Res. Natl. Bur. Stand. 49, 409–436 (1952)
Polyak, B.T.: The conjugate gradient method in extremal problems. USSR Comput. Math. Math. Phys. 9, 94–112 (1969)
Gilbert, J.C., Nocedal, J.: Global convergence properties of conjugate gradient methods for optimization. SIAM J. Optim. 2, 21–42 (1992)
Dai, Y., Yuan, Y.: A nonlinear conjugate gradient method with a strong global convergence property. SIAM J. Optim. 10, 177–182 (1999)
Hager, W.W., Zhang, H.: A survey of nonlinear conjugate gradient methods. Pac. J. Optim. 2, 35–58 (2006)
Al-Baali, M.: Descent property and global convergence of the Fletcher–Reeves method with inexact line search. IMA J. Numer. Anal. 5(1), 121–124 (1985)
Hager, W.W., Zhang, H.: A new conjugate gradient method with guaranteed descent and an efficient line search. SIAM J. Optim. 16, 170–192 (2005)
Narushima, Y., Yabe, H., Ford, J.A.: A three-term conjugate gradient method with sufficient descent property for unconstrained optimization. SIAM J. Optim. 21, 212–230 (2011)
Dai, Y.H., Kou, C.X.: A nonlinear conjugate gradient algorithm with an optimal property and an improved Wolfe line search. SIAM J. Optim. 23, 296–320 (2013)
Zhang, L., Zhou, W., Li, D.: A descent modified Polak–Ribi\(\grave{e}\)re–Polyak conjugate gradient method and its global convergence. IMA J. Numer. Anal. 26, 629–640 (2006)
Andrei, N.: A simple three-term conjugate gradient algorithm for unconstrained optimization. J. Comput. Appl. Math. 241, 19–29 (2013)
Cheng, W.: A two-term PRP-based descent method. Numer. Funct. Anal. Opt. 28(11–12), 1217–1230 (2007)
Babaie-Kafaki, S., Reza, G.: A descent extension of the Polak–Ribi\(\grave{e}\)re–Polyak conjugate gradient method. Comput. Math. Appl. 68(12), 2005–2010 (2014)
Yuan, G., Wei, Z., Yang, Y.: The global convergence of the Polak–Ribi\(\grave{e}\)re–Polyak conjugate gradient algorithm under inexact line search for nonconvex functions. J. Comput. Appl. Math. 281, 239–249 (2015)
Yao, S., Feng, Q., Li, L., Xu, J.: A class of globally convergent three-term Dai–Liao conjugate gradient methods. Appl. Numer. Math. 151, 354–366 (2020)
Dong, X., Liu, H., He, Y., Yang, X.: A modified Hestenes–Stiefel conjugate gradient method with sufficient descent condition and conjugacy condition. J. Comput. Appl. Math. 281, 239–249 (2015)
Aminifard, Z., Babaie-Kafaki, S.: A modified descent Polak–Ribi\(\grave{e}\)re–Polyak conjugate gradient method with global convergence property for nonconvex functions. Calcolo (2019). https://doi.org/10.1007/s10092-019-0312-9
Dong, X., Han, D., Dai, Z., Li, X., Zhu, J.: An accelerated three-term conjugate gradient method with sufficient descent condition and conjugacy condition. J. Optim. Theory Appl. 179(3), 944–961 (2018)
Wolfe, P.: Convergence conditions for ascent methods. SIAM Rev. 11, 226–235 (1969)
Andrei, N.: An unconstrained optimization test functions collection. Adv. Model. Optim. 10, 147–161 (2008)
Dolan, E.D., Moré, J.J.: Benchmarking optimization software with performance profiles. Math. Program. 91(2), 201–213 (2002)
Acknowledgements
This work is supported by National Natural Science Foundation of China (11601012), National Science Fund for Distinguished Young Scholars (11625105). We would like to thank Professors N. Andrei for his THREECG code for numerical comparison.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
The work of the first author was done while he was visiting Xi’an Shiyou University.
Rights and permissions
About this article
Cite this article
Dong, X. A modified nonlinear Polak–Ribière–Polyak conjugate gradient method with sufficient descent property. Calcolo 57, 30 (2020). https://doi.org/10.1007/s10092-020-00378-2
Received:
Revised:
Accepted:
Published:
DOI: https://doi.org/10.1007/s10092-020-00378-2
Keywords
- Polak–Ribière–Polyak conjugate gradient method
- Sufficient descent condition
- Global convergence
- Numerical comparison