Skip to main content
Log in

Two Improved Nonlinear Conjugate Gradient Methods with the Strong Wolfe Line Search

  • Original Paper
  • Published:
Bulletin of the Iranian Mathematical Society Aims and scope Submit manuscript

Abstract

Two improved nonlinear conjugate gradient methods are proposed by using the second inequality of the strong Wolfe line search. Under usual assumptions, we proved that the improved methods possess the sufficient descent property and global convergence. By testing the unconstrained optimization problems which taken from the CUTE library and other usual test collections, some large-scale numerical experiments for the presented methods and their comparisons are executed. The detailed results are listed in tables and their corresponding performance profiles are reported in figures, which show that our improved methods are superior to their comparisons.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6

Similar content being viewed by others

References

  1. Andrei, N.: An unconstrained optimization test functions collection. Adv. Model. Optim. 10(1), 147–161 (2008)

    MathSciNet  MATH  Google Scholar 

  2. Bongartz, I., Conn, A.R., Gould, N., Toint, P.L.: CUTE: constrained and unconstrained testing environment. ACM Trans. Math. Softw. 21(1), 123–160 (1995)

  3. Dai, Y.H., Yuan, Y.X.: A nonlinear conjugate gradient method with a strong global convergence property. SIAM J. Optim. 10(1), 177–182 (1999)

    Article  MathSciNet  Google Scholar 

  4. Dai, Y.H., Yuan, Y.X.: An efficient hybrid conjugate gradient method for unconstrained optimization. Ann. Oper. Res. 103, 33–47 (2001)

    Article  MathSciNet  Google Scholar 

  5. Dolan, E.D., Moré, J.J.: Benchmarking optimization software with performance profiles. Math. Program. 91(2), 201–213 (2002)

    Article  MathSciNet  Google Scholar 

  6. Esmaeili, H., Shabani, S., Kimiaei, M.: A new generalized shrinkage conjugate gradient method for sparse recovery. Calcolo 56(1), 1–38 (2019)

    Article  MathSciNet  Google Scholar 

  7. Fletcher, R., Reeves, C.M.: Function minimization by conjugate gradients. Comput. J. 7(2), 149–154 (1964)

    Article  MathSciNet  Google Scholar 

  8. Glibert, J.C., Nocedal, J.: Global covergence properties of conjugate gradient method for optimization. SIAM J. Optim. 2(1), 21–42 (1992)

    Article  MathSciNet  Google Scholar 

  9. Hager, W.W., Zhang, H.C.: A new conjugate gradient method with guaranteed descent and an efficient line search. SIAM J. Optim. 16(1), 170–192 (2005)

    Article  MathSciNet  Google Scholar 

  10. Hager, W.W., Zhang, H.C.: A survey of nonlinear conjugate gradient methods. Pac. J. Optim. 2(1), 35–58 (2006)

    MathSciNet  MATH  Google Scholar 

  11. Hestenes, M.R., Stiefel, E.: Methods of conjugate gradients for solving linear systems. J. Res. Natl. Bur. Stand. 49(6), 409 (1952)

    Article  MathSciNet  Google Scholar 

  12. Huang, H., Li, Y., Wei, Z.: Global convergence of a modified PRP conjugate gradient method. J. Math. Res. Expos. 30(1), 141–148 (2010)

    MathSciNet  MATH  Google Scholar 

  13. Jiang, X., Jian, J.: Improved Fletcher–Reeves and Dai–Yuan conjugate gradient methods with the strong Wolfe line search. J. Comput. Appl. Math. 348, 525–534 (2019)

    Article  MathSciNet  Google Scholar 

  14. Kou, C., Dai, Y.: A modified self-scaling memoryless Broyden–Fletcher–Goldfarb–Shanno method for unconstrained optimization. J. Optim. Theory Appl. 165(1), 209–224 (2015)

    Article  MathSciNet  Google Scholar 

  15. Li, H.: Solution of inverse blackbody radiation problem with conjugate gradient method. IEEE Trans. Antennas Propag. 53(5), 1840–1842 (2005)

    Article  MathSciNet  Google Scholar 

  16. Liu, Z., Zhu, S., Ge, Y., Shan, F., Zeng, L., Liu, W.: Geometry optimization of two-stage thermoelectric generators using simplified conjugate-gradient method. Appl. Energy 190, 540–552 (2017)

    Article  Google Scholar 

  17. Moré, J.J., Garbow, B.S., Hillstrom, K.E.: Testing unconstrained optimization software. ACM Trans. Math. Softw. 7(1), 17–41 (1981)

    Article  MathSciNet  Google Scholar 

  18. Polak, E., Ribière, G.: Note sur la convergence de méthodes de directions conjuées. Rev. Fr. Inform. Rech. Oper. 3(16), 35–43 (1969)

    MATH  Google Scholar 

  19. Polyak, B...T.: The conjugate gradient method in extremal problems. USSR Comput. Math. Math. Phys. 9(4), 94–112 (1969)

  20. Tang, C., Li, S., Cui, Z.: Least-squares-based three-term conjugate gradient methods. J. Inequal. Appl. (2020). https://doi.org/10.1186/s13660-020-2301-6

    Article  MathSciNet  MATH  Google Scholar 

  21. Yang, K., Jiang, G., Qu, Q., Peng, H., Gao, X.: A new modified conjugate gradient method to identify thermal conductivity of transient non-homogeneous problems based on radial integration boundary element method. Int. J. Heat Mass Transf. 133, 669–676 (2019)

    Article  Google Scholar 

  22. Yang, L.F., Luo, J.Y., Xu, Y., Zhang, Z.R., Dong, Z.Y.: A distributed dual consensus ADMM based on partition for DC-DOPF with carbon emission trading. IEEE Trans. Ind. Inform. 16(3), 1858–1872 (2020)

    Article  Google Scholar 

  23. Yang, L.F., Zhang, C., Jian, J.B., Meng, K., Xu, Y., Dong, Z.Y.: A novel projected two-binary-variable formulation for unit commitment in power systems. Appl. Energy 187, 732–745 (2017)

    Article  Google Scholar 

  24. Zhang, L.: An improved Wei–Yao–Liu nonlinear conjugate gradient method for optimization computation. J. Comput. Appl. Math. 215(6), 2269–2274 (2009)

    Article  MathSciNet  Google Scholar 

  25. Zhou, X., Lu, L.: The global convergence of modified DY conjugate gradient methods under the Wolfe line search. J. Chongqing Normal Univ. (Nat. Sci. Ed.) 33(3), 6–10 (2016)

    MATH  Google Scholar 

  26. Zhu, Z., Zhang, D., Wang, S.: Two modified DY conjugate gradient methods for unconstrained optimization problems. Appl. Math. Comput. 373(15), 125004 (2020)

    MathSciNet  MATH  Google Scholar 

Download references

Acknowledgements

This work was supported by the National Natural Science Foundation of China (11771383), the Natural Science Foundation of Guangxi Province (2020GXNSFDA238017, 2018GXNSFFA281007) and Research Project of Guangxi University for Nationalities (2018KJQ-D02).

Author information

Authors and Affiliations

Authors

Contributions

All authors read and approved the final manuscript. JJ mainly contributed to the algorithm design; PL and XJ mainly contributed to the convergence analysis, numerical results and drafted the manuscript; BH mainly contributed to the convergence analysis.

Corresponding author

Correspondence to Xianzhen Jiang.

Ethics declarations

Conflict of interest

The authors declare that they have no conflict of interest.

Additional information

Communicated by Rohollah Yousefpour.

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Jian, J., Liu, P., Jiang, X. et al. Two Improved Nonlinear Conjugate Gradient Methods with the Strong Wolfe Line Search. Bull. Iran. Math. Soc. 48, 2297–2319 (2022). https://doi.org/10.1007/s41980-021-00647-y

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s41980-021-00647-y

Keywords

Mathematics Subject Classification

Navigation