Skip to main content

Some Properties of A New Conjugate Gradient Method

  • Chapter
Advances in Nonlinear Programming

Part of the book series: Applied Optimization ((APOP,volume 14))

Abstract

It is proved that the new conjugate gradient method proposed by Dai and Yuan [5] produces a descent direction at each iteration for strictly convex problems. Consequently, the global convergence of the method can be established if the Goldstein line search is used. Further, if the function is uniformly convex, two Armijo-type line searches, the first of which is the standard Armijo line search, are also shown to guarantee the convergence of the new method.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 109.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. L. Armijo, Minimization of functions having Lipschitz continuous first partial derivatives, Pacific Journal of Mathematics, 16 (1) (1966), pp. 1–3

    MathSciNet  MATH  Google Scholar 

  2. M. Al-Baali, Descent property and global convergence of the Fletcher- Reeves method with inexact line search, IMA J. Numer. Anal. 5 (1985), pp. 121–124

    Article  MathSciNet  MATH  Google Scholar 

  3. RH. Byrd and J. Nocedal, A Tool for the Analysis of Quasi-Newton Methods with Application To Unconstrained Minimization, SIAM J. Numer. Anal. 26 (3), (1989), pp. 727–739

    Article  MathSciNet  MATH  Google Scholar 

  4. YH. Dai, Analyses of Nonlinear Conjugate Gradient Methods, Ph. D. thesis, Institute of Computational Mathematics and Scientific/Engineering Computing, Chinese Academy of Sciences, 1997. (in Chinese)

    Google Scholar 

  5. YH. Dai and Y. Yuan, A Nonlinear conjugate Gradient Method with Nice Global Convergence Properties, Research report ICM-95-038, Institute of Computational Mathematics and Scientific/Engineering Computing, Chinese Academy of Sciences, 1995.

    Google Scholar 

  6. YH. Dai and Y. Yuan, Further studies on the Polak-Ribière-Polyak method, Research report ICM-95-040, Institute of Computational Mathematics and Scientific/Engineering Computing, Chinese Academy of Sciences, 1995.

    Google Scholar 

  7. YH. Dai and Y. Yuan, Convergence properties of the Fletcher-Reeves method, IMA J. Numer. Anal. Vol. 16 No. 2 (1996), pp. 155–164

    Article  MathSciNet  MATH  Google Scholar 

  8. YH. Dai and Y. Yuan, Convergence Properties of the Conjugate Descent Method, Mathematical Advances, No. 6 (1996), pp. 552–562

    MathSciNet  Google Scholar 

  9. R. Fletcher (1987), Practical Methods of Optimization vol. 1: Unconstrained Optimization, John Wiley & Sons (New York)

    Google Scholar 

  10. R. Fletcher and C. Reeves, Function minimization by conjugate gradients, Comput. J. 7 (1964), pp. 149–154

    Article  MathSciNet  MATH  Google Scholar 

  11. JC. Gilbert and J. Nocedal, Global convergence properties of conjugate gradient methods for optimization, SIAM J. Optimization. Vol. 2 No. 1 (1992), pp. 21–42.

    Article  MathSciNet  MATH  Google Scholar 

  12. AA. Goldstein, On steepest descent, SIAM J. on Control A, Vol. 3, No. 1, (1965), pp. 147–151

    Google Scholar 

  13. L. Grippo, F. Lampariello, and S. Lucidi, Global convergence and stabilization of unconstrained minimization methods without derivatives, Journal of Optimization Theory and Applications 56 (1988) pp. 385–406

    Article  MathSciNet  MATH  Google Scholar 

  14. L. Grippo and S. Lucidi, A globally convergent version of the Polak-Ribière conjugate gradient method, Math. Prog. 78 (1997), pp. 375–391

    MathSciNet  MATH  Google Scholar 

  15. MR. Hestenes and E. L. Stiefel, Methods of conjugate gradients for solving linear systems, J. Res. Nat. Bur. Standards Sect. 5, 49 (1952), 409–436

    MathSciNet  Google Scholar 

  16. JC. Lagarias, JA. Reeds, MH. Wright and PE. Wright, Convergence Properties of the Nelder-Mead Simplex Algorithm in Low Dimensions, Technical Report 96-4-07, Computing Sciences Research Center, Bell Laboratories, 1996

    Google Scholar 

  17. R. De Leone, M. Gaudioso, and L. Grippo, Stopping criteria for linesearch methods without derivatives, Mathematical Programming 30 (1984) 285–300

    Article  MathSciNet  MATH  Google Scholar 

  18. GH. Liu, JY. Han and HX. Yin, Global convergence of the Fletcher- Reeves algorithm with an inexact line search, Report, Institute of Applied Mathematics, Chinese Academy of Sciences, 1993.

    Google Scholar 

  19. JJ. Morè, BS. Garbow and KE. Hillstrom, Testing unconstrained optimization software, ACM Transactions on Mathematical Software 7 (1981), pp. 17–41

    Article  MATH  Google Scholar 

  20. JL. Nazareth, A view of conjugate gradient-related algorithms for nonlinear optimization, in: Linear and Nonlinear Conjugate Gradient-Related Methods, L. Adams and J. L. Nazareth, eds., SIAM, Philadelphia, 1996, pp. 149–163

    Google Scholar 

  21. J. Nocedal, Theory of Algorithm for Unconstrained Optimization, Acta Numerica (1991), pp. 199–242.

    Google Scholar 

  22. MJD. Powell, Nonconvex minimization calculations and the conjugate gradient method, in: Lecture Notes in Mathematics vol. 1066, Springer- Verlag (Berlin) (1984), pp. 122–141

    Google Scholar 

  23. E. Polak and G. Ribière, Note sur la convergence de directions conjugées, Rev. Francaise Informat Recherche Opertionelle, 3e Année 16 (1969), pp. 35–43

    Google Scholar 

  24. BT. Polyak, The conjugate gradient method in extremem problems, USSR Comp. Math, and Math. Phys. 9 (1969), pp. 94–112

    Article  Google Scholar 

  25. J. Werner, Über die globale konvergenz von Variable-Metric Verfahren mith nichtexakter Schrittweitenbestimmung, Numer. Math. 31 (1978), pp. 321–334

    Article  MathSciNet  MATH  Google Scholar 

  26. P. Wolfe, Convergence conditions for ascent methods, SIAM Review 11 (1969), pp. 226–235

    Article  MathSciNet  MATH  Google Scholar 

  27. G. Zoutendijk, Nonlinear Programming, Computational Methods, in: Integer and Nonlinear Programming (J. Abadie, ed. ), North-Holland (Amsterdam), 1970, pp. 37–86

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 1998 Kluwer Academic Publishers

About this chapter

Cite this chapter

Dai, Y.H., Yuan, Y. (1998). Some Properties of A New Conjugate Gradient Method. In: Yuan, Yx. (eds) Advances in Nonlinear Programming. Applied Optimization, vol 14. Springer, Boston, MA. https://doi.org/10.1007/978-1-4613-3335-7_11

Download citation

  • DOI: https://doi.org/10.1007/978-1-4613-3335-7_11

  • Publisher Name: Springer, Boston, MA

  • Print ISBN: 978-1-4613-3337-1

  • Online ISBN: 978-1-4613-3335-7

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics