Skip to main content

Conjugate Gradient and Quasi-Newton

  • Chapter
  • First Online:
Optimization

Part of the book series: Springer Texts in Statistics ((STS,volume 95))

  • 12k Accesses

Abstract

Our discussion of Newton’s method has highlighted both its strengths and its weaknesses. Related algorithms such as scoring and Gauss-Newton exploit special features of the objective function \(f(\boldsymbol{x})\) in overcoming the defects of Newton’s method. We now consider algorithms that apply to generic functions \(f(\boldsymbol{x})\). These algorithms also operate by locally approximating \(f(\boldsymbol{x})\) by a strictly convex quadratic function. Indeed, the guiding philosophy behind many modern optimization algorithms is to see what techniques work well with quadratic functions and then to modify the best techniques to accommodate generic functions.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

eBook
USD 16.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 129.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 179.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Acton FS (1990) Numerical methods that work. Mathematical Association of America, Washington, DC

    MATH  Google Scholar 

  2. Beltrami EJ (1970) An algorithmic approach to nonlinear analysis and optimization. Academic, New York

    MATH  Google Scholar 

  3. Byrd RH, Nocedal J (1989) A tool for the analysis of quasi-Newton methods with application to unconstrained minimization. SIAM J Numer Anal 26:727–739

    Article  MathSciNet  MATH  Google Scholar 

  4. Ciarlet PG (1989) Introduction to numerical linear algebra and optimization. Cambridge University Press, Cambridge

    Google Scholar 

  5. Conn AR, Gould NIM, Toint PL (1991) Convergence of quasi-Newton matrices generated by the symmetric rank one update. Math Program 50:177–195

    Article  MathSciNet  MATH  Google Scholar 

  6. Davidon WC (1959) Variable metric methods for minimization. AEC Research and Development Report ANL–5990, Argonne National Laboratory, Argonne

    Google Scholar 

  7. Dennis JE Jr, Schnabel RB (1996) Numerical methods for unconstrained optimization and nonlinear equations. SIAM, Philadelphia

    Book  MATH  Google Scholar 

  8. Fiacco AV, McCormick GP (1968) Nonlinear programming: sequential unconstrained minimization techniques. Wiley, Hoboken

    MATH  Google Scholar 

  9. Fletcher R (2000) Practical methods of optimization, 2nd edn. Wiley, Hoboken

    Google Scholar 

  10. Fletcher R, Powell MJD (1963) A rapidly convergent descent method for minimization. Comput J 6:163–168

    Article  MathSciNet  MATH  Google Scholar 

  11. Fletcher R, Reeves CM (1964) Function minimization by conjugate gradients. Comput J 7:149–154

    Article  MathSciNet  MATH  Google Scholar 

  12. Golub GH, Van Loan CF (1996) Matrix computations, 3rd edn. Johns Hopkins University Press, Baltimore

    MATH  Google Scholar 

  13. Henrici P (1982) Essentials of numerical analysis with pocket calculator demonstrations. Wiley, Hoboken

    MATH  Google Scholar 

  14. Hestenes MR, Stiefel E (1952) Methods of conjugate gradients for solving linear systems. J Res Natl Bur Stand 29:409–439

    Article  MathSciNet  Google Scholar 

  15. Kelley CT (1999) Iterative methods for optimization. SIAM, Philadelphia

    Book  MATH  Google Scholar 

  16. Khalfan HF, Byrd RH, Schnabel RB (1993) A theoretical and experimental study of the symmetric rank-one update. SIAM J Optim 3:1–24

    Article  MathSciNet  MATH  Google Scholar 

  17. Miller KS (1987) Some eclectic matrix theory. Robert E Krieger Publishing, Malabar

    MATH  Google Scholar 

  18. Moré JJ, Sorensen DC (1983) Computing a trust region step. SIAM J Sci Stat Comput 4:553–572

    Article  MATH  Google Scholar 

  19. Nazareth L (1979) A relationship between the BFGS and conjugate gradient algorithms and its implications for new algorithms. SIAM J Numer Anal 16:794–800

    Article  MathSciNet  MATH  Google Scholar 

  20. Nocedal J, Wright S (2006) Numerical optimization, 2nd edn. Springer, New York

    MATH  Google Scholar 

  21. Press WH, Teukolsky SA, Vetterling WT, Flannery BP (1992) Numerical recipes in Fortran: the art of scientific computing, 2nd edn. Cambridge University Press, Cambridge

    Google Scholar 

  22. Sorensen DC (1997) Minimization of a large-scale quadratic function subject to spherical constraints. SIAM J Optim 7:141–161

    Article  MathSciNet  MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Rights and permissions

Reprints and permissions

Copyright information

© 2013 Springer Science+Business Media New York

About this chapter

Cite this chapter

Lange, K. (2013). Conjugate Gradient and Quasi-Newton. In: Optimization. Springer Texts in Statistics, vol 95. Springer, New York, NY. https://doi.org/10.1007/978-1-4614-5838-8_11

Download citation

Publish with us

Policies and ethics