Abstract
Our discussion of Newton’s method has highlighted both its strengths and its weaknesses. Related algorithms such as scoring and Gauss-Newton exploit special features of the objective function \(f(\boldsymbol{x})\) in overcoming the defects of Newton’s method. We now consider algorithms that apply to generic functions \(f(\boldsymbol{x})\). These algorithms also operate by locally approximating \(f(\boldsymbol{x})\) by a strictly convex quadratic function. Indeed, the guiding philosophy behind many modern optimization algorithms is to see what techniques work well with quadratic functions and then to modify the best techniques to accommodate generic functions.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Acton FS (1990) Numerical methods that work. Mathematical Association of America, Washington, DC
Beltrami EJ (1970) An algorithmic approach to nonlinear analysis and optimization. Academic, New York
Byrd RH, Nocedal J (1989) A tool for the analysis of quasi-Newton methods with application to unconstrained minimization. SIAM J Numer Anal 26:727–739
Ciarlet PG (1989) Introduction to numerical linear algebra and optimization. Cambridge University Press, Cambridge
Conn AR, Gould NIM, Toint PL (1991) Convergence of quasi-Newton matrices generated by the symmetric rank one update. Math Program 50:177–195
Davidon WC (1959) Variable metric methods for minimization. AEC Research and Development Report ANL–5990, Argonne National Laboratory, Argonne
Dennis JE Jr, Schnabel RB (1996) Numerical methods for unconstrained optimization and nonlinear equations. SIAM, Philadelphia
Fiacco AV, McCormick GP (1968) Nonlinear programming: sequential unconstrained minimization techniques. Wiley, Hoboken
Fletcher R (2000) Practical methods of optimization, 2nd edn. Wiley, Hoboken
Fletcher R, Powell MJD (1963) A rapidly convergent descent method for minimization. Comput J 6:163–168
Fletcher R, Reeves CM (1964) Function minimization by conjugate gradients. Comput J 7:149–154
Golub GH, Van Loan CF (1996) Matrix computations, 3rd edn. Johns Hopkins University Press, Baltimore
Henrici P (1982) Essentials of numerical analysis with pocket calculator demonstrations. Wiley, Hoboken
Hestenes MR, Stiefel E (1952) Methods of conjugate gradients for solving linear systems. J Res Natl Bur Stand 29:409–439
Kelley CT (1999) Iterative methods for optimization. SIAM, Philadelphia
Khalfan HF, Byrd RH, Schnabel RB (1993) A theoretical and experimental study of the symmetric rank-one update. SIAM J Optim 3:1–24
Miller KS (1987) Some eclectic matrix theory. Robert E Krieger Publishing, Malabar
Moré JJ, Sorensen DC (1983) Computing a trust region step. SIAM J Sci Stat Comput 4:553–572
Nazareth L (1979) A relationship between the BFGS and conjugate gradient algorithms and its implications for new algorithms. SIAM J Numer Anal 16:794–800
Nocedal J, Wright S (2006) Numerical optimization, 2nd edn. Springer, New York
Press WH, Teukolsky SA, Vetterling WT, Flannery BP (1992) Numerical recipes in Fortran: the art of scientific computing, 2nd edn. Cambridge University Press, Cambridge
Sorensen DC (1997) Minimization of a large-scale quadratic function subject to spherical constraints. SIAM J Optim 7:141–161
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2013 Springer Science+Business Media New York
About this chapter
Cite this chapter
Lange, K. (2013). Conjugate Gradient and Quasi-Newton. In: Optimization. Springer Texts in Statistics, vol 95. Springer, New York, NY. https://doi.org/10.1007/978-1-4614-5838-8_11
Download citation
DOI: https://doi.org/10.1007/978-1-4614-5838-8_11
Published:
Publisher Name: Springer, New York, NY
Print ISBN: 978-1-4614-5837-1
Online ISBN: 978-1-4614-5838-8
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)