Abstract
The conjugate gradient method is also aimed at accelerating the methods of Chap. 2. Its first motivation is to solve in n iterations a linear system with symmetric positive definite matrix (or, equivalently, to minimize in n iterations a quadratic strongly convex function on ℝn), without storing an additional matrix, without even storing the matrix of the system. In fact, to solve Ax + b = 0 (A symmetric positive definite), the conjugate gradient method just needs a “black box” (a subroutine) which, given the vector u, computes the vector v = Au. Naturally, this becomes particularly interesting when, while n is large, A is sparse and/or enjoys some structure allowing automatic calculations. Typical examples come from the discretization of partial differential equations.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2003 Springer-Verlag Berlin Heidelberg
About this chapter
Cite this chapter
Bonnans, J.F., Gilbert, J.C., Lemaréchal, C., Sagastizábal, C.A. (2003). Conjugate Gradient. In: Numerical Optimization. Universitext. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-662-05078-1_5
Download citation
DOI: https://doi.org/10.1007/978-3-662-05078-1_5
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-00191-1
Online ISBN: 978-3-662-05078-1
eBook Packages: Springer Book Archive