Some History of Conjugate Gradients and Other Krylov Subspace Methods October 2009 SIAM Applied Linear Algebra Meeting 2009 Dianne P. O'Leary c 2009 1 Some History of Conjugate Gradients Dianne P. O'Leary Computer Science Dept. and Institute for Advanced Computer Studies University of Maryland
[email protected] http://www.cs.umd.edu/users/oleary 2 A Tale of 2 Cities Los Angeles, California Z¨urich,Switzerland 3 ... and a Tale of 2 Men 4 ... or maybe 3 Men 5 ... or maybe 5 Men 6 The Conjugate Gradient Algorithm 7 Notation • We solve the linear system Ax∗ = b where A 2 Rn×n is symmetric positive definite and b 2 Rn. • We assume, without loss of generality, that our initial guess for the solution is x(0) = 0 : • We denote the Krylov subspace of dimension m by m−1 Km(A; b) = spanfb; Ab; : : : ; A bg : (k) • Then the conjugate gradient algorithm chooses x 2 Kk(A; b) to minimize (x − x∗)T A(x − x∗). { work per iteration: 1 matrix-vector multiply and a few dot products and combinations of vectors. { storage: 3 vectors, plus original data. 8 Context for the Conjugate Gradient Algorithm Hestenes and Stiefel (1952) presented the conjugate gradient algorithm in the Journal of Research of the NBS. Context: Two existing classes of algorithms • Direct methods, like Gauss elimination, modified a tableau of matrix entries in a systematic way in order to compute the solution. These methods were finite but required a rather large amount of computational effort, with work growing as the cube of the number of unknowns.