Math 5620 - Introduction to Numerical Analysis - Class Notes

Math 5620 - Introduction to Numerical Analysis - Class Notes

Math 5620 - Introduction to Numerical Analysis - Class Notes Fernando Guevara Vasquez Version 1990. Date: January 17, 2012. 3 Contents 1. Disclaimer 4 Chapter 1. Iterative methods for solving linear systems 5 1. Preliminaries 5 2. Neumann series based methods 7 3. Conjugate gradient and other Krylov subspace methods 15 na.tex 3 Rev: 1990, January 17, 2012 4 CHAPTER 0. CONTENTS 1. Disclaimer These notes are work in progress and may be incomplete and contain errors. These notes are based on the following sources: Burden and Faires, Numerical Analysis, ninth ed. ² Kincaid and Cheney, Numerical Analysis: Mathematics of scien- ² tific computing Stoer and Burlisch, Introduction to Numerical Analysis, Springer ² 1992 Golub and Van Loan, Matrix Computations, John Hopkins 1996 ² Rev: 1990, January 17, 2012 4 na.tex CHAPTER 1. CONTENTS 5 CHAPTER 1 Iterative methods for solving linear systems 1. Preliminaries 1.1. Convergence in Rn. Let be a norm defined on Rn, i.e. it sat- k ¢ k isfies the properties: i. x Rn x 0 (non-negativity) 8 2 k k ¸ ii. x 0 x 0 (definiteness) k k Æ , Æ iii. ¸ R, x Rn ¸x ¸ x (multiplication by a scalar) 8 2 2 k k Æ j jk k iv. x,y Rn x y x y (triangle inequality) 8 2 k Å k · k k Å k k Some important examples of norms in Rn are: ¡Pn 2¢1/2 (1) x 2 i 1 xi (Euclidean or `2 norm) k k Æ Pn Æ j j (2) x 1 i 1 xi (`1 norm) k k Æ Æ j j (3) x maxi 1,...,n xi (` or max norm) k k1 Æ¡Pn Æ p ¢1/j pj 1 (4) x p i 1 xi (for p 1, `p norm) k k Æ Æ j j ¸ (k) Rn Rn A sequence {v }k1 1 in converges to v if and only if Æ 2 (1) lim v(k) v . k k ¡ k !1 Since Rn is a finite dimensional vector space the notion of conver- gence is independent of the norm. This follows from the fact that in a finite dimensional vector space all norms are equivalent. Two norms k¢k and are equivalent if there are constants ®,¯ 0 such that kj ¢ jk È (2) x ® x x ¯ x . 8 kj jk · k k · kj jk Another property of Rn is that it is complete, meaning that all Cauchy Rn (k) sequences converge in . A sequence {v }k1 1 is said to be a Cauchy sequence when Æ (3) ² 0 N i, j N v(i) v(j) ². 8 È 9 8 ¸ k ¡ k Ç In English: A Cauchy sequence is a sequence for which any two iterates can be made as close as we want provided that we are far enough in the sequence. na.tex 5 Rev: 1990, January 17, 2012 6 CHAPTER 1. ITERATIVE METHODS FOR SOLVING LINEAR SYSTEMS n n 1.2. Induced matrix norms. Let A R £ be a matrix (this notation 2 means that the matrix has n rows and n columns) and let be a norm k¢¢¢k on Rn. The operator or induced matrix norm of A is defined by: Ax (4) A sup k k sup Ax . k k Æ x Rn ,x 0 x Æ x 1 k k 2 6Æ k k k kÆ The induced matrix norm measures what is the maximum dilation of the image of a vector through the matrix A. It is a good exercise to show that it satisfies the axiom of a norm. Some important examples of induced matrix norms: (1) A 1 maxj 1,...,n cj 1, where cj is the j th column of A. k k Æ Æ k k ¡ (2) A maxi 1,...,n ri 1, where ri is the i th row of A. k k1 Æ Æ k k ¡ (3) A square root of largest eigenvalue of AT A. When A is sym- k k2 Æ metric we have A ¸ , with ¸ being the largest eigen- k k2 Æ j max j max value of A in magnitude. A matrix norm that is not an induced matrix norm is the Frobenius norm: à n !1/2 X 2 (5) A F ai j . k k Æ i,j 1 j j Æ Some properties of induced matrix norms: (1) Ax A x . k k · k kk k (2) AB A B . k k · k kk k (3) Ak A k . k k · k k 1.3. Eigenvalues. The eigenvalues of a n n matrix A are the roots of £ the characteristic polynomial (6) p(¸) det(¸I A). Æ ¡ This is a polynomial of degree n, so it has at most n complex roots. An eigenvector v associated with an eigenvalue ¸ is a nonzero vector such that Av ¸v. Sometimes it is convenient to refer to an eigenvalue ¸ and Æ corresponding eigenvector v 0 as an eigenpair of A. 6Æ The spectral radius ½(A) is defined as the magnitude of the largest eigenvalue in magnitude of a matrix A i.e. (7) ½(A) max{ ¸ det(¸I A) 0}. Æ j j j ¡ Æ The spectral radius is the radius of the smallest circle in C containing all eigenvalues of A. Two matrices A and B are said to be similar if there is an invertible matrix X such that (8) AX XB. Æ Rev: 1990, January 17, 2012 6 na.tex CHAPTER 1. ITERATIVE METHODS FOR SOLVING LINEAR SYSTEMS 7 Similar matrices have the same characteristic polynomial and thus the same eigenvalues. This follows from the properties of the determi- nant since (9) 1 1 det(¸I A) det(X¡ )det(¸I A)det(X) det(¸I X¡ AX) det(¸I B). ¡ Æ ¡ Æ ¡ Æ ¡ 2. Neumann series based methods 2.1. Neumann series. We start by a fundamental theorem that is the theoretical basis for several methods for solving the linear system Ax b. Æ This theorem can be seen as a generalization to matrices of the geometric 1 2 series identity (1 x)¡ 1 x x ... for x 1. ¡ Æ Å Å Å j j Ç n n THEOREM 1. Let A R £ be such that A 1 for some induced ma- 2 k k Ç trix norm , then: k ¢ k i. I A is invertible ¡ 1 2 P k ii. (I A)¡ I A A k1 0 A . ¡ Æ Å Å Å ¢¢¢ Æ Æ PROOF. Assume for contradiction that I A is singular. This means ¡ there is a x 0 such that (I A)x 0. Taking x such that x 1, we have 6Æ ¡ Æ k k Æ (10) 1 x Ax A x A , Æ k k Æ k k · k kk k Æ k k which contradicts the hypothesis A 1. k k Ç 1 We now need to show convergence to (I A)¡ of the partial series ¡ m X (11) Ak . k 0 Æ Observe that: m m X k X k k 1 0 m 1 (12) (I A) A A A Å A A Å . ¡ k 0 Æ k 0 ¡ Æ ¡ Æ Æ Therefore m X k m 1 m 1 (13) (I A) A I A Å A Å 0 as m . k ¡ k 0 ¡ k Æ k k · k k ! ! 1 Æ Here is an application of this theorem to estimate the norm 1 X1 k X1 k 1 (14) (I A)¡ A A . k ¡ k · k 0 k k · k 0 k k Æ 1 A Æ Æ ¡ k k Here is a generalization of the Neumann series theorem. THEOREM 2. If A and B are n n matrices such that I AB 1 for £ k ¡ k Ç some induced matrix norm then i. A and B are invertible. na.tex 7 Rev: 1990, January 17, 2012 8 CHAPTER 1. ITERATIVE METHODS FOR SOLVING LINEAR SYSTEMS ii. The inverses are: 1 X1 k A¡ B (I AB) , Æ k 0 ¡ (15) " Æ # 1 X1 k B¡ (I AB) A. Æ k 0 ¡ Æ PROOF. By using the Neumann series theorem, AB I (I AB) is in- Æ ¡ ¡ vertible and 1 X1 k (16) (AB)¡ (I AB) . Æ k 0 ¡ Æ Therefore 1 1 X1 k A¡ B(AB)¡ B (I AB) , Æ Æ k 0 ¡ (17) " Æ # 1 1 X1 k B¡ (AB)¡ A (I AB) A. Æ Æ k 0 ¡ Æ 2.2. Iterative refinement. Let A be an invertible matrix. The itera- tive refinement method is a method for generating successively better approximations to the solution of the linear system Ax b. Assume we 1 Æ have an invertible matrix B such that x Bb A¡ b and applying B is Æ ¼ much cheaper than applying solving a system with the matrix A (we shall see how good the approximation needs to be later). This approximate in- verse B may come for example from an incomplete LU factorization or from running a few steps of an iterative method to solve Ax b. Can we Æ use successively refine the approximations given by this method? The idea is to look at the iteration x(0) Bb (18) Æ (k) (k 1) (k 1) x x ¡ B(b Ax ¡ ). Æ Å ¡ If this iteration converges, then the limit must satisfy (19) x x B(b Ax), Æ Å ¡ i.e. if the method converges it converges to a solution of Ax b. Æ THEOREM 3. The iterative refinement method (18) generates iterates of the form m X (20) x(m) B (I AB)k b, m 0. Æ k 0 ¡ ¸ Æ Rev: 1990, January 17, 2012 8 na.tex CHAPTER 1.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    15 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us