Cholesky Decomposition

Total Page:16

File Type:pdf, Size:1020Kb

Cholesky Decomposition Cholesky Decomposition Martin Licht UC San Diego Winter Quarter 2021 Cholesky Decomposition The class of symmetric positive definite matrices appears in many applications. We will see that for those matrices, the LU decomposition can always be performed without pivoting. Moreover, the LU decomposition of symmetric positive definete matrices leads to the Cholesky decomposition. Symmetric Positive Definite Matrices LU Decomposition of Symmetric Positive Definite Matrices Specialized Algorithms Symmetric Positive Definite Matrices × Let A 2 Rn n be a matrix. We call A symmetric if A = At: We call A positive definite if n 8x 2 R n f0g : hx; Axi > 0: A matrix that is both symmetric and positive definite is called symmetric positive definite. Symmetric Positive Definite Matrices × Suppose that M 2 Rn n is invertible. Let A = MtM. Then A is symmetric because At = (MtM)t = MtM = A Also, A is positive definite: for any x 2 Rn with x 6= 0 we can write y := Mx and observe y 6= 0. Then hx; Axi = hx; MtMxi = hMx; Mxi = hy; yi > 0: It follows that A is symmetric positive definite. Symmetric Positive Definite Matrices Example 01 1 05 1 05 1 01 @ 1 A ; @ 10 A ; @1 5 1A 1 0:17 0 1 5 Symmetric Positive Definite Matrices The previous example is already exhaustive for the class of symmetric positive definite matrices: all spd matrices A can be written in the form A = MtM for some invertible matrix M. In fact, we can achieve that M is upper triangular. Cholesky Decomposition Example 01 1 01 1 1 01 1 1 @1 1 A @ 1 1A = @1 2 1A 1 1 1 1 2 02 1 02 3 11 04 6 21 @3 2 A @ 2 2A = @6 13 5A 1 2 1 1 2 5 6 0p 1 0p 1 2 2 p−1 0 0 1 q 2 2 −1 0 B p−1 3 C B q 3 q 3 C B 2 C B − C = @−1 2 −1A @ 2 q A @ 2 2 A 0 − 3 p2 p2 0 −1 2 2 3 3 Symmetric Positive Definite Matrices LU Decomposition of Symmetric Positive Definite Matrices Specialized Algorithms Cholesky Decomposition × × If A 2 Rn n is invertible and R 2 Rn n is upper triangular with A = RtR, then we call R a Cholesky factor of A. We call A = RtR the Cholesky decomposition. Note that A invertible implies R invertible. Cholesky Decomposition We will show the Cholesky decomposition exists precisely for the class of symmetric positive definite matrices. The strategy will be showing that symmetric positive definite matrices allow for an LU decomposition without pivoting, and then we modify that LU decomposition to get a Cholesky decomposition. Cholesky Decomposition Lemma Suppose that A is symmetric positive definite. Then A satisfies the principle minor criterion. Proof. Let A be symmetric positive definite n × n matrix. We want to show that all principle minors Ak, 1 ≤ k ≤ n are invertible. k×k Suppose that there exists a principle minor Ak 2 R with 1 ≤ k ≤ n that is not invertible. So there exists v 2 Rk with v 6= 0 n but Akv = 0. Let ev 2 R be the extension of v to an n vector. We have ev 6= 0. We get 0 < hev; Aevi = hv; Akvi = hv; 0i = 0: This is a contradiction. So Ak must have been invertible in the first place. Cholesky Decomposition Corollary The LU decomposition without pivoting exists for symmetric definite positive matrices. Cholesky Decomposition Lemma The LU decomposition of a symmetric positive definite matrix has U with positive diagonal entries. Proof. × Let A be s.p.d. and let L; U 2 Rn n be the matrices from its LU decomposition. We want to show that Uii > 0 for all 1 ≤ i ≤ n. Suppose that instead there exists 1 ≤ i ≤ n with Uii ≤ 0. Then Uii = 0 cannot hold because U is invertible, so Uii < 0 must hold. n t Let v 2 R be the solution to L v = ei, and then let w := Uv. We obtain v by backward substitution and find that vi = 1 and that vj = 0 for all j > i. Hence wi = uiivi = uii. t 0 < hv; Avi = hv; LUvi = hL v; Uvi = hei; wi = wi = uii < 0 This is a contradiction. Hence uii > 0 must hold. Cholesky Decomposition Let A = LU be the LU decomposition of a symmetric positive × definite matrix A 2 Rn n. We now define the factors in the Cholesky decomposition.p Let n×n D 2 R be the diagonal matrix with Dii = Uii. We define new n × n matrices R := D−1U; S := LD: We get that A = LU = LDD−1U = SR: The upper triangular matrix R will be the Cholesky factor in our Cholesky decomposition. To verify that, it only remains need to show that St = R. Cholesky Decomposition To see that, we first use the symmetry of A to find SR = A = At = RtSt: Since R and S are invertible, we thus get St = R−tSR: If we can show that R−tS = I, then we get St = R. Cholesky Decomposition Since R is upper triangular, Rt is lower triangular. The inverse of a lower triangular matrix is again lower triangular. Hence R−t is lower triangular. The diagonal entries of R−t are the reciprocals of the diagonal entries of Rt. Furthermore, we recall that Rt and S have the same diagonal entries. Hence the diagonal entries of R−t are the reciprocals of the diagonal entries of S. We conclude that R−tS has only 1’s on the diagonal. Cholesky Decomposition Let us abbreviate T := R−tS. So far we have T = R−tS is a unit lower triangular matrix. Suppose that T has any non-zero entry below the diagonal, say, in row i and column j. Let i be so small that all entries to the left of ti;j are zero. Then the entry (i; j) of TR will be ti;jrj;i 6= 0, which contradicts St = TR being upper triangular. Hence T is also not only a unit lower triangular matrix but also a diagonal matrix with only 1’s on the diagonal. In other words, T is the unit matrix. So R−tS = I must hold. Putting this together, we finally get St = R. Cholesky Decomposition Theorem Every symmetric positive definite matrix has a Cholesky decomposition. But we have also established before that every matrix which has a Cholesky decomposition is symmetric positive definite. We conclude the following theorem. Theorem A matrix has a Cholesky decomposition if and only if it is symmetric positive definite. Cholesky Decomposition The above arguments also show the following procedure for computing the Cholesky decomposition of a symmetric positive definite matrix: 1: CholeskyDecomposition 2: Compute the factor U in the LU Decomposition 3: for m = 1; 2;:::; n do 4: for k = m; m + 1;:::; n do p 5: rm;k = um;k= um;m 6: end for 7: end for Specialized Algorithms × Consider the matrix A 2 R3 3 given by 04 2 0 1 @2 10 −3A : 0 −3 5 Its LU decomposition can be calculated by 04 2 0 1 0 4 2 0 1 0 4 2 0 1 @2 10 −3A @1=29 −3A @1=29 −3A ; 0 −3 5 0 −3 5 0 −1=34 that is, 0 1 0 01 04 2 0 1 04 2 0 1 @1=2 1 0A @0 9 −3A = @2 10 −3A 0 −1=3 1 0 0 4 0 −3 5 Specialized Algorithms ... that is, 0 1 0 01 04 2 0 1 04 2 0 1 @1=2 1 0A @0 9 −3A = @2 10 −3A 0 −1=3 1 0 0 4 0 −3 5 We extract the diagonal, 0 1 0 1 0 1 1 0 1 1 0 0 2 2 4 2 0 1 ··· = @1=2 1 0A @ 3 A @ 3 A @0 9 −3A − = 1 0 1 3 1 2 2 0 0 4 Specialized Algorithms We extract the diagonal, 0 1 0 1 0 1 1 0 1 1 0 0 2 2 4 2 0 1 ··· = @1=2 1 0A @ 3 A @ 3 A @0 9 −3A − = 1 0 1 3 1 2 2 0 0 4 Bringing the diagonal entries of triangular matrices to matching values, we get 02 0 01 02 1 0 1 04 2 0 1 @1 3 0A @0 3 −1A = @2 10 −3A 0 −1 2 0 0 2 0 −3 5 Note that we have divided the i-th row of U by the square root of its i-th diagonal entry, and we have multiplied the i-th column of L by the corresponding entry. Symmetric Positive Definite Matrices LU Decomposition of Symmetric Positive Definite Matrices Specialized Algorithms Specialized Algorithms The Cholesky decomposition can be calculated from the LU decomposition with only minimal effort. However, a different algorithm seems to more common in practice, as we will discuss in the next subsection. Specialized Algorithms Let A be a positive definite matrix, and let A = RtR be its Cholesky decomposition. Then R is an upper triangular matrix with positive diagonal entries. 0 1 r1;1 r1;2 r1;3 ::: r1;n B r2;2 r2;3 ::: r2;nC B C B r ::: r C R = B 3;3 3;nC : B .. C @ . A rn;n Specialized Algorithms From A = RtR, 0 1 0 1 r1;1 r1;1 r1;2 r1;3 ::: r1;n Br1;2 r2;2 C B r2;2 r2;3 ::: r2;nC B C B C Br r r C B r ::: r C A = B 1;3 2;3 3;3 C B 3;3 3;nC : B .
Recommended publications
  • Conjugate Gradient Method (Part 4)  Pre-Conditioning  Nonlinear Conjugate Gradient Method
    18-660: Numerical Methods for Engineering Design and Optimization Xin Li Department of ECE Carnegie Mellon University Pittsburgh, PA 15213 Slide 1 Overview Conjugate Gradient Method (Part 4) Pre-conditioning Nonlinear conjugate gradient method Slide 2 Conjugate Gradient Method Step 1: start from an initial guess X(0), and set k = 0 Step 2: calculate ( ) ( ) ( ) D 0 = R 0 = B − AX 0 Step 3: update solution D(k )T R(k ) X (k +1) = X (k ) + µ (k )D(k ) where µ (k ) = D(k )T AD(k ) Step 4: calculate residual ( + ) ( ) ( ) ( ) R k 1 = R k − µ k AD k Step 5: determine search direction R(k +1)T R(k +1) (k +1) = (k +1) + β (k ) β = D R k +1,k D where k +1,k (k )T (k ) D R Step 6: set k = k + 1 and go to Step 3 Slide 3 Convergence Rate k ( + ) κ(A) −1 ( ) X k 1 − X ≤ ⋅ X 0 − X κ(A) +1 Conjugate gradient method has slow convergence if κ(A) is large I.e., AX = B is ill-conditioned In this case, we want to improve convergence rate by pre- conditioning Slide 4 Pre-Conditioning Key idea Convert AX = B to another equivalent equation ÃX̃ = B̃ Solve ÃX̃ = B̃ by conjugate gradient method Important constraints to construct ÃX̃ = B̃ à is symmetric and positive definite – so that we can solve it by conjugate gradient method à has a small condition number – so that we can achieve fast convergence Slide 5 Pre-Conditioning AX = B −1 −1 L A⋅ X = L B L−1 AL−T ⋅ LT X = L−1B A X B ̃ ̃ ̃ L−1AL−T is symmetric and positive definite, if A is symmetric and positive definite T (L−1 AL−T ) = L−1 AL−T T X T L−1 AL−T X = (L−T X ) ⋅ A⋅(L−T X )> 0 Slide
    [Show full text]
  • Recursive Approach in Sparse Matrix LU Factorization
    51 Recursive approach in sparse matrix LU factorization Jack Dongarra, Victor Eijkhout and the resulting matrix is often guaranteed to be positive Piotr Łuszczek∗ definite or close to it. However, when the linear sys- University of Tennessee, Department of Computer tem matrix is strongly unsymmetric or indefinite, as Science, Knoxville, TN 37996-3450, USA is the case with matrices originating from systems of Tel.: +865 974 8295; Fax: +865 974 8296 ordinary differential equations or the indefinite matri- ces arising from shift-invert techniques in eigenvalue methods, one has to revert to direct methods which are This paper describes a recursive method for the LU factoriza- the focus of this paper. tion of sparse matrices. The recursive formulation of com- In direct methods, Gaussian elimination with partial mon linear algebra codes has been proven very successful in pivoting is performed to find a solution of Eq. (1). Most dense matrix computations. An extension of the recursive commonly, the factored form of A is given by means technique for sparse matrices is presented. Performance re- L U P Q sults given here show that the recursive approach may per- of matrices , , and such that: form comparable to leading software packages for sparse ma- LU = PAQ, (2) trix factorization in terms of execution time, memory usage, and error estimates of the solution. where: – L is a lower triangular matrix with unitary diago- nal, 1. Introduction – U is an upper triangular matrix with arbitrary di- agonal, Typically, a system of linear equations has the form: – P and Q are row and column permutation matri- Ax = b, (1) ces, respectively (each row and column of these matrices contains single a non-zero entry which is A n n A ∈ n×n x where is by real matrix ( R ), and 1, and the following holds: PPT = QQT = I, b n b, x ∈ n and are -dimensional real vectors ( R ).
    [Show full text]
  • Lecture 5 - Triangular Factorizations & Operation Counts
    Lecture 5 - Triangular Factorizations & Operation Counts LU Factorization We have seen that the process of GE essentially factors a matrix A into LU. Now we want to see how this factorization allows us to solve linear systems and why in many cases it is the preferred algorithm compared with GE. Remember on paper, these methods are the same but computationally they can be different. First, suppose we want to solve A~x = ~b and we are given the factorization A = LU. It turns out that the system LU~x = ~b is \easy" to solve because we do a forward solve L~y = ~b and then back solve U~x = ~y. We have seen that we can easily implement the equations for the back solve and for homework you will write out the equations for the forward solve. Example If 0 2 −1 2 1 0 1 0 0 1 0 2 −1 2 1 A = @ 4 1 9 A = LU = @ 2 1 0 A @ 0 3 5 A 8 5 24 4 3 1 0 0 1 solve the linear system A~x = ~b where ~b = (0; −5; −16)T . ~ We first solve L~y = b to get y1 = 0, 2y1 +y2 = −5 implies y2 = −5 and 4y1 +3y2 +y3 = −16 T implies y3 = −1. Now we solve U~x = ~y = (0; −5; −1) . Back solving yields x3 = −1, 3x2 + 5x3 = −5 implies x2 = 0 and finally 2x1 − x2 + 2x3 = 0 implies x1 = 1 giving the solution (1; 0; −1)T . If GE and LU factorization are equivalent on paper, why would one be computationally advantageous in some settings? Recall that when we solve A~x = ~b by GE we must also multiply the right hand side by the Gauss transformation matrices.
    [Show full text]
  • LU Factorization LU Decomposition LU Decomposition: Motivation LU Decomposition
    LU Factorization LU Decomposition To further improve the efficiency of solving linear systems Factorizations of matrix A : LU and QR A matrix A can be decomposed into a lower triangular matrix L and upper triangular matrix U so that LU Factorization Methods: Using basic Gaussian Elimination (GE) A = LU ◦ Factorization of Tridiagonal Matrix ◦ LU decomposition is performed once; can be used to solve multiple Using Gaussian Elimination with pivoting ◦ right hand sides. Direct LU Factorization ◦ Similar to Gaussian elimination, care must be taken to avoid roundoff Factorizing Symmetrix Matrices (Cholesky Decomposition) ◦ errors (partial or full pivoting) Applications Special Cases: Banded matrices, Symmetric matrices. Analysis ITCS 4133/5133: Intro. to Numerical Methods 1 LU/QR Factorization ITCS 4133/5133: Intro. to Numerical Methods 2 LU/QR Factorization LU Decomposition: Motivation LU Decomposition Forward pass of Gaussian elimination results in the upper triangular ma- trix 1 u12 u13 u1n 0 1 u ··· u 23 ··· 2n U = 0 0 1 u3n . ···. 0 0 0 1 ··· We can determine a matrix L such that LU = A , and l11 0 0 0 l l 0 ··· 0 21 22 ··· L = l31 l32 l33 0 . ···. . l l l 1 n1 n2 n3 ··· ITCS 4133/5133: Intro. to Numerical Methods 3 LU/QR Factorization ITCS 4133/5133: Intro. to Numerical Methods 4 LU/QR Factorization LU Decomposition via Basic Gaussian Elimination LU Decomposition via Basic Gaussian Elimination: Algorithm ITCS 4133/5133: Intro. to Numerical Methods 5 LU/QR Factorization ITCS 4133/5133: Intro. to Numerical Methods 6 LU/QR Factorization LU Decomposition of Tridiagonal Matrix Banded Matrices Matrices that have non-zero elements close to the main diagonal Example: matrix with a bandwidth of 3 (or half band of 1) a11 a12 0 0 0 a21 a22 a23 0 0 0 a32 a33 a34 0 0 0 a43 a44 a45 0 0 0 a a 54 55 Efficiency: Reduced pivoting needed, as elements below bands are zero.
    [Show full text]
  • LU-Factorization 1 Introduction 2 Upper and Lower Triangular Matrices
    MAT067 University of California, Davis Winter 2007 LU-Factorization Isaiah Lankham, Bruno Nachtergaele, Anne Schilling (March 12, 2007) 1 Introduction Given a system of linear equations, a complete reduction of the coefficient matrix to Reduced Row Echelon (RRE) form is far from the most efficient algorithm if one is only interested in finding a solution to the system. However, the Elementary Row Operations (EROs) that constitute such a reduction are themselves at the heart of many frequently used numerical (i.e., computer-calculated) applications of Linear Algebra. In the Sections that follow, we will see how EROs can be used to produce a so-called LU-factorization of a matrix into a product of two significantly simpler matrices. Unlike Diagonalization and the Polar Decomposition for Matrices that we’ve already encountered in this course, these LU Decompositions can be computed reasonably quickly for many matrices. LU-factorizations are also an important tool for solving linear systems of equations. You should note that the factorization of complicated objects into simpler components is an extremely common problem solving technique in mathematics. E.g., we will often factor a polynomial into several polynomials of lower degree, and one can similarly use the prime factorization for an integer in order to simply certain numerical computations. 2 Upper and Lower Triangular Matrices We begin by recalling the terminology for several special forms of matrices. n×n A square matrix A =(aij) ∈ F is called upper triangular if aij = 0 for each pair of integers i, j ∈{1,...,n} such that i>j. In other words, A has the form a11 a12 a13 ··· a1n 0 a22 a23 ··· a2n A = 00a33 ··· a3n .
    [Show full text]
  • Hybrid Algorithms for Efficient Cholesky Decomposition And
    Hybrid Algorithms for Efficient Cholesky Decomposition and Matrix Inverse using Multicore CPUs with GPU Accelerators Gary Macindoe A dissertation submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy of UCL. 2013 ii I, Gary Macindoe, confirm that the work presented in this thesis is my own. Where information has been derived from other sources, I confirm that this has been indicated in the thesis. Signature : Abstract The use of linear algebra routines is fundamental to many areas of computational science, yet their implementation in software still forms the main computational bottleneck in many widely used algorithms. In machine learning and computational statistics, for example, the use of Gaussian distributions is ubiquitous, and routines for calculating the Cholesky decomposition, matrix inverse and matrix determinant must often be called many thousands of times for com- mon algorithms, such as Markov chain Monte Carlo. These linear algebra routines consume most of the total computational time of a wide range of statistical methods, and any improve- ments in this area will therefore greatly increase the overall efficiency of algorithms used in many scientific application areas. The importance of linear algebra algorithms is clear from the substantial effort that has been invested over the last 25 years in producing low-level software libraries such as LAPACK, which generally optimise these linear algebra routines by breaking up a large problem into smaller problems that may be computed independently. The performance of such libraries is however strongly dependent on the specific hardware available. LAPACK was originally de- veloped for single core processors with a memory hierarchy, whereas modern day computers often consist of mixed architectures, with large numbers of parallel cores and graphics process- ing units (GPU) being used alongside traditional CPUs.
    [Show full text]
  • Matrix Inversion Using Cholesky Decomposition
    Matrix Inversion Using Cholesky Decomposition Aravindh Krishnamoorthy, Deepak Menon ST-Ericsson India Private Limited, Bangalore [email protected], [email protected] Abstract—In this paper we present a method for matrix inversion Upper triangular elements, i.e. : based on Cholesky decomposition with reduced number of operations by avoiding computation of intermediate results; further, we use fixed point simulations to compare the numerical ( ∑ ) accuracy of the method. … (4) Keywords-matrix, inversion, Cholesky, LDL. Note that since older values of aii aren’t required for computing newer elements, they may be overwritten by the I. INTRODUCTION value of rii, hence, the algorithm may be performed in-place Matrix inversion techniques based on Cholesky using the same memory for matrices A and R. decomposition and the related LDL decomposition are efficient Cholesky decomposition is of order and requires techniques widely used for inversion of positive- operations. Matrix inversion based on Cholesky definite/symmetric matrices across multiple fields. decomposition is numerically stable for well conditioned Existing matrix inversion algorithms based on Cholesky matrices. decomposition use either equation solving [3] or triangular matrix operations [4] with most efficient implementation If , with is the linear system with requiring operations. variables, and satisfies the requirement for Cholesky decomposition, we can rewrite the linear system as In this paper we propose an inversion algorithm which … (5) reduces the number of operations by 16-17% compared to the existing algorithms by avoiding computation of some known By letting , we have intermediate results. … (6) In section 2 of this paper we review the Cholesky and LDL decomposition techniques, and discuss solutions to linear and systems based on them.
    [Show full text]
  • LU, QR and Cholesky Factorizations Using Vector Capabilities of Gpus LAPACK Working Note 202 Vasily Volkov James W
    LU, QR and Cholesky Factorizations using Vector Capabilities of GPUs LAPACK Working Note 202 Vasily Volkov James W. Demmel Computer Science Division Computer Science Division and Department of Mathematics University of California at Berkeley University of California at Berkeley Abstract GeForce Quadro GeForce GeForce GPU name We present performance results for dense linear algebra using 8800GTX FX5600 8800GTS 8600GTS the 8-series NVIDIA GPUs. Our matrix-matrix multiply routine # of SIMD cores 16 16 12 4 (GEMM) runs 60% faster than the vendor implementation in CUBLAS 1.1 and approaches the peak of hardware capabilities. core clock, GHz 1.35 1.35 1.188 1.458 Our LU, QR and Cholesky factorizations achieve up to 80–90% peak Gflop/s 346 346 228 93.3 of the peak GEMM rate. Our parallel LU running on two GPUs peak Gflop/s/core 21.6 21.6 19.0 23.3 achieves up to ~300 Gflop/s. These results are accomplished by memory bus, MHz 900 800 800 1000 challenging the accepted view of the GPU architecture and memory bus, pins 384 384 320 128 programming guidelines. We argue that modern GPUs should be viewed as multithreaded multicore vector units. We exploit bandwidth, GB/s 86 77 64 32 blocking similarly to vector computers and heterogeneity of the memory size, MB 768 1535 640 256 system by computing both on GPU and CPU. This study flops:word 16 18 14 12 includes detailed benchmarking of the GPU memory system that Table 1: The list of the GPUs used in this study. Flops:word is the reveals sizes and latencies of caches and TLB.
    [Show full text]
  • AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 3: Positive-Definite Systems; Cholesky Factorization
    AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 3: Positive-Definite Systems; Cholesky Factorization Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical Analysis I 1 / 11 Symmetric Positive-Definite Matrices n×n Symmetric matrix A 2 R is symmetric positive definite (SPD) if T n x Ax > 0 for x 2 R nf0g n×n Hermitian matrix A 2 C is Hermitian positive definite (HPD) if ∗ n x Ax > 0 for x 2 C nf0g SPD matrices have positive real eigenvalues and orthogonal eigenvectors Note: Most textbooks only talk about SPD or HPD matrices, but a positive-definite matrix does not need to be symmetric or Hermitian! A real matrix A is positive definite iff A + AT is SPD T n If x Ax ≥ 0 for x 2 R nf0g, then A is said to be positive semidefinite Xiangmin Jiao Numerical Analysis I 2 / 11 Properties of Symmetric Positive-Definite Matrices SPD matrix often arises as Hessian matrix of some convex functional I E.g., least squares problems; partial differential equations If A is SPD, then A is nonsingular Let X be any n × m matrix with full rank and n ≥ m. Then T I X X is symmetric positive definite, and T I XX is symmetric positive semidefinite n×m If A is n × n SPD and X 2 R has full rank and n ≥ m, then X T AX is SPD Any principal submatrix (picking some rows and corresponding columns) of A is SPD and aii > 0 Xiangmin Jiao Numerical Analysis I 3 / 11 Cholesky Factorization If A is symmetric positive definite, then there is factorization of A A = RT R where R is upper triangular, and all its diagonal entries are positive Note: Textbook calls is “Cholesky
    [Show full text]
  • 1. Positive Definite Matrices a Matrix a Is Positive Definite If X>Ax > 0 for All Nonzero X
    CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 8 GENE H. GOLUB 1. Positive Definite Matrices A matrix A is positive definite if x>Ax > 0 for all nonzero x. A positive definite matrix has real and positive eigenvalues, and its leading principal submatrices all have positive determinants. From the definition, it is easy to see that all diagonal elements are positive. To solve the system Ax = b where A is positive definite, we can compute the Cholesky decom- position A = F >F where F is upper triangular. This decomposition exists if and only if A is symmetric and positive definite. In fact, attempting to compute the Cholesky decomposition of A is an efficient method for checking whether A is symmetric positive definite. It is important to distinguish the Cholesky decomposition from the square root factorization.A square root of a matrix A is defined as a matrix S such that S2 = SS = A. Note that the matrix F in A = F >F is not the square root of A, since it does not hold that F 2 = A unless A is a diagonal matrix. The square root of a symmetric positive definite A can be computed by using the fact that A has an eigendecomposition A = UΛU > where Λ is a diagonal matrix whose diagonal elements are the positive eigenvalues of A and U is an orthogonal matrix whose columns are the eigenvectors of A. It follows that A = UΛU > = (UΛ1/2U >)(UΛ1/2U >) = SS and so S = UΛ1/2U > is a square root of A. 2. The Cholesky Decomposition The Cholesky decomposition can be computed directly from the matrix equation A = F >F .
    [Show full text]
  • Cholesky Decomposition 1 Cholesky Decomposition
    Cholesky decomposition 1 Cholesky decomposition In linear algebra, the Cholesky decomposition or Cholesky triangle is a decomposition of a symmetric, positive-definite matrix into the product of a lower triangular matrix and its conjugate transpose. It was discovered by André-Louis Cholesky for real matrices and is an example of a square root of a matrix. When it is applicable, the Cholesky decomposition is roughly twice as efficient as the LU decomposition for solving systems of linear equations.[1] Statement If A has real entries and is symmetric (or more generally, is Hermitian) and positive definite, then A can be decomposed as where L is a lower triangular matrix with strictly positive diagonal entries, and L* denotes the conjugate transpose of L. This is the Cholesky decomposition. The Cholesky decomposition is unique: given a Hermitian, positive-definite matrix A, there is only one lower triangular matrix L with strictly positive diagonal entries such that A = LL*. The converse holds trivially: if A can be written as LL* for some invertible L, lower triangular or otherwise, then A is Hermitian and positive definite. The requirement that L have strictly positive diagonal entries can be dropped to extend the factorization to the positive-semidefinite case. The statement then reads: a square matrix A has a Cholesky decomposition if and only if A is Hermitian and positive semi-definite. Cholesky factorizations for positive-semidefinite matrices are not unique in general. In the special case that A is a symmetric positive-definite matrix with real entries, L can be assumed to have real entries as well.
    [Show full text]
  • LU and Cholesky Decompositions. J. Demmel, Chapter 2.7
    LU and Cholesky decompositions. J. Demmel, Chapter 2.7 1/14 Gaussian Elimination The Algorithm — Overview Solving Ax = b using Gaussian elimination. 1 Factorize A into A = PLU Permutation Unit lower triangular Non-singular upper triangular 2/14 Gaussian Elimination The Algorithm — Overview Solving Ax = b using Gaussian elimination. 1 Factorize A into A = PLU Permutation Unit lower triangular Non-singular upper triangular 2 Solve PLUx = b (for LUx) : − LUx = P 1b 3/14 Gaussian Elimination The Algorithm — Overview Solving Ax = b using Gaussian elimination. 1 Factorize A into A = PLU Permutation Unit lower triangular Non-singular upper triangular 2 Solve PLUx = b (for LUx) : − LUx = P 1b − 3 Solve LUx = P 1b (for Ux) by forward substitution: − − Ux = L 1(P 1b). 4/14 Gaussian Elimination The Algorithm — Overview Solving Ax = b using Gaussian elimination. 1 Factorize A into A = PLU Permutation Unit lower triangular Non-singular upper triangular 2 Solve PLUx = b (for LUx) : − LUx = P 1b − 3 Solve LUx = P 1b (for Ux) by forward substitution: − − Ux = L 1(P 1b). − − 4 Solve Ux = L 1(P 1b) by backward substitution: − − − x = U 1(L 1P 1b). 5/14 Example of LU factorization We factorize the following 2-by-2 matrix: 4 3 l 0 u u = 11 11 12 . (1) 6 3 l21 l22 0 u22 One way to find the LU decomposition of this simple matrix would be to simply solve the linear equations by inspection. Expanding the matrix multiplication gives l11 · u11 + 0 · 0 = 4, l11 · u12 + 0 · u22 = 3, (2) l21 · u11 + l22 · 0 = 6, l21 · u12 + l22 · u22 = 3.
    [Show full text]