MATH 3795 Lecture 5. Solving Linear Systems 3

Total Page:16

File Type:pdf, Size:1020Kb

MATH 3795 Lecture 5. Solving Linear Systems 3 MATH 3795 Lecture 5. Solving Linear Systems 3 Dmitriy Leykekhman Fall 2008 Goals I Positive definite and definite matrices. I Cholesky decomposition. I LU-Decomposition of Tridiagonal Systems I Applications. D. Leykekhman - MATH 3795 Introduction to Computational Mathematics Symmetric and Banded Matrices { 1 n×n I A 2 R is called m- banded if aij = 0 for ji − jj > m. n×n I If A 2 R is symmetric positive definite, then the LU decomposition can be computed in a stable way without permutation, i.e., A = LU I Can we use the structure of A, i.e. symmetry and the positive definiteness of A or the property that A is m-banded to compute LU-decomposition more efficiently? Symmetric Positive Definite Matrices n×n T I A 2 R is called symmetric if A = A . n×n T I A 2 R is called symmetric positive definite if A = A and vT Av > 0 for all v 2 Rn, v 6= 0. D. Leykekhman - MATH 3795 Introduction to Computational Mathematics Symmetric and Banded Matrices { 2 n×n I If A 2 R is symmetric positive definite, then the LU decomposition can be computed in a stable way without permutation, i.e., A = LU I Can we use the structure of A, i.e. symmetry and the positive definiteness of A or the property that A is m-banded to compute LU-decomposition more efficiently? Symmetric Positive Definite Matrices n×n T I A 2 R is called symmetric if A = A . n×n T I A 2 R is called symmetric positive definite if A = A and vT Av > 0 for all v 2 Rn, v 6= 0. n×n I A 2 R is called m- banded if aij = 0 for ji − jj > m. D. Leykekhman - MATH 3795 Introduction to Computational Mathematics Symmetric and Banded Matrices { 2 I Can we use the structure of A, i.e. symmetry and the positive definiteness of A or the property that A is m-banded to compute LU-decomposition more efficiently? Symmetric Positive Definite Matrices n×n T I A 2 R is called symmetric if A = A . n×n T I A 2 R is called symmetric positive definite if A = A and vT Av > 0 for all v 2 Rn, v 6= 0. n×n I A 2 R is called m- banded if aij = 0 for ji − jj > m. n×n I If A 2 R is symmetric positive definite, then the LU decomposition can be computed in a stable way without permutation, i.e., A = LU D. Leykekhman - MATH 3795 Introduction to Computational Mathematics Symmetric and Banded Matrices { 2 Symmetric Positive Definite Matrices n×n T I A 2 R is called symmetric if A = A . n×n T I A 2 R is called symmetric positive definite if A = A and vT Av > 0 for all v 2 Rn, v 6= 0. n×n I A 2 R is called m- banded if aij = 0 for ji − jj > m. n×n I If A 2 R is symmetric positive definite, then the LU decomposition can be computed in a stable way without permutation, i.e., A = LU I Can we use the structure of A, i.e. symmetry and the positive definiteness of A or the property that A is m-banded to compute LU-decomposition more efficiently? D. Leykekhman - MATH 3795 Introduction to Computational Mathematics Symmetric and Banded Matrices { 2 n×n I Fact 2: If A 2 R is symmetric positive definite and A = LU, then the diagonal entries of U are positive. I Any upper triangular matrix U we can write as U = DU~, where U~ is unit upper triangular and D = diag(u11; : : : ; unn): Symmetric Positive Definite Matrices I Fact 1: If L is a unit lower triangular matrix and U is an upper triangular matrix such that A = LU, then L and U are unique, i.e., if there L~ is a unit lower triangular matrix and U~ is an upper triangular matrix such that A = L~U~, then L = L~ and U = U~ . D. Leykekhman - MATH 3795 Introduction to Computational Mathematics Symmetric and Banded Matrices { 3 I Any upper triangular matrix U we can write as U = DU~, where U~ is unit upper triangular and D = diag(u11; : : : ; unn): Symmetric Positive Definite Matrices I Fact 1: If L is a unit lower triangular matrix and U is an upper triangular matrix such that A = LU, then L and U are unique, i.e., if there L~ is a unit lower triangular matrix and U~ is an upper triangular matrix such that A = L~U~, then L = L~ and U = U~ . n×n I Fact 2: If A 2 R is symmetric positive definite and A = LU, then the diagonal entries of U are positive. D. Leykekhman - MATH 3795 Introduction to Computational Mathematics Symmetric and Banded Matrices { 3 Symmetric Positive Definite Matrices I Fact 1: If L is a unit lower triangular matrix and U is an upper triangular matrix such that A = LU, then L and U are unique, i.e., if there L~ is a unit lower triangular matrix and U~ is an upper triangular matrix such that A = L~U~, then L = L~ and U = U~ . n×n I Fact 2: If A 2 R is symmetric positive definite and A = LU, then the diagonal entries of U are positive. I Any upper triangular matrix U we can write as U = DU~, where U~ is unit upper triangular and D = diag(u11; : : : ; unn): D. Leykekhman - MATH 3795 Introduction to Computational Mathematics Symmetric and Banded Matrices { 3 Using that A = AT and LU decomposition is uniqie A = LU = U~ T DLT = U~ T (DLT ) = (lower unit triangular)×(upper triangular): Thus, L = U~ T and U = DLT : Symmetric Positive Definite Matrices Thus, A = LU = LDU:~ On the other hand AT = (LDU~)T = U~ T DLT : D. Leykekhman - MATH 3795 Introduction to Computational Mathematics Symmetric and Banded Matrices { 4 Symmetric Positive Definite Matrices Thus, A = LU = LDU:~ On the other hand AT = (LDU~)T = U~ T DLT : Using that A = AT and LU decomposition is uniqie A = LU = U~ T DLT = U~ T (DLT ) = (lower unit triangular)×(upper triangular): Thus, L = U~ T and U = DLT : D. Leykekhman - MATH 3795 Introduction to Computational Mathematics Symmetric and Banded Matrices { 4 I Recall that D = diag(u11; : : : ; unn) has positive diagonal entries. So we can define 1=2 p p D = diag( u11;::: unn): Define R := D1=2LT , then A = LDLT = LD1=2D1=2LT = RT R : Cholesky-decomposition Symmetric Positive Definite Matrices I Above we showed that if A is a symmetric positive definite matrix, then A = LDLT : D. Leykekhman - MATH 3795 Introduction to Computational Mathematics Symmetric and Banded Matrices { 5 Symmetric Positive Definite Matrices I Above we showed that if A is a symmetric positive definite matrix, then A = LDLT : I Recall that D = diag(u11; : : : ; unn) has positive diagonal entries. So we can define 1=2 p p D = diag( u11;::: unn): Define R := D1=2LT , then A = LDLT = LD1=2D1=2LT = RT R : Cholesky-decomposition D. Leykekhman - MATH 3795 Introduction to Computational Mathematics Symmetric and Banded Matrices { 5 Example 0 4 −4 8 1 0 1 0 0 1 0 4 −4 8 1 @ −4 8 −4 A = @ −1 1 0 A @ 0 4 4 A 8 −4 29 2 1 1 0 0 9 | {z } | {z } | {z } A L U 0 1 0 0 1 0 4 0 0 1 0 1 −1 2 1 = @ −1 1 0 A @ 0 4 0 A @ 0 1 1 A 2 1 1 0 0 9 0 0 1 | {z } | {z } | {z } L D LT 0 2 0 0 1 0 2 −2 4 1 = @ −2 2 0 A @ 0 2 2 A 4 2 3 0 0 3 | {z } | {z } RT R D. Leykekhman - MATH 3795 Introduction to Computational Mathematics Symmetric and Banded Matrices { 6 T T I So far we know how to compute LDL and R R by first computing the LU-decomposition and then derive LDLT or RT R from it. No savings. Not very useful. Symmetric Positive Definite Matrices n×n I A matrix A 2 R is symmetric positive definite if and only if there exists a unit lower triangular matrix L and a positive definite diagonal matrix D such that A = LDLT . n×n I A matrix A 2 R is symmetric positive definite if and only if there exists an upper triangular matrix R with rii > 0, i = 1; : : : ; n, such that A = RT R. D. Leykekhman - MATH 3795 Introduction to Computational Mathematics Symmetric and Banded Matrices { 7 Symmetric Positive Definite Matrices n×n I A matrix A 2 R is symmetric positive definite if and only if there exists a unit lower triangular matrix L and a positive definite diagonal matrix D such that A = LDLT . n×n I A matrix A 2 R is symmetric positive definite if and only if there exists an upper triangular matrix R with rii > 0, i = 1; : : : ; n, such that A = RT R. T T I So far we know how to compute LDL and R R by first computing the LU-decomposition and then derive LDLT or RT R from it. No savings. Not very useful. D. Leykekhman - MATH 3795 Introduction to Computational Mathematics Symmetric and Banded Matrices { 7 I Can easily derive the algorithm I Fix i = 1 and let j = 1 : n; a = r r . This implies p 1j 11 1j r11 = a11, j = 1 and r1j = a1j =r11, j = 2 : n. I Fix i =p 2 and let j = 2 : n; a2j = r12r1j + r22r2j .
Recommended publications
  • Conjugate Gradient Method (Part 4)  Pre-Conditioning  Nonlinear Conjugate Gradient Method
    18-660: Numerical Methods for Engineering Design and Optimization Xin Li Department of ECE Carnegie Mellon University Pittsburgh, PA 15213 Slide 1 Overview Conjugate Gradient Method (Part 4) Pre-conditioning Nonlinear conjugate gradient method Slide 2 Conjugate Gradient Method Step 1: start from an initial guess X(0), and set k = 0 Step 2: calculate ( ) ( ) ( ) D 0 = R 0 = B − AX 0 Step 3: update solution D(k )T R(k ) X (k +1) = X (k ) + µ (k )D(k ) where µ (k ) = D(k )T AD(k ) Step 4: calculate residual ( + ) ( ) ( ) ( ) R k 1 = R k − µ k AD k Step 5: determine search direction R(k +1)T R(k +1) (k +1) = (k +1) + β (k ) β = D R k +1,k D where k +1,k (k )T (k ) D R Step 6: set k = k + 1 and go to Step 3 Slide 3 Convergence Rate k ( + ) κ(A) −1 ( ) X k 1 − X ≤ ⋅ X 0 − X κ(A) +1 Conjugate gradient method has slow convergence if κ(A) is large I.e., AX = B is ill-conditioned In this case, we want to improve convergence rate by pre- conditioning Slide 4 Pre-Conditioning Key idea Convert AX = B to another equivalent equation ÃX̃ = B̃ Solve ÃX̃ = B̃ by conjugate gradient method Important constraints to construct ÃX̃ = B̃ à is symmetric and positive definite – so that we can solve it by conjugate gradient method à has a small condition number – so that we can achieve fast convergence Slide 5 Pre-Conditioning AX = B −1 −1 L A⋅ X = L B L−1 AL−T ⋅ LT X = L−1B A X B ̃ ̃ ̃ L−1AL−T is symmetric and positive definite, if A is symmetric and positive definite T (L−1 AL−T ) = L−1 AL−T T X T L−1 AL−T X = (L−T X ) ⋅ A⋅(L−T X )> 0 Slide
    [Show full text]
  • Recursive Approach in Sparse Matrix LU Factorization
    51 Recursive approach in sparse matrix LU factorization Jack Dongarra, Victor Eijkhout and the resulting matrix is often guaranteed to be positive Piotr Łuszczek∗ definite or close to it. However, when the linear sys- University of Tennessee, Department of Computer tem matrix is strongly unsymmetric or indefinite, as Science, Knoxville, TN 37996-3450, USA is the case with matrices originating from systems of Tel.: +865 974 8295; Fax: +865 974 8296 ordinary differential equations or the indefinite matri- ces arising from shift-invert techniques in eigenvalue methods, one has to revert to direct methods which are This paper describes a recursive method for the LU factoriza- the focus of this paper. tion of sparse matrices. The recursive formulation of com- In direct methods, Gaussian elimination with partial mon linear algebra codes has been proven very successful in pivoting is performed to find a solution of Eq. (1). Most dense matrix computations. An extension of the recursive commonly, the factored form of A is given by means technique for sparse matrices is presented. Performance re- L U P Q sults given here show that the recursive approach may per- of matrices , , and such that: form comparable to leading software packages for sparse ma- LU = PAQ, (2) trix factorization in terms of execution time, memory usage, and error estimates of the solution. where: – L is a lower triangular matrix with unitary diago- nal, 1. Introduction – U is an upper triangular matrix with arbitrary di- agonal, Typically, a system of linear equations has the form: – P and Q are row and column permutation matri- Ax = b, (1) ces, respectively (each row and column of these matrices contains single a non-zero entry which is A n n A ∈ n×n x where is by real matrix ( R ), and 1, and the following holds: PPT = QQT = I, b n b, x ∈ n and are -dimensional real vectors ( R ).
    [Show full text]
  • Lecture 5 - Triangular Factorizations & Operation Counts
    Lecture 5 - Triangular Factorizations & Operation Counts LU Factorization We have seen that the process of GE essentially factors a matrix A into LU. Now we want to see how this factorization allows us to solve linear systems and why in many cases it is the preferred algorithm compared with GE. Remember on paper, these methods are the same but computationally they can be different. First, suppose we want to solve A~x = ~b and we are given the factorization A = LU. It turns out that the system LU~x = ~b is \easy" to solve because we do a forward solve L~y = ~b and then back solve U~x = ~y. We have seen that we can easily implement the equations for the back solve and for homework you will write out the equations for the forward solve. Example If 0 2 −1 2 1 0 1 0 0 1 0 2 −1 2 1 A = @ 4 1 9 A = LU = @ 2 1 0 A @ 0 3 5 A 8 5 24 4 3 1 0 0 1 solve the linear system A~x = ~b where ~b = (0; −5; −16)T . ~ We first solve L~y = b to get y1 = 0, 2y1 +y2 = −5 implies y2 = −5 and 4y1 +3y2 +y3 = −16 T implies y3 = −1. Now we solve U~x = ~y = (0; −5; −1) . Back solving yields x3 = −1, 3x2 + 5x3 = −5 implies x2 = 0 and finally 2x1 − x2 + 2x3 = 0 implies x1 = 1 giving the solution (1; 0; −1)T . If GE and LU factorization are equivalent on paper, why would one be computationally advantageous in some settings? Recall that when we solve A~x = ~b by GE we must also multiply the right hand side by the Gauss transformation matrices.
    [Show full text]
  • LU Factorization LU Decomposition LU Decomposition: Motivation LU Decomposition
    LU Factorization LU Decomposition To further improve the efficiency of solving linear systems Factorizations of matrix A : LU and QR A matrix A can be decomposed into a lower triangular matrix L and upper triangular matrix U so that LU Factorization Methods: Using basic Gaussian Elimination (GE) A = LU ◦ Factorization of Tridiagonal Matrix ◦ LU decomposition is performed once; can be used to solve multiple Using Gaussian Elimination with pivoting ◦ right hand sides. Direct LU Factorization ◦ Similar to Gaussian elimination, care must be taken to avoid roundoff Factorizing Symmetrix Matrices (Cholesky Decomposition) ◦ errors (partial or full pivoting) Applications Special Cases: Banded matrices, Symmetric matrices. Analysis ITCS 4133/5133: Intro. to Numerical Methods 1 LU/QR Factorization ITCS 4133/5133: Intro. to Numerical Methods 2 LU/QR Factorization LU Decomposition: Motivation LU Decomposition Forward pass of Gaussian elimination results in the upper triangular ma- trix 1 u12 u13 u1n 0 1 u ··· u 23 ··· 2n U = 0 0 1 u3n . ···. 0 0 0 1 ··· We can determine a matrix L such that LU = A , and l11 0 0 0 l l 0 ··· 0 21 22 ··· L = l31 l32 l33 0 . ···. . l l l 1 n1 n2 n3 ··· ITCS 4133/5133: Intro. to Numerical Methods 3 LU/QR Factorization ITCS 4133/5133: Intro. to Numerical Methods 4 LU/QR Factorization LU Decomposition via Basic Gaussian Elimination LU Decomposition via Basic Gaussian Elimination: Algorithm ITCS 4133/5133: Intro. to Numerical Methods 5 LU/QR Factorization ITCS 4133/5133: Intro. to Numerical Methods 6 LU/QR Factorization LU Decomposition of Tridiagonal Matrix Banded Matrices Matrices that have non-zero elements close to the main diagonal Example: matrix with a bandwidth of 3 (or half band of 1) a11 a12 0 0 0 a21 a22 a23 0 0 0 a32 a33 a34 0 0 0 a43 a44 a45 0 0 0 a a 54 55 Efficiency: Reduced pivoting needed, as elements below bands are zero.
    [Show full text]
  • LU-Factorization 1 Introduction 2 Upper and Lower Triangular Matrices
    MAT067 University of California, Davis Winter 2007 LU-Factorization Isaiah Lankham, Bruno Nachtergaele, Anne Schilling (March 12, 2007) 1 Introduction Given a system of linear equations, a complete reduction of the coefficient matrix to Reduced Row Echelon (RRE) form is far from the most efficient algorithm if one is only interested in finding a solution to the system. However, the Elementary Row Operations (EROs) that constitute such a reduction are themselves at the heart of many frequently used numerical (i.e., computer-calculated) applications of Linear Algebra. In the Sections that follow, we will see how EROs can be used to produce a so-called LU-factorization of a matrix into a product of two significantly simpler matrices. Unlike Diagonalization and the Polar Decomposition for Matrices that we’ve already encountered in this course, these LU Decompositions can be computed reasonably quickly for many matrices. LU-factorizations are also an important tool for solving linear systems of equations. You should note that the factorization of complicated objects into simpler components is an extremely common problem solving technique in mathematics. E.g., we will often factor a polynomial into several polynomials of lower degree, and one can similarly use the prime factorization for an integer in order to simply certain numerical computations. 2 Upper and Lower Triangular Matrices We begin by recalling the terminology for several special forms of matrices. n×n A square matrix A =(aij) ∈ F is called upper triangular if aij = 0 for each pair of integers i, j ∈{1,...,n} such that i>j. In other words, A has the form a11 a12 a13 ··· a1n 0 a22 a23 ··· a2n A = 00a33 ··· a3n .
    [Show full text]
  • The Inverse Along a Lower Triangular Matrix∗
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Universidade do Minho: RepositoriUM The inverse along a lower triangular matrix∗ Xavier Marya, Pedro Patr´ıciob aUniversit´eParis-Ouest Nanterre { La D´efense,Laboratoire Modal'X, 200 avenuue de la r´epublique,92000 Nanterre, France. email: [email protected] bDepartamento de Matem´aticae Aplica¸c~oes,Universidade do Minho, 4710-057 Braga, Portugal. email: [email protected] Abstract In this paper, we investigate the recently defined notion of inverse along an element in the context of matrices over a ring. Precisely, we study the inverse of a matrix along a lower triangular matrix, under some conditions. Keywords: Generalized inverse, inverse along an element, Dedekind-finite ring, Green's relations, rings AMS classification: 15A09, 16E50 1 Introduction In this paper, R is a ring with identity. We say a is (von Neumann) regular in R if a 2 aRa.A particular solution to axa = a is denoted by a−, and the set of all such solutions is denoted by af1g. Given a−; a= 2 af1g then x = a=aa− satisfies axa = a; xax = a simultaneously. Such a solution is called a reflexive inverse, and is denoted by a+. The set of all reflexive inverses of a is denoted by af1; 2g. Finally, a is group invertible if there is a# 2 af1; 2g that commutes with a, and a is Drazin invertible if ak is group invertible, for some non-negative integer k. This is equivalent to the existence of aD 2 R such that ak+1aD = ak; aDaaD = aD; aaD = aDa.
    [Show full text]
  • Hybrid Algorithms for Efficient Cholesky Decomposition And
    Hybrid Algorithms for Efficient Cholesky Decomposition and Matrix Inverse using Multicore CPUs with GPU Accelerators Gary Macindoe A dissertation submitted in partial fulfillment of the requirements for the degree of Doctor of Philosophy of UCL. 2013 ii I, Gary Macindoe, confirm that the work presented in this thesis is my own. Where information has been derived from other sources, I confirm that this has been indicated in the thesis. Signature : Abstract The use of linear algebra routines is fundamental to many areas of computational science, yet their implementation in software still forms the main computational bottleneck in many widely used algorithms. In machine learning and computational statistics, for example, the use of Gaussian distributions is ubiquitous, and routines for calculating the Cholesky decomposition, matrix inverse and matrix determinant must often be called many thousands of times for com- mon algorithms, such as Markov chain Monte Carlo. These linear algebra routines consume most of the total computational time of a wide range of statistical methods, and any improve- ments in this area will therefore greatly increase the overall efficiency of algorithms used in many scientific application areas. The importance of linear algebra algorithms is clear from the substantial effort that has been invested over the last 25 years in producing low-level software libraries such as LAPACK, which generally optimise these linear algebra routines by breaking up a large problem into smaller problems that may be computed independently. The performance of such libraries is however strongly dependent on the specific hardware available. LAPACK was originally de- veloped for single core processors with a memory hierarchy, whereas modern day computers often consist of mixed architectures, with large numbers of parallel cores and graphics process- ing units (GPU) being used alongside traditional CPUs.
    [Show full text]
  • LU Decompositions We Seek a Factorization of a Square Matrix a Into the Product of Two Matrices Which Yields an Efficient Method
    LU Decompositions We seek a factorization of a square matrix A into the product of two matrices which yields an efficient method for solving the system where A is the coefficient matrix, x is our variable vector and is a constant vector for . The factorization of A into the product of two matrices is closely related to Gaussian elimination. Definition 1. A square matrix is said to be lower triangular if for all . 2. A square matrix is said to be unit lower triangular if it is lower triangular and each . 3. A square matrix is said to be upper triangular if for all . Examples 1. The following are all lower triangular matrices: , , 2. The following are all unit lower triangular matrices: , , 3. The following are all upper triangular matrices: , , We note that the identity matrix is only square matrix that is both unit lower triangular and upper triangular. Example Let . For elementary matrices (See solv_lin_equ2.pdf) , , and we find that . Now, if , then direct computation yields and . It follows that and, hence, that where L is unit lower triangular and U is upper triangular. That is, . Observe the key fact that the unit lower triangular matrix L “contains” the essential data of the three elementary matrices , , and . Definition We say that the matrix A has an LU decomposition if where L is unit lower triangular and U is upper triangular. We also call the LU decomposition an LU factorization. Example 1. and so has an LU decomposition. 2. The matrix has more than one LU decomposition. Two such LU factorizations are and .
    [Show full text]
  • Handout 9 More Matrix Properties; the Transpose
    Handout 9 More matrix properties; the transpose Square matrix properties These properties only apply to a square matrix, i.e. n £ n. ² The leading diagonal is the diagonal line consisting of the entries a11, a22, a33, . ann. ² A diagonal matrix has zeros everywhere except the leading diagonal. ² The identity matrix I has zeros o® the leading diagonal, and 1 for each entry on the diagonal. It is a special case of a diagonal matrix, and A I = I A = A for any n £ n matrix A. ² An upper triangular matrix has all its non-zero entries on or above the leading diagonal. ² A lower triangular matrix has all its non-zero entries on or below the leading diagonal. ² A symmetric matrix has the same entries below and above the diagonal: aij = aji for any values of i and j between 1 and n. ² An antisymmetric or skew-symmetric matrix has the opposite entries below and above the diagonal: aij = ¡aji for any values of i and j between 1 and n. This automatically means the digaonal entries must all be zero. Transpose To transpose a matrix, we reect it across the line given by the leading diagonal a11, a22 etc. In general the result is a di®erent shape to the original matrix: a11 a21 a11 a12 a13 > > A = A = 0 a12 a22 1 [A ]ij = A : µ a21 a22 a23 ¶ ji a13 a23 @ A > ² If A is m £ n then A is n £ m. > ² The transpose of a symmetric matrix is itself: A = A (recalling that only square matrices can be symmetric).
    [Show full text]
  • Matrix Inversion Using Cholesky Decomposition
    Matrix Inversion Using Cholesky Decomposition Aravindh Krishnamoorthy, Deepak Menon ST-Ericsson India Private Limited, Bangalore [email protected], [email protected] Abstract—In this paper we present a method for matrix inversion Upper triangular elements, i.e. : based on Cholesky decomposition with reduced number of operations by avoiding computation of intermediate results; further, we use fixed point simulations to compare the numerical ( ∑ ) accuracy of the method. … (4) Keywords-matrix, inversion, Cholesky, LDL. Note that since older values of aii aren’t required for computing newer elements, they may be overwritten by the I. INTRODUCTION value of rii, hence, the algorithm may be performed in-place Matrix inversion techniques based on Cholesky using the same memory for matrices A and R. decomposition and the related LDL decomposition are efficient Cholesky decomposition is of order and requires techniques widely used for inversion of positive- operations. Matrix inversion based on Cholesky definite/symmetric matrices across multiple fields. decomposition is numerically stable for well conditioned Existing matrix inversion algorithms based on Cholesky matrices. decomposition use either equation solving [3] or triangular matrix operations [4] with most efficient implementation If , with is the linear system with requiring operations. variables, and satisfies the requirement for Cholesky decomposition, we can rewrite the linear system as In this paper we propose an inversion algorithm which … (5) reduces the number of operations by 16-17% compared to the existing algorithms by avoiding computation of some known By letting , we have intermediate results. … (6) In section 2 of this paper we review the Cholesky and LDL decomposition techniques, and discuss solutions to linear and systems based on them.
    [Show full text]
  • LU, QR and Cholesky Factorizations Using Vector Capabilities of Gpus LAPACK Working Note 202 Vasily Volkov James W
    LU, QR and Cholesky Factorizations using Vector Capabilities of GPUs LAPACK Working Note 202 Vasily Volkov James W. Demmel Computer Science Division Computer Science Division and Department of Mathematics University of California at Berkeley University of California at Berkeley Abstract GeForce Quadro GeForce GeForce GPU name We present performance results for dense linear algebra using 8800GTX FX5600 8800GTS 8600GTS the 8-series NVIDIA GPUs. Our matrix-matrix multiply routine # of SIMD cores 16 16 12 4 (GEMM) runs 60% faster than the vendor implementation in CUBLAS 1.1 and approaches the peak of hardware capabilities. core clock, GHz 1.35 1.35 1.188 1.458 Our LU, QR and Cholesky factorizations achieve up to 80–90% peak Gflop/s 346 346 228 93.3 of the peak GEMM rate. Our parallel LU running on two GPUs peak Gflop/s/core 21.6 21.6 19.0 23.3 achieves up to ~300 Gflop/s. These results are accomplished by memory bus, MHz 900 800 800 1000 challenging the accepted view of the GPU architecture and memory bus, pins 384 384 320 128 programming guidelines. We argue that modern GPUs should be viewed as multithreaded multicore vector units. We exploit bandwidth, GB/s 86 77 64 32 blocking similarly to vector computers and heterogeneity of the memory size, MB 768 1535 640 256 system by computing both on GPU and CPU. This study flops:word 16 18 14 12 includes detailed benchmarking of the GPU memory system that Table 1: The list of the GPUs used in this study. Flops:word is the reveals sizes and latencies of caches and TLB.
    [Show full text]
  • AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 3: Positive-Definite Systems; Cholesky Factorization
    AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 3: Positive-Definite Systems; Cholesky Factorization Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical Analysis I 1 / 11 Symmetric Positive-Definite Matrices n×n Symmetric matrix A 2 R is symmetric positive definite (SPD) if T n x Ax > 0 for x 2 R nf0g n×n Hermitian matrix A 2 C is Hermitian positive definite (HPD) if ∗ n x Ax > 0 for x 2 C nf0g SPD matrices have positive real eigenvalues and orthogonal eigenvectors Note: Most textbooks only talk about SPD or HPD matrices, but a positive-definite matrix does not need to be symmetric or Hermitian! A real matrix A is positive definite iff A + AT is SPD T n If x Ax ≥ 0 for x 2 R nf0g, then A is said to be positive semidefinite Xiangmin Jiao Numerical Analysis I 2 / 11 Properties of Symmetric Positive-Definite Matrices SPD matrix often arises as Hessian matrix of some convex functional I E.g., least squares problems; partial differential equations If A is SPD, then A is nonsingular Let X be any n × m matrix with full rank and n ≥ m. Then T I X X is symmetric positive definite, and T I XX is symmetric positive semidefinite n×m If A is n × n SPD and X 2 R has full rank and n ≥ m, then X T AX is SPD Any principal submatrix (picking some rows and corresponding columns) of A is SPD and aii > 0 Xiangmin Jiao Numerical Analysis I 3 / 11 Cholesky Factorization If A is symmetric positive definite, then there is factorization of A A = RT R where R is upper triangular, and all its diagonal entries are positive Note: Textbook calls is “Cholesky
    [Show full text]