Section 6.1.2: the Schur Decomposition 6.1.3: Nonunitary Decompositions

Section 6.1.2: the Schur Decomposition 6.1.3: Nonunitary Decompositions

Section 6.1.2: The Schur Decomposition 6.1.3: Nonunitary Decompositions Jim Lambers March 22, 2021 • Now that we have answered the first of the two questions posed earlier, by identifying a class of matrices for which eigenvalues are easily obtained, we move to the second question: how can a general square matrix A be reduced to (block) triangular form, in such a way that its eigenvalues are preserved? • That is, we need to find a similarity transformation (that is, A P −1AP for some P ) that can bring A closer to triangular form, in some sense. n n×n • With this goal in mind, suppose that x 2 C is a normalized eigenvector of A 2 C (that is, kxk2 = 1), with eigenvalue λ. • Furthermore, suppose that P is a Householder reflection such that P x = e1. (this can be done based on Section 4.2) • Because P is complex, it is Hermitian, meaning that P H = P , and unitary, meaning that P H P = I. • This the generalization of a real Householder reflection being symmetric and orthogonal. • Therefore, P is its own inverse, so P e1 = x. • It follows that the matrix P H AP , which is a similarity transformation of A, satisfies H H H P AP e1 = P Ax = λP x = λP x = λe1: H H • That is, e1 is an eigenvector of P AP with eigenvalue λ, and therefore P AP has the block structure λ vH P H AP = : (1) 0 B • Therefore, λ(A) = fλg [ λ(B), which means that we can now focus on the (n − 1) × (n − 1) matrix B to find the rest of the eigenvalues of A. • This process of reducing the eigenvalue problem for A to that of B is called deflation. • Repeating this process of deflation, we obtain the Schur Decomposition A = QT QH (2) n×n where T 2 C is an upper triangular matrix whose diagonal entries are the eigenvalues of n×n A, and Q 2 C is a unitary matrix. 1 • Every square complex matrix has a Schur Decomposition. • The Matlab function schur computes the Schur Decomposition of a given matrix A. • The statement >> [Q,T]=schur(A) computes matrices Q and T such that A = QT QH . • If A is complex, then Q is unitary and T is upper triangular, with the eigenvalues of A along its diagonal. • If A is real, then Q is a real orthogonal matrix and T is real and block upper triangular, with 1×1 diagonal blocks corresponding to real eigenvalues, and 2×2 diagonal blocks corresponding to complex-conjugate pairs of eigenvalues. • This form of the Schur Decomposition is called the Real Schur form. • If the complex Schur Decomposition is desired, one can use the statement [Q,T]=schur(A,'complex'). • If only one output is specified in the call to schur, then only the matrix T is returned. • The columns of Q are called Schur vectors, and the first Schur vector q1 is an eigenvector as well. • However, for a general matrix A, there is no simple relationship between the other Schur vectors of A and eigenvectors of A, as each Schur vector qj satisfies Aqj = AQej = QT ej. • That is, Aqj is a linear combination of q1;:::; qj. • It follows that for j = 1; 2; : : : ; n, the first j Schur vectors q1; q2;:::; qj span an invariant subspace of A. • The Schur vectors are eigenvectors of A when A is a normal matrix, which means that AH A = AAH . • Any real symmetric or skew-symmetric matrix (A = −AT ), for example, is normal. • It can be shown that in this case, the normalized eigenvectors of A form an orthonormal basis n for C . • It follows that if λ1; λ2; : : : ; λn are the eigenvalues of A, with corresponding (orthonormal) eigenvectors q1; q2;:::; qn, then we have AQ = QD; Q = q1 ··· qn ;D = diag(λ1; : : : ; λn): • Because Q is a unitary matrix, it follows that QH AQ = QH QD = D; and A is similar to a diagonal matrix. • We therefore say that A is diagonalizable. 2 • Furthermore, because D can be obtained from A by a similarity transformation involving a unitary matrix, we say that A is unitarily diagonalizable. • Even if A is not a normal matrix, it may still be diagonalizable, meaning that there exists an invertible matrix P such that P −1AP = D, where D is a diagonal matrix. • If this is the case, then, because AP = PD, the columns of P are eigenvectors of A, and the rows of P −1 are eigenvectors of AT (as well as the left eigenvectors of A, if P is real). • Recall that the Matlab function eig, when used with only one output argument, returns a column vector containing the eigenvalues of a given matrix. • The usage [X,D]=eig(A) returns a matrix X, the columns of which contain the eigenvectors of A, and a diagonal matrix D containing the eigenvalues of A. • These matrices satisfy the relationship A*X = X*D. • Which kinds of non-normal matrices are diagonalizable? • To help answer this question, suppose that λ1 and λ2 are distinct eigenvalues, with corre- sponding eigenvectors x1 and x2, respectively. • Furthermore, suppose that x1 and x2 are linearly dependent. • This means that they must be parallel; that is, there exists a nonzero constant c such that x2 = cx1. • This implies that Ax2 = λ2x2 and Ax2 = cAx1 = cλ1x1 = λ1x2. • However, because λ1 6= λ2, this is a contradiction. Therefore, x1 and x2 must be linearly independent. • It follows that if A has n distinct eigenvalues, then it has a set of n linearly independent eigenvectors. • If X is an n × n matrix whose columns are these eigenvectors, then AX = XD, where D is a diagonal matrix of the eigenvectors. • Because the columns of X are linearly independent, X is invertible, and therefore X−1AX = D, and A is diagonalizable. • Now, suppose that the eigenvalues of A are not distinct; that is, the characteristic polynomial has repeated roots. • Then an eigenvalue with multiplicity m does not necessarily correspond to m linearly inde- pendent eigenvectors. • To facilitate discussion of repeated eigenvalues, we introduce some terminology. n×n n Definition. Let A 2 C , and let λ 2 C be an eigenvalue of A. The subspace of C spanned by all eigenvectors of A corresponding to λ is called an eigenspace. n×n • By definition, an eigenvalue of A 2 C corresponds to at least one eigenvector. 3 • Because any nonzero scalar multiple of an eigenvector is also an eigenvector, an eigenvector n defines a one-dimensional invariant subspace of C . • A repeated eigenvalue may have more than one linearly independent eigenvector; the set of all such eigenvectors is a basis of its eigenspace. • Any invariant subspace of a diagonalizable matrix A is a union of eigenspaces. Definition. The algebraic multiplicity of an eigenvalue λ is the number of times that λ occurs as a root of the characteristic polynomial. The geometric multiplicity of λ is the dimension of the eigenspace corresponding to λ, which is equal to the maximal size of a set of linearly independent eigenvectors corresponding to λ. • The geometric multiplicity of an eigenvalue λ is always less than or equal to its algebraic multiplicity. • When it is strictly less, then we say that the eigenvalue is defective. • The matrix A is said to be defective if any of its eigenvalues are defective. • When both multiplicities are equal to one, then we say that the eigenvalue is simple. n×n n • Suppose that A 2 C is defective, which means its eigenvectors do not form a basis for C . That means A is not diagonalizable. • Since there is no similarity transformation that would reduce A to diagonal form, is there at least a similarity transformation that would reduce A to block diagonal form? • The Jordan Canonical Form of an n×n matrix A is a decomposition that yields information about the eigenspaces of A, as well as the multiplicities of the eigenvalues. • It has the form A = XJX−1 where J has the block diagonal structure 2 3 J1 0 ··· 0 6 .. 7 6 0 J2 . 7 J = 6 7 : 6 . .. .. 7 4 . 0 5 0 ··· 0 Jp • Each diagonal block Jp is a Jordan block that has the form 2 3 λi 1 6 .. 7 6 λi . 7 Ji = 6 7 ; i = 1; 2; : : : ; p: 4 λi 1 5 λi • The number of Jordan blocks, p, is equal to the number of linearly independent eigenvectors of A. 4 • The number of Jordan blocks associated with λi is equal to the geometric multiplicity of λi. • The sum of the sizes of these blocks is equal to the algebraic multiplicity of λi. • If A is diagonalizable, then each Jordan block is 1 × 1. Example. Consider a matrix with Jordan Canonical Form 2 2 1 0 3 6 0 2 1 7 6 7 6 0 0 2 7 J = 6 7 : 6 3 1 7 6 7 4 0 3 5 2 The eigenvalues of this matrix are 2, with algebraic multiplicity 4, and 3, with algebraic multiplicity 2. The geometric multiplicity of the eigenvalue 2 is 2, because it is associated with two Jordan blocks. The geometric multiplicity of the eigenvalue 3 is 1, because it is associated with only one Jordan block. Therefore, there are a total of three linearly independent eigenvectors, and the matrix is not diagonalizable. 2 • The Jordan Canonical Form, while very informative about the eigensystem of A, is not practical to compute using floating-point arithmetic. • This is due to the fact that while the eigenvalues of a matrix are continuous functions of its entries, the Jordan Canonical Form is not.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    5 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us