MA251 Algebra I: Advanced Linear Algebra Revision Guide
Written by David McCormick ii MA251 Algebra I: Advanced Linear Algebra
Contents
1 Change of Basis 1
2 The Jordan Canonical Form 1 2.1 Eigenvalues and Eigenvectors ...... 1 2.2 Minimal Polynomials ...... 2 2.3 Jordan Chains, Jordan Blocks and Jordan Bases ...... 3 2.4 Computing the Jordan Canonical Form ...... 4 2.5 Exponentiation of a Matrix ...... 6 2.6 Powers of a Matrix ...... 7
3 Bilinear Maps and Quadratic Forms 8 3.1 Definitions ...... 8 3.2 Change of Variable under the General Linear Group ...... 9 3.3 Change of Variable under the Orthogonal Group ...... 10 3.4 Unitary, Hermitian and Normal Matrices ...... 11
4 Finitely Generated Abelian Groups 12 4.1 Generators and Cyclic Groups ...... 13 4.2 Subgroups and Cosets ...... 13 4.3 Quotient Groups and the First Isomorphism Theorem ...... 14 4.4 Abelian Groups and Matrices Over Z ...... 15
Introduction
This revision guide for MA251 Algebra I: Advanced Linear Algebra has been designed as an aid to revision, not a substitute for it. While it may seem that the module is impenetrably hard, there’s nothing in Algebra I to be scared of. The underlying theme is normal forms for matrices, and so while there is some theory you have to learn, most of the module is about doing computations. (After all, this is mathematics, not philosophy.) Finding books for this module is hard. My personal favourite book on linear algebra is sadly out-of- print and bizarrely not in the library, but if you can find a copy of Evar Nering’s “Linear Algebra and Matrix Theory” then it’s well worth it (though it doesn’t cover the abelian groups section of the course). Three classic seminal books that cover pretty much all first- and second-year algebra are Michael Artin’s “Algebra”, P. M. Cohn’s “Classic Algebra” and I. N. Herstein’s “Topics in Algebra”, but all of them focus on the theory and not on the computation, and they often take a different order (for instance, most do modules before doing JCFs). Your best source of computation questions is old past paper questions, not only for the present module but also for its predecessors MA242 Algebra I and MA245 Algebra II. So practise, practise, PRACTISE, and good luck on the exam!
Disclaimer: Use at your own risk. No guarantee is made that this revision guide is accurate or complete, or that it will improve your exam performance, or that it will make you 20% cooler. Use of this guide will increase entropy, contributing to the heat death of the universe.
Authors Written by D. S. McCormick ([email protected]). Edited by C. I. Midgley ([email protected]) Based upon lectures given by Dr. Derek Holt and Dr. Dmitri˘ıRumynin at the University of Warwick, 2006–2008, and later Dr. David Loeffler, 2011–2012. Any corrections or improvements should be entered into our feedback form at http://tinyurl.com/WMSGuides (alternatively email [email protected]). MA251 Algebra I: Advanced Linear Algebra 1
1 Change of Basis
A major theme in MA106 Linear Algebra is change of bases. Since this is fundamental to what follows, we recall some notation and the key theorem here. Let T : U → V be a linear map between U and V . To express T as a matrix requires picking a basis 0 {ei} of U and a basis {fj} of V . To change between two bases {ei} and {ei} of U, we simply take the 0 identity map IU : U → U and use the basis {ei} in the domain and the basis {ei} in the codomain; the 0 matrix so formed is the change of basis matrix from the basis of eis to the eis. Any such change of basis matrix is invertible — note that this version is the inverse of the one learnt in MA106 Linear Algebra! Proposition 1.1. Let u ∈ U, and let u and u0 denote the column vectors associated with u in the bases 0 0 0 e1,..., en and e1,..., en respectively. Then if P is the change of basis matrix, we have P u = u.
Theorem 1.2. Let A be the matrix of T : U → V with respect to the bases {ei} of U and {fj} of V , and 0 0 let B be the matrix of T with respect to the bases {ei} of U and {fj} of V . Let P be the change of basis 0 0 −1 matrix from {ei} to {ei}, and let Q be the change of basis matrix from {fj} to {fj}. Then B = Q AP . 0 0 Throughout this course we are concerned primarily with U = V , and {ei} = {fj}, {ei} = {fj}, so that P = Q and hence B = P −1AP . In this course, the aim is to find so-called normal forms for matrices and their corresponding linear maps. Given a matrix, we want to know what it does in simple terms, and to be able to compare matrices that look completely different; so, how should we change bases to get the matrix into a nice form? There are three very different answers:
1. When working in a finite-dimensional vector space over C, we can always change bases so that T is as close to diagonal as possible, with only eigenvalues on the diagonal and possibly some 1s on the superdiagonal; this is the Jordan Canonical Form, discussed in section 2. 2. We may want the change of basis not just to get a matrix into a nice form, but also preserve its geometric properties. This leads to the study of bilinear and quadratic forms, and along with it the theory of orthogonal matrices, in section 3.
3. Alternatively, we may consider matrices with entries in Z, and try and diagonalise them; this leads, perhaps surprisingly, to a classification of all finitely-generated abelian groups, which (along with some basic group theory) is the subject of section 4.
2 The Jordan Canonical Form 2.1 Eigenvalues and Eigenvectors We first recall some facts on eigenvalues and eigenvectors from MA106 Linear Algebra. Definition 2.1. Let V be a vector space over K and let T : V → V be a linear map, with associated matrix A. If T (v) = λv for some λ ∈ K and v ∈ V with v 6= 0, then λ is an eigenvalue of T (and of A), and v a corresponding eigenvector of T (and of A). We call the subspace {v ∈ V | T (v) = λv} the eigenspace of T with respect to λ.
Definition 2.2. For an n × n matrix A, cA(x) := det(A − xIn) is the characteristic polynomial of A.
Theorem 2.3. Let A be an n × n matrix. Then λ is an eigenvalue of A if and only if det(A − λIn) = 0. Recall that n × n matrices A and B are similar if there is an invertible n × n matrix P such that B = P −1AP . Since similar matrices have the same characteristic equation, changing bases does not change the eigenvalues of a linear map. You have already seen one “normal form”, which occurs when the matrix has distinct eigenvalues: Theorem 2.4. Let T : V → V be a linear map. Then the matrix of T is diagonal with respect to some basis of V if and only if V has a basis consisting of eigenvectors of T . 2 MA251 Algebra I: Advanced Linear Algebra
Theorem 2.5. Let λ1, . . . , λr be distinct eigenvalues of a linear map T : V → V and let v1,..., vr be the corresponding eigenvectors. Then v1,..., vr are linearly independent. Corollary 2.6. If the linear map T : V → V has n distinct eigenvalues, where dim V = n, then T is diagonalisable.
Of course, the converse is not true; T may be diagonalisable even though it has repeated eigenvalues.
2.2 Minimal Polynomials We denote the set of all polynomials in a single variable x with coefficients in a field K by K[x]. We recall some properties of polynomials from MA132 Foundations; these often resemble properties of Z. We write a | b to mean a divides b; e.g. (x − 4) | (x2 − 3x − 4). Given two polynomials p, q 6= 0, we can divide with remainder, where the remainder has degree less than p. For example, if p = x2 − 4x and q = x3 + 2x2 + 5, then q = sp + r, where s = x + 6 and r = 24x + 5. This is the Euclidean algorithm.
Definition 2.7. A polynomial in K[x] is called monic if the coefficient of the highest power of x is 1.
Definition 2.8. The greatest common divisor of p, q ∈ K[x] is the unique monic polynomial r such that r | p and r | q, and for any other polynomial r0 such that r0 | p and r0 | q, we have r0 | r. Similarly, the lowest common multiple of p, q ∈ K[x] is the unique monic polynomial r such that p | r and q | r, and for any other polynomial r0 such that p | r0 and q | r0, we have r | r0.
We first observe a very interesting fact about characteristic polynomials:
Theorem 2.9 (Cayley–Hamilton Theorem). Let cA(x) be the characteristic polynomial of an n × n matrix A over an arbitrary field K. Then cA(A) = 0. So we know that there is at least some polynomial p ∈ K[x] such that p(A) = 0. The following theorem allows us to define more:
Theorem 2.10. Let A be an n × n matrix over K representing the linear map T : V → V . Then there is a unique monic non-zero polynomial p ∈ K[x] with minimal degree such that p(A) = 0. Furthermore, if q ∈ K[x] also satisfies q(A) = 0, then p | q.
Proof. We can assume such a polynomial is monic. The Cayley–Hamilton Theorem tells us that there is p ∈ K[x] such that p(A) = 0. If there were two distinct polynomials p1, p2 of minimal degree s.t. p1(A) = p2(A) = 0, then p = p1 −p2 would be non-zero and of lower degree, contradicting minimality. Thus p is unique. Furthermore, suppose q(A) = 0 but p - q. Then we can write q = sp + r, with deg(r) < deg(p), and r 6= 0. But then r(A) = q(A) − s(A)p(A) = 0, contradicting minimality of p.
Definition 2.11. The unique monic non-zero polynomial µA(x) of minimal degree with µA(A) = 0 is called the minimal polynomial of A, or of the corresponding linear map T .
Combining the last two theorems we observe that µA(x) divides cA(x). Furthermore, similar matrices have the same minimal polynomial, so the minimal polynomial of a linear map does not depend on bases. Similarly to above we may define µAv to be the unique monic polynomial p of minimal degree such that p(T )(v) = 0; since p(T ) = 0 if and only if p(T )(v) = 0 for all v ∈ V , µA is the least common multiple of µAv for v ∈ V . In fact, we only need consider vectors in a basis of V ; i.e. if {e1,..., en} is
a basis of V then µA = lcm{µAei : 1 ≤ i ≤ n}. This allows us to calculate µA: for v ∈ V , we compute 2 µAv by calculating v, T (v), T (v), and so on until the sequence becomes linearly dependent.
5 0 −1 T T Example 2.12. Let K = R and consider A = 3 4 −3 . Let e1 = (1, 0, 0) , e2 = (0, 1, 0) , e3 = 1 0 3 (0, 0, 1)T be the standard basis of R3. Then: T 2 T 2 • Ae1 = (5, 3, 1) , A e1 = (24, 24, 8) = 8Ae1 − 16e1, so (A − 8A + 16)e1 = 0, thus µAe1 (x) = x2 − 8x + 16 = (x − 4)2. T • Ae2 = (0, 4, 0) = 4e2, so (A − 4)e2 = 0, thus µAe2 (x) = (x − 4). T 2 T 2 2 • Ae3 = (−1, −3, 3) , A e3 = (−8, −24, 8) = 8Ae3 − 16e3, thus µAe3 (x) = x − 8x + 16 = (x − 4) . 2 3 Thus µA = lcm{µAe1 , µAe2 , µAe3 } = (x − 4) . One may compute that cA(x) = det(A − xI) = (4 − x) . MA251 Algebra I: Advanced Linear Algebra 3
Lemma 2.13. (x − λ) divides the minimal polynomial µA(x) if and only if λ is an eigenvalue of A.
Proof. Suppose (x − λ) | µA(x); then as µA(x) | cA(x), we have (x − λ) | cA(x), and hence λ is an eigenvalue of A. Conversely, if λ is an eigenvalue of A then there exists v 6= 0 such that (A − λI)v = 0, hence µAv = (x − λ), and since µA = lcm{µAv : v ∈ V } we have (x − λ) | µA(x).
2.3 Jordan Chains, Jordan Blocks and Jordan Bases
We assume henceforth that K = C, so that all polynomials in K[x] factorise into linear factors (by the Fundamental Theorem of Algebra). We now seek to generalise our notions of eigenvalue and eigenvector in order to be able to find a “normal form” for a matrix with any eigenvalues, not just distinct ones.
n,1 Definition 2.14. A Jordan chain of length k is a sequence of non-zero vectors v1,..., vk ∈ C (that is, column vectors of length n with entries in C) that satisfies
Av1 = λv1 and Avi = λvi + vi−1 for 2 ≤ i ≤ k
for some eigenvalue λ of A. Equivalently, (A − λIn)v1 = 0 and (A − λIn)vi = vi−1 for 2 ≤ i ≤ k, so i (A − λIn) vi = 0 for 1 ≤ i ≤ k.
i Definition 2.15. A non-zero vector v ∈ V such that (A − λIn) v = 0 for some i > 0 is called a i generalised eigenvector of A with respect to the eigenvalue λ. The set {v ∈ V |(A − λIn) v = 0} is called i the generalised eigenspace of index i of A with respect to λ; it is the nullspace of (A − λIn) . (Note that when i = 1, these definitions reduce to ordinary eigenvectors and eigenspaces.)
3 1 0 3,1 For example, consider the matrix A = 0 3 1 . For the standard basis e1, e2, e3 of , we have 0 0 3 C Ae1 = 3e1, Ae2 = 3e2 + e1, Ae3 = 3e3 + e2. Thus e1, e2, e3 is a Jordan chain of length 3 for the eigenvalue 3 of A. The generalised eigenspaces of index 1, 2 and 3 respectively are he1i, he1, e2i, and he1, e2, e3i. i Note that the dimension of a generalised eigenspace of A is the nullity of (T − λIV ) , which depends only on the linear map T associated with A; thus the dimensions of corresponding eigenspaces of similar matrices are the same.
Definition 2.16. A Jordan block with eigenvalue λ of degree k is the k × k matrix Jλ,k = (γij) where γii = λ for 1 ≤ i ≤ k, γi,i+1 = 1 for 1 ≤ i < k, and γij = 0 if j 6= i, i + 1. For example, the following are Jordan blocks:
2 − i 1 0 2 1 J = ,J = 0 2 − i 1 . 2,2 0 2 (2−i),3 0 0 2 − i
n,1 It is a fact that the matrix A of T with respect to the basis v1,..., vn of C is a Jordan block of degree n if and only if v1,..., vn is a Jordan chain for A. k Note that the minimal polynomial of Jλ,k is µJλ,k (x) = (x − λ) , and its characteristic polynomial is k cJλ,k (x) = (λ − x) .
Definition 2.17. A Jordan basis for A is a basis of Cn,1 which is a union of disjoint Jordan chains. For an m × m matrix A and an n × n matrix B, we can form the (m + n) × (m + n) matrix A 0 A ⊕ B = m,n . For example, 0n,m B 1 2 0 0 1 2 2 3 0 1 0 0 ⊕ = . 0 1 4 −1 0 0 2 3 0 0 4 −1
Suppose A has eigenvalues λ1, . . . , λr, and suppose wi,1,..., wi,ki is a Jordan chain for A for the
eigenvalue λi, such that w1,1,..., w1,k1 , w2,1,..., w2,k2 ,..., wr,1,..., wr,kr is a Jordan basis for A. Then 4 MA251 Algebra I: Advanced Linear Algebra the matrix of the linear map T corresponding to A with respect to this Jordan basis is the direct sum of
Jordan blocks Jλ1,k1 ⊕ Jλ2,k2 ⊕ · · · ⊕ Jλr ,kr . The main theorem of this section is that we can always find a Jordan matrix for any n × n matrix A over C; the corresponding matrix which is a direct sum of the Jordan blocks is called the Jordan canonical form of A:
Theorem 2.18. Let A be an n × n matrix over C. Then there exists a Jordan basis for A, and hence A is similar to a matrix J which is a direct sum of Jordan blocks. The Jordan blocks occurring in J are uniquely determined by A, so J is uniquely determined up to the order of the blocks. J is said to be the Jordan canonical form of A. The proof of this theorem is hard and non-examinable. What’s far more important is calculating the Jordan canonical form (JCF) of a matrix, and the matrix P whose columns are the vectors of the Jordan basis; then by theorem 1.2, we have that P −1AP = J. For 2 × 2 and 3 × 3 matrices, the JCF of a matrix is in fact determined solely by its minimal and characteristic polynomials. In higher dimensions, we must consider the generalised eigenspaces.
2.4 Computing the Jordan Canonical Form
Suppose A is an n×n matrix with eigenvalues λ1, . . . , λr, and that the Jordan blocks for eigenvalue λi are J ,...,J , where k ≥ k ≥ · · · ≥ k . Then the characteristic polynomial of J (and hence λi,ki,1 λi,ki,ji i,1 i,2 i,ji Qr ki of A) is the product of the characteristic polynomials of the Jordan blocks; thus cJ (x) = i=1(λi − x) , where ki = ki,1 + ··· + ki,ji ; i.e. each (λi − x) occurs raised to the power of the sum of the sizes of the Jordan blocks of that eigenvalue. The minimal polynomial of J (hence of A) is the least common multiple of the minimal polynomials Qr ki,1 of the Jordan blocks; since we have arranged them in descending order of size, µJ (x) = i=1(x − λi) ; i.e. each (x − λi) occurs raised to the power of the biggest Jordan block for that eigenvalue. In 2 and 3 dimensions, this restricts the possible Jordan blocks enough to determine the JCF solely by looking at the minimal and characteristic polynomials. We must then determine the Jordan basis; note that it is often easier to find the vectors in a Jordan chain in reverse order.
2.4.1 2 × 2 Matrices For a 2 × 2 matrix, there are two possibilities for its characteristic polynomial; it must either have two 2 distinct roots, e.g. (λ1 − x)(λ2 − x), or it must have one repeated root, (λ1 − x) . By lemma 2.13, in the first case the minimal polynomial must be (x − λ1)(x − λ2), and the only possibility is one Jordan block for each eigenvalue of size 1. (This accords with corollary 2.6.) In the second case, we can have 2 µA(x) = (λ1 − x) or µA(x) = (λ1 − x) , which correspond to two Jordan blocks of size 1 and one Jordan block of size 2 for the only eigenvalue. (In fact, when we have two Jordan blocks of size 1 for λ 0 the same eigenvalue, the JCF is just a scalar matrix J = 0 λ which commutes with all matrices, thus A = PJP −1 = J, i.e. A is its own JCF.) Table 1 summarises the possibilities.
Characteristic Polynomial Minimal Polynomial Jordan Canonical Form (λ − x)(λ − x) (x − λ )(x − λ ) J ⊕ J = λ1 0 1 2 1 2 λ1,1 λ2,1 0 λ2 (x − λ )2 J = λ1 1 2 1 λ1,2 0 λ1 (λ1 − x) (x − λ ) J ⊕ J = λ1 0 1 λ1,1 λ1,1 0 λ1
Table 1: The possible JCFs of a 2 × 2 matrix.
2 4 Example 2.19. A = ( 5 3 ) has characteristic polynomial cA(x) = (x + 2)(x − 7), so A has eigenvalues −2 0 T −2 and 7, and thus the JCF is J = 0 7 . An eigenvector for the eigenvalue −2 is (1, −1) , and an T 1 4 −1 eigenvector for the eigenvalue 7 is (4, 5) ; setting P = −1 5 , we may calculate that P AP = J. 2 1 2 2 Example 2.20. A = −1 4 has cA(x) = (3 − x) , and one may calculate µA(x) = (x − 3) . Thus its 3 1 JCF is J = ( 0 3 ). To find a Jordan basis we choose any v2 such that (A − 3I)v2 6= 0, and then choose MA251 Algebra I: Advanced Linear Algebra 5
T T T v1 = (A − 3I)v2; for example v2 = (0, 1) has (A − 3I)v2 = (1, 1) , so set v1 = (1, 1) ; then Av1 = 3v1 1 0 −1 and Av2 = 3v2 + v1 as required. Thus setting P = ( 1 1 ), we may calculate that P AP = J.
2.4.2 3 × 3 Matrices For 3 × 3 matrices, we can do the same kind of case analysis that we did for 2 × 2 matrices. It is a very good test of understanding to go through and derive all the possibilities for yourself, so DO IT NOW! Once you have done so, turn to the next page and check the results in table 2.
Characteristic Polynomial Minimal Polynomial Jordan Canonical Form
λ1 0 0 0 λ 0 (λ1 − x)(λ2 − x)(λ3 − x) (x − λ1)(x − λ2)(x − λ3) Jλ1,1 ⊕ Jλ2,1 ⊕ Jλ3,1 = 2 0 0 λ3 λ1 1 0 2 0 λ 0 2 (x − λ1) (x − λ2) Jλ1,2 ⊕ Jλ2,1 = 1 (λ1 − x) (λ2 − x) 0 0 λ2 λ1 0 0 0 λ 0 (x − λ1)(x − λ2) Jλ1,1 ⊕ Jλ1,1 ⊕ Jλ2,1 = 1 0 0 λ2 λ1 1 0 3 0 λ 1 (x − λ1) Jλ1,3 = 1 3 0 0 λ1 (λ1 − x) λ1 1 0 2 0 λ 0 (x − λ1) Jλ1,2 ⊕ Jλ1,1 = 1 0 0 λ1 λ1 0 0 0 λ 0 (x − λ1) Jλ1,1 ⊕ Jλ1,1 ⊕ Jλ1,1 = 1 0 0 λ1
Table 2: The possible JCFs of a 3 × 3 matrix.
5 0 −1 3 2 Example 2.21. Consider A = 3 4 −3 . We previously calculated cA(x) = (4 − x) , µA(x) = (x − 4) . 1 0 3 4 1 0 This tells us that the JCF of A is J = 0 4 0 . There are two Jordan chains, one of length 2 and one 0 0 4 of length 1. For the first we need v1, v2 such that (A − 4I)v2 = v1, and (A − 4I)v1 = 0. We calculate 1 0 −1 A − 4I = 3 0 −3 . The minimal polynomial tells us that (A − 4I)2v = 0 for all v ∈ V , so we can choose 1 0 −1 T T whatever we like for v2; say v2 = (1, 0, 0) ; then v1 = (A − 4I)v2 = (1, 3, 1) . For the second chain we T need an eigenvector v3 which is linearly independent of v1; v3 = (1, 0, 1) is as good as any. Setting 1 1 1 P = 3 0 0 we find J = P −1AP . 1 0 1
4 −1 −1 3 Example 2.22. Consider A = −4 9 4 . One may tediously compute that cA(x) = (3 − x) , and 7 −10 −4 that 1 −1 −1 −2 3 2 2 3 A − 3I = −4 6 4 , (A − 3I) = 0 0 0 , (A − 3I) = 0. 7 −10 −7 −2 3 2
3 Thus µA(x) = (x − 3) . Thus we have one Jordan chain of length 3; that is, we need nonzero vectors v1, v2, v3 such that (A − 3I)v3 = v2,(A − 3I)v2 = v1, and (A − 3I)v1 = 0. For v3, we need (A − 3I)v3 2 T T and (A − 3I) v3 to be nonzero; we may choose v3 = (1, 1, 0) ; we can then compute v2 = (0, 2, −3) T 1 0 1 −1 3 1 0 and v1 = (1, 0, 1) . Putting P = 0 2 1 , we obtain P AP = 0 3 1 . 1 −3 0 0 0 3
2.4.3 Higher Dimensions: The General Case For dimensions higher than 3, the characteristic polynomial and minimal polynomial do not always determine the JCF uniquely. In 4 dimensions, for example, Jλ,2 ⊕ Jλ,2 and Jλ,2 ⊕ Jλ,1 ⊕ Jλ,1 both have 4 2 cA(x) = (λ − x) and µA(x) = (x − λ) . In general, we can compute the JCF from the dimensions of the generalised eigenspaces, as follows:
Theorem 2.23. Let λ be an eigenvalue of A and let J be the JCF of A. Then: (i) The number of Jordan blocks of J with eigenvalue λ is equal to nullity(A − λIn). 6 MA251 Algebra I: Advanced Linear Algebra
(ii) More generally, for i > 0, the number of Jordan blocks of J with eigenvalue λ and degree at least i i−1 i is equal to nullity((A − λIn) ) − nullity((A − λIn) ). (Recall that nullity(T ) = dim(ker(T )).) The proof of this need not be learnt, but the theorem is vital as a tool for calculating JCFs, as the following example shows. 1 0 −1 1 0 ! −4 1 −3 2 1 5 Example 2.24. Let A = −2 −1 0 1 1 . One may tediously compute that cA(x) = (2 − x) , and −3 −1 −3 4 1 −8 −2 −7 5 4 that −1 0 −1 1 0 ! 0 0 0 0 0 ! −4 −1 −3 2 1 0 0 0 0 0 A − 2I = −2 −1 −2 1 1 , (A − 2I)2 = −1 0 −1 1 0 , (A − 2I)3 = 0. −3 −1 −3 2 1 −1 0 −1 1 0 −8 −2 −7 5 2 −1 0 −1 1 0 3 th This gives that µA(x) = (x − 2) . Let rj denote the j row of (A − 2I); then one may observe that r4 = r1 + r3, and r5 = 2r1 + r2 + r3, but that r1, r2, r3 are linearly independent, so rank(A − 2I) = 3, and thus by the dimension theorem nullity(A − 2I) = 5 − 3 = 2. Thus there are two Jordan blocks for eigenvalue 2. Furthermore, it is clear that rank(A − 2I)2 = 1 and hence nullity(A − 2I)2 = 4, so there are 4 − 2 = 2 blocks of size at least 2. As nullity(A − 2I)3 = 5, we have 5 − 4 = 1 block of size at least 3. Since the largest block has size 3 (by the minimal polynomial), we now know that there are two Jordan blocks, one of size 3 and one of size 2. To find the Jordan chains, we need v1, v2, v3, v4, v5 such that
(A − 2I)v5 = v4, (A − 2I)v4 = 0;(A − 2I)v3 = v2, (A − 2I)v2 = v1, (A − 2I)v1 = 0.
T T For the chain of length 3, we may choose v3 = (0, 0, 0, 1, 0) , since then v2 = (A−2I)v3 = (1, 2, 1, 2, 5) 6= 2 T 0 and v1 = (A − 2I) v3 = (0, 0, 1, 1, 1) 6= 0. For the chain of length 2, we must choose v5 so that 2 v4 = (A − 2I)v5 6= 0, but so that (A − 2I) v5 = 0, and so that all the vi are linearly independent. In T general there is no easy way of doing this; we choose v5 = (−1, 0, 1, 0, 0) , so that v4 = (A − 2I)v5 = 0 1 0 0 −1 ! 2 1 0 0 0 ! T 0 2 0 1 0 −1 0 2 1 0 0 (0, 1, 0, 0, 1) . Then, setting P = 1 1 0 0 1 , we find J = P AP = 0 0 2 0 0 . 1 2 1 0 0 0 0 0 2 1 1 5 0 1 0 0 0 0 0 2
2.5 Exponentiation of a Matrix In this section, we define eA where A is a matrix.
Definition 2.25. If A ∈ Cn,n, we define eA to be the infinite series ∞ A2 A3 X Ak eA = I + A + + + ... = n 2 6 k! k=1
Warning: It is not in general true that eA+B = eAeB — though this does hold if AB = BA.
Lemma 2.26. 1. Let A and B ∈ Cn,n be similar, so B = P −1AP for some invertible matrix P . Then eB = P −1eAP . d At At 2. dt e = Ae The first point on this lemma gives us a hint as to how we might compute the exponential of a matrix A A1 An — using the Jordan form! Given A = A1 ⊕ ... ⊕ An, we have that e = e ⊕ ... ⊕ e , so it will suffice to consider exponentiation of a single Jordan block.
Jt Theorem 2.27. If J = Jλ,s is a Jordan block, then e is the matrix whose (i, j) entry is given by
( tj−ieλt (j−i)! if j ≥ i 0 if j < i
2 1 0 e2t te2t 0 Example 2.28. Given J = 0 2 0, we have eJt = 0 e2t 0 0 0 1 0 0 et MA251 Algebra I: Advanced Linear Algebra 7
We can use this to compute the solution to differential equations. Check your notes for MA133 Differential Equations for solution methods, remembering that we can now take exponents and powers of matrices directly. We can also use a slight variation on this method to find the solution to difference equations, using matrix powers.
2.6 Powers of a Matrix Na¨ıvely, we can use a similar strategy to exponentiation, using the Jordan canonical form. Observe that if A = PJP −1, where J is the Jordan form of A, then An = PJ nP −1. Again, it suffices to consider only a single Jordan block:
n Theorem 2.29. If J = Jλ,s is a Jordan block, then J is the matrix whose (i, j) entry is given by
( n n−(j−i) j−i λ 0 if j < i