<<

We’ll see soon that if a linear operator on an n- dimensional space has n distinct eigenvalues, then it’s diagonalizable. But first, a preliminary theo- rem. Diagonalizable operators Math 130 Theorem 2. Eigenvectors that are associated to D Joyce, Fall 2013 distinct eigenvalues are independent. That is, if λ1, λ2,..., λk are different eigenvalues of an oper- Some linear operators T : V → V have the nice ator T , and an eigenvector vi is associated to each property that there is some for V so that the eigenvalue λi, then the set of vectors v1, v2,..., vk representing T is a matrix. We’ll is linearly independent. call those operators diagonalizable operators. We’ll Proof. Assume by induction that the first k − 1 of call a A a if it is the eigenvectors are independent. We’ll show all k conjugate to a , that is, there exists of them are. Suppose some linear combination of an P so that P −1AP is a diagonal all k of them equals 0: matrix. That’s the same as saying that under a , A becomes a diagonal matrix. c v + c v + ··· + c v = 0. Reflections are examples of diagonalizable oper- 1 1 2 2 k k ators as are rotations if C is your field of scalars. Take T − λkI of both sides of that equation. The Not all linear operators are diagonalizable. The left side simplifies simplest one is R2 → R2,(x, y) → (y, 0) whose ma- 0 1 trix is A = . No conjugate of it is diagonal. (T − λkI)(c1v1 + ··· + ckvk) 0 0 = c T (v ) − λ c v + ··· + c T (v ) − λ c v It’s an example of a matrix, since some 1 1 k 1 1 k k k k k power of it, namely A2, is the 0-matrix. In general, = c1(λ1 − λk)v1 + ··· + ck(λk − λk)vk nilpotent matrices aren’t diagonalizable. There are = c1(λ1 − λk)v1 + ··· + ck−1(λk−1 − λk)vk−1 many other matrices that aren’t diagonalizable as well. and, of course, the right side is 0. That gives us a linear combination of the first k − 1 vectors which Theorem 1. A linear operator on an n- equals 0, so all their coefficients are 0: dimensional is diagonalizable if and only if it has a basis of n eigenvectors, in which case c1(λ1 − λk) = ··· = ck−1(λk−1 − λk) = 0 the diagonal entries are the eigenvalues for those eigenvectors. Since λk does not equal any of the other λi’s, there- fore all the ci’s are 0: Proof. If it’s diagonalizable, then there’s a basis for which the matrix representing it is diagonal. The c1 = ··· = ck−1 = 0 transformation therefore acts on the ith basis vector th by multiplying it by the i diagonal entry, so it’s an The original equation now says ckvk = 0, and since eigenvector. Thus, all the vectors in that basis are the eigenvector vk is not 0, therefore ck = 0. Thus eigenvectors for their associated diagonal entries. all k eigenvectors are linearly independent. q.e.d. Conversely, if you have a basis of n eigenvectors, then the matrix representing the transformation is Corollary 3. If a linear operator on an n- diagonal since each eigenvector is multiplied by its dimensional vector space has n distinct eigenvalues, associated eigenvalue. q.e.d. then it’s diagonalizable.

1 Proof. Take an eigenvector for each eigenvalue. By the preceding theorem, they’re independent, and since there are n of them, they form a basis of the n-dimensional vector space. The matrix represent- ing the transformation with respect to this basis is diagonal and has the eigenvalues displayed down the diagonal. q.e.d.

Math 130 Home Page at http://math.clarku.edu/~djoyce/ma130/

2