Chapter 6: Complex Matrices We assume that the reader has some experience with matrices and determinants. We can easily extend basic theory of linear algebra by allowing taking complex numbers as matrix entries. However, we should pay more attention to features unique to complex matrices, especially the notion of adjoint, which is the matrix version of complex conjugate. Among the first things we learned from linear algebra is the intimate relation between matrices and linear mappings. To describe this relation within our convention, we need to identify each vector in Cn as a column, that is, as an n 1 matrix. Thus a vector in Cn, × say x = (x1, x2, . , xn), will be considered as the same as x1 x2 x = . [x1 x2 xn]⊤. ≡ ··· xn We are safeguarded from confusion by different types of brackets. From now on, let us adopt the following rule: things in a row surrounded by the round brackets “(” and “)” is the same things arranged in a column surrounded by the square brackets “[” and “]”, e.g. dog (dog, cat) = . We have the following cat “Matrix Representation Theorem” A map T from Cn to Cm is linear if and only if there exists an m n matrix A such that T x = Ax for all x Cn. × ∈ Furthermore, the matrix A here is uniquely determined by T . (Recall that a mapping T from Cn to Cm (we write T : Cn Cm) is linear if the following identity holds → for all vectors x, y in Cn and all scalars α, β: T (αx + βy) = αT x + βT y.) Given a complex matrix A, we define the adjoint of A, denoted by A∗, to be the conjugate transpose of A. In other words, A∗ is obtained by taking complex conjugate of all entries of A, followed by taking the transpose: A∗ = A⊤. Thus a11 a12 a1n a11 a21 an1 a a ··· a a a ··· a 21 22 ··· 2n 12 22 ··· n2 A = = A∗ = . ⇒ am1 am2 amn a1m a2m anm ··· ··· 1 As we have mentioned, the adjoint is the matrix version of the complex conjugate. Example 6.1. Regarding a vector v = (a , a , , a ] in Cn as a matrix, we have 1 2 ··· n a a a a a a a1 1 1 1 2 1 n a a a a ··· a a a2 2 1 2 2 ··· 2 n v = . , v∗ = [a1 a2 an], vv∗ = , . ··· an ana1 ana2 anan ··· 2 2 2 and v∗v = a + a + + a = v, v . | 1| | 2| ··· | n| For n n matrices A and B, and for a complex number α, we have × (A + B)∗ = A∗ + B∗, (αA)∗ = aA∗ (AB)∗ = B∗A∗ The last identity tells us that in general (AB)∗ = A∗B∗ is false. The following identity is the most basic feature concerning the adjoint of a matrix: for every n n matrix A, and all vectors x, y in the complex vector space Cn, we have × Ax, y = x,A∗y We check this identity only for 2 2 matrices. Suppose × a a x y A = 11 12 , x = 1 y = 1 . a a x y 21 22 2 2 Then a11x1 + a12x2 a11y1 + a21y2 Ax = and A∗y = . a x + a x a y + a y 21 1 22 2 12 1 22 2 So Ax, y = a x y + a x y + a x y + a x y and x,A∗y = x a y + x a y + 11 1 1 12 2 1 21 1 2 22 2 2 1 11 1 1 21 2 x a y + x a y . Comparing them, we see that Ax, y = x,A∗y . 2 12 1 2 22 2 We say that an n n matrix is self–adjoint or Hermitian if A∗ = A. The last × identity can be regarded as the matrix version of z = z. So being Hermitian is the matrix analogue of being real for numbers. We say that a matrix A is unitary if A∗A = AA∗ = I, that is, the adjoint A∗ is equal to the inverse of A. The identity A∗A = AA∗ = I is the matrix analogue of zz = 1, or z = 1. Thus, being unitary is a | | matrix analogue of being unit modular for complex numbers. Denote by U(n) the set of 2 all n n unitary matrices. It is easy to check that U(n) forms a group under the usual × matrix multiplication. For example, A, B U(n) implies A∗A = B∗B = I and hence ∈ (AB)(AB)∗ = ABB∗A∗ = AIA∗ = AA∗ = I, etc. The group U(n) is called the unitary group. It plays a basic role in the geometry of the complex vector space Cn. Let A be an n n unitary matrix and denote by v , v ,..., v its column × 1 2 n vectors. Thus we have A = [v1 v2 ... vn] and hence v v1∗v1 v1∗v2 v1∗vn v1, v1 v1, v2 v1, vn 1∗ ··· · · · v v2∗v1 v2∗v2 v2∗vn v2, v1 v2, v2 v2, vn 2∗ ··· · · · A∗ = . A∗A= = . v∗ n vn∗ v1 vn∗ v2 vn∗ vn vn, v1 vn, v2 vn, vn ··· · · · Thus A∗A = I tells us that v , v = δ , meaning that the columns v , v ,..., v j k jk 1 2 n form an orthonormal basis in Cn. We have shown that the columns of a unitary matrix form an orthonormal basis. It turns out that the converse is also true. We have arrived at the following characterization of unitary matrices: An n n matrix is unitary iff its columns form an orthonormal basis in Cn. × Here “iff” stands for “if and only if”, a short hand invented by Paul Halmos. We also have the “real version” of the above statement: A real n n matrix is orthogonal iff its columns × form an orthonormal basis in Rn. Now we give examples of unitary matrices which are used in practice: communication theory (but exactly how they are used is too lengthy to be explained here). Example 6.2. The matrix 1 1 1 1/√2 1/√2 H1 = wih columns v1 = , v2 = √2 1 1 1/√2 1/√2 − − is an orthogonal matrix, since we can check that its columns v1, v2 form an orthonormal 2 basis in R . Now we describe a process to define the Hadamard matrix Hn. Let a a A = 11 12 a a 21 22 be a 2 2 matrix and let B be an n n matrix. We define their tensor product A B × × ⊗ to be the 2n 2n matrix given × a B a B A B = 11 12 . ⊗ a21B a22B 3 We have the following basic identities about tensor products of matrices: aA bB = ab(A B), (A B)∗ = A∗ B∗, (A B)(C D) = AC BD. (6.1) ⊗ ⊗ ⊗ ⊗ ⊗ ⊗ ⊗ A consequence of these identities is: if A and B are unitary (or orthogonal), then so is A B. For example ⊗ 1 1 1 1 1 1 1 1 1 1 1 1 1 1 H2 H1 H1 = = − − ≡ ⊗ 2 1 1 ⊗ 1 1 2 1 1 1 1 − − − − 1 1 1 1 − − We can define Hn inductively by putting 1 Hn 1 Hn 1 Hn = H1 Hn 1 = − − ⊗ − √2 Hn 1 Hn 1 − − − which is a 2n 2n orthogonal matrix, called the Hadamard matrix. We remark that × tensoring is an important operation used in many areas, such as quantum information and quantum computation. Example 6.3. Let ω = e2πi/n. The columns of the following matrix is the orthonor- mal basis of Cn described in Example 5.2 of the last chapter and hence is a unitary matrix: 1 1 1 1 1 1 2 3 4 ··· n 1 1 ω ω ω ω ω − 2 4 6 8 ··· 2(n 1) 1 ω ω ω ω ω − 1 ··· F = √n n 1 2(n 1) 3(n 1) 4(n 1) (n 1)(n 1) 1 ω − ω − ω − ω − ω − − ··· The linear mapping associated with this matrix is called the finite Fourier transform. To speed up this transform by using some special methods is related to saving the cost of communication network in recent years. The rediscovery of so–called FFT (Fast Fourier Transform) has great practical significance. Now the historian can trace back FFT method as early as Gauss. The material in the rest of the present chapter is optional. We say that an n n complex matrix A is orthogonally diagonalizable if there × is an orthonormal basis = e , e ,..., e consisting of eigenvectors of A, that is, E { 1 2 n} 4 for each k, Aek = λkek, where λk is the eigenvalue corresponding to the eigenvector ek. Now we use the basis vectors (considered as columns) in to form the unitary matrix E U = [e1 e2 ... en]. In the next step, we make use of Aek = λkek, but somehow we find it incorrect because we need to consider the scalar λ as a 1 1 matrix, while the k × vector e on its right hand side is n 1. To adjust this, we rewrite λ e as e λ . Thus k × k k k k we have Aek = ekλk. Now the way is clear for the following matrix manipulation: AU = A[e1 e2 ... en] = [Ae1 Ae2 ...Aen] = [e1λ1 e2λ2 ... enλn] = [e1 e2 ... en]D = UD where D is the diagonal matrix given by λ1 0 0 0 0 λ2 0 0 0 0 λ 0 D = 3 0 0 0 λn 1 Thus we have A = UDU − . The above steps can go backward. So we have proved: 1 Fact. A is orthogonally diagonalizable if and only if A = UDU − UDU ∗ for some ≡ unitary U and diagonal D.
Details
-
File Typepdf
-
Upload Time-
-
Content LanguagesEnglish
-
Upload UserAnonymous/Not logged-in
-
File Pages8 Page
-
File Size-