Canonical Forms

Total Page:16

File Type:pdf, Size:1020Kb

Canonical Forms C H A P T E R 8 Canonical Forms Recall that at the beginning of Section 7.5 we stated that a canonical form for T ∞ L(V) is simply a representation in which the matrix takes on an especially simple form. For example, if there exists a basis of eigenvectors of T, then the matrix representation will be diagonal. In this case, it is then quite trivial to assess the various properties of T such as its rank, determinant and eigenval- ues. Unfortunately, while this is generally the most desirable form for a matrix representation, it is also generally impossible to achieve. We now wish to determine certain properties of T that will allow us to learn as much as we can about the possible forms its matrix representation can take. There are three major canonical forms that we will consider in this chapter : triangular, rational and Jordan. (This does not count the Smith form, which is really a tool, used to find the rational and Jordan forms.) As we have done before, our approach will be to study each of these forms in more than one way. By so doing, we shall gain much insight into their meaning, as well as learning additional techniques that are of great use in various branches of mathematics. 8.1 ELEMENTARY CANONICAL FORMS In order to ease into the subject, this section presents a simple and direct method of treating two important results: the triangular form for complex matrices and the diagonalization of normal matrices. To begin with, suppose 382 8.1 ELEMENTARY CANONICAL FORMS 383 that we have a matrix A ∞ Mñ(ç). We define the adjoint (or Hermitian adjoint) of A to be the matrix A¿ = A*T. In other words, the adjoint of A is its complex conjugate transpose. From Theorem 3.18(d), it is easy to see that (AB)¿ = B¿A¿ . If it so happens that A¿ = A, then A is said to be a Hermitian matrix. If a matrix U ∞ Mñ(ç) has the property that U¿ = Uî, then we say that U is unitary. Thus a matrix U is unitary if UU¿ = U¿U = I. (Note that by Theorem 3.21, it is only necessary to require either UU¿ = I or U¿U = I.) We also see that the product of two unitary matrices U and V is unitary since (UV)¿UV = V¿U¿UV = V¿IV = V¿V = I. If a matrix N ∞ Mñ(ç) has the property that it commutes with its adjoint, i.e., NN¿ = N¿N, then N is said to be a normal matrix. Note that Hermitian and unitary matrices are auto- matically normal. Example 8.1 Consider the matrix A ∞ Mì(ç) given by 1 "1 !1% A = $ '!!. 2 #i !!i& Then the adjoint of A is given by † 1 "!!1 !i% A = $ ' 2 #!1 !i& and we leave it to the reader to verify that AA¿ = A¿A = 1, and hence show that A is unitary. ∆ We will devote considerable time in Chapter 10 to the study of these matrices. However, for our present purposes, we wish to point out one impor- tant property of unitary matrices. Note that since U ∞ Mñ(ç), the rows Uá and columns Ui of U are just vectors in çn. This means that we can take their inner product relative to the standard inner product on çn (see Example 2.9). Writing out the relation UU¿ = I in terms of components, we have n n n † † (UU )ij = !uiku kj = !uiku jk* = !u jk*uik = U j ,!Ui = "ij k=1 k=1 k=1 and from U¿U = I we see that n n † † i j (U U)ij = !u ikukj = !uki*ukj = U ,!U = "ij !!. k=1 k=1 384 CANONICAL FORMS In other words, a matrix is unitary if and only if its rows (or columns) each form an orthonormal set. Note we have shown that if the rows (columns) of U ∞ Mñ(ç) form an orthonormal set, then so do the columns (rows), and either of these is a sufficient condition for U to be unitary. For example, the reader can easily verify that the matrix A in Example 8.1 satisfies these conditions. It is also worth pointing out that Hermitian and unitary matrices have important analogues over the real number system. If A ∞ Mñ(®) is Hermitian, then A = A¿ = AT, and we say that A is symmetric. If U ∞ Mñ(®) is unitary, then Uî = U¿ = UT, and we say that U is orthogonal. Repeating the above calculations over ®, it is easy to see that a real matrix is orthogonal if and only if its rows (or columns) form an orthonormal set. It will also be useful to recall from Section 3.6 that if A and B are two matrices for which the product AB is defined, then the ith row of AB is given by (AB)á = AáB and the ith column of AB is given by (AB)i = ABi. We now prove yet another version of the triangular form theorem. Theorem 8.1 (Schur Canonical Form) If A ∞ Mñ(ç), then there exists a unitary matrix U ∞ Mñ(ç) such that U¿AU is upper-triangular. Furthermore, the diagonal entries of U¿AU are just the eigenvalues of A. Proof If n = 1 there is nothing to prove, so we assume that the theorem holds for any square matrix of size n - 1 ˘ 1, and suppose A is of size n. Since we are dealing with the algebraically closed field ç, we know that A has n (not necessarily distinct) eigenvalues (see Section 7.3). Let ¬ be one of these eigenvalues, and denote the corresponding eigenvector by Uÿ1. By Theorem 2.10 we extend Uÿ1 to a basis for çn, and by the Gram-Schmidt process (Theorem 2.21) we assume that this basis is orthonormal. From our discussion above, we see that this basis may be used as the columns of a unitary matrix Uÿ with Uÿ1 as its first column. We then see that (U! †AU! )1 = U! †(AU! )1 = U! †(AU! 1) = U! †(!U! 1) = !(U! †U! 1) ! † ! 1 1 = !(U U) = !I and hence Uÿ¿AUÿ has the form " ! *!!"!!* % $ ' $ 0 ! ! ' U! †AU! = $ # ! B ! ' $ ' $ 0 !! !! ' # & 8.1 ELEMENTARY CANONICAL FORMS 385 where B ∞ Mn-1(ç) and the *’s are (in general) nonzero scalars. By our induc- tion hypothesis, we may choose a unitary matrix W ∞ Mn-1(ç) such that W¿BW is upper-triangular. Let V ∞ Mñ(ç) be a unitary matrix of the form ! 1 0!!!!!0 $ # & # 0 ! ! & V = # " ! W ! & # & # 0 !! !! & " % and define the unitary matrix U = UÿV ∞ Mñ(ç). Then U¿AU = (UÿV)¿A(UÿV) = V¿(Uÿ¿AUÿ)V is upper-triangular since (in an obvious shorthand notation) ! 1 0 $!' *$!1 0 $ ! 1 0 $!' * $ V †(U! †AU! )V = # † &# &# & = # † &# & " 0 W %"0 B%"0 W % " 0 W %"0 BW % !' * $ = # & 0 W †BW " % and W¿BW is upper-triangular by the induction hypothesis. Noting that ¬I - U¿AU is upper-triangular, it is easy to see (using Theorem 4.5) that the roots of det(¬I - U¿AU) are just the diagonal entries of U¿AU. But det(¬I - U¿AU) = det[U¿(¬I - A)U] = det(¬I - A) so that A and U¿AU have the same eigenvalues. ˙ Corollary If A ∞ Mñ(®) has all its eigenvalues in ®, then the matrix U defined in Theorem 8.1 may be chosen to have all real entries. Proof If ¬ ∞ ® is an eigenvalue of A, then A - ¬I is a real matrix with deter- minant det(A - ¬I) = 0, and therefore the homogeneous system of equations (A - ¬I)X = 0 has a real solution. Defining Uÿ1 = X, we may now proceed as in Theorem 8.1. The details are left to the reader (see Exercise 8.8.1). ˙ We say that two matrices A, B ∞ Mñ(ç) are unitarily similar (written A – B) if there exists a unitary matrix U such that B = U¿AU = UîAU. Since this 386 CANONICAL FORMS defines an equivalence relation on the set of all matrices in Mñ(ç), many authors say that A and B are unitarily equivalent. However, we will be using the term “equivalent” in a somewhat more general context later in this chapter, and the word “similar” is in accord with our earlier terminology. We leave it to the reader to show that if A and B are unitarily similar and A is normal, then B is also normal (see Exercise 8.8.2). In particular, suppose that U is unitary and N is such that U¿NU = D is diagonal. Since any diagonal matrix is automatically normal, it follows that N must be normal also. We now show that the converse is also true, i.e., that any normal matrix is unitarily similar to a diagonal matrix. To see this, suppose N is normal, and let U¿NU = D be the Schur canoni- cal form of N. Then D is both upper-triangular and normal (since it is unitarily similar to a normal matrix). We claim that the only such matrices are diagonal. For, consider the (1, 1) elements of DD¿ and D¿D. From what we showed above, we have (DD¿)èè = ÓDè, DèÔ = \dèè\2 + \dèì\2 + ~ ~ ~ + \dèñ\2 and (D¿D)èè = ÓD1, D1Ô = \dèè\2 + \dìè\2 + ~ ~ ~ + \dñè\2 . But D is upper-triangular so that dìè = ~ ~ ~ = dñè = 0. By normality we must have (DD¿)èè = (D¿D)èè, and therefore dèì = ~ ~ ~ = dèñ = 0 also. In other words, with the possible exception of the (1, 1) entry, all entries in the first row and column of D must be zero.
Recommended publications
  • 3·Singular Value Decomposition
    i “book” — 2017/4/1 — 12:47 — page 91 — #93 i i i Embree – draft – 1 April 2017 3 Singular Value Decomposition · Thus far we have focused on matrix factorizations that reveal the eigenval- ues of a square matrix A n n,suchastheSchur factorization and the 2 ⇥ Jordan canonical form. Eigenvalue-based decompositions are ideal for ana- lyzing the behavior of dynamical systems likes x0(t)=Ax(t) or xk+1 = Axk. When it comes to solving linear systems of equations or tackling more general problems in data science, eigenvalue-based factorizations are often not so il- luminating. In this chapter we develop another decomposition that provides deep insight into the rank structure of a matrix, showing the way to solving all variety of linear equations and exposing optimal low-rank approximations. 3.1 Singular Value Decomposition The singular value decomposition (SVD) is remarkable factorization that writes a general rectangular matrix A m n in the form 2 ⇥ A = (unitary matrix) (diagonal matrix) (unitary matrix)⇤. ⇥ ⇥ From the unitary matrices we can extract bases for the four fundamental subspaces R(A), N(A), R(A⇤),andN(A⇤), and the diagonal matrix will reveal much about the rank structure of A. We will build up the SVD in a four-step process. For simplicity suppose that A m n with m n.(Ifm<n, apply the arguments below 2 ⇥ ≥ to A n m.) Note that A A n n is always Hermitian positive ⇤ 2 ⇥ ⇤ 2 ⇥ semidefinite. (Clearly (A⇤A)⇤ = A⇤(A⇤)⇤ = A⇤A,soA⇤A is Hermitian. For any x n,notethatx A Ax =(Ax) (Ax)= Ax 2 0,soA A is 2 ⇤ ⇤ ⇤ k k ≥ ⇤ positive semidefinite.) 91 i i i i i “book” — 2017/4/1 — 12:47 — page 92 — #94 i i i 92 Chapter 3.
    [Show full text]
  • MATH 210A, FALL 2017 Question 1. Let a and B Be Two N × N Matrices
    MATH 210A, FALL 2017 HW 6 SOLUTIONS WRITTEN BY DAN DORE (If you find any errors, please email [email protected]) Question 1. Let A and B be two n × n matrices with entries in a field K. −1 Let L be a field extension of K, and suppose there exists C 2 GLn(L) such that B = CAC . −1 Prove there exists D 2 GLn(K) such that B = DAD . (that is, turn the argument sketched in class into an actual proof) Solution. We’ll start off by proving the existence and uniqueness of rational canonical forms in a precise way. d d−1 To do this, recall that the companion matrix for a monic polynomial p(t) = t + ad−1t + ··· + a0 2 K[t] is defined to be the (d − 1) × (d − 1) matrix 0 1 0 0 ··· 0 −a0 B C B1 0 ··· 0 −a1 C B C B0 1 ··· 0 −a2 C M(p) := B C B: : : C B: :: : C @ A 0 0 ··· 1 −ad−1 This is the matrix representing the action of t in the cyclic K[t]-module K[t]=(p(t)) with respect to the ordered basis 1; t; : : : ; td−1. Proposition 1 (Rational Canonical Form). For any matrix M over a field K, there is some matrix B 2 GLn(K) such that −1 BMB ' Mcan := M(p1) ⊕ M(p2) ⊕ · · · ⊕ M(pm) Here, p1; : : : ; pn are monic polynomials in K[t] with p1 j p2 j · · · j pm. The monic polynomials pi are 1 −1 uniquely determined by M, so if for some C 2 GLn(K) we have CMC = M(q1) ⊕ · · · ⊕ M(qk) for q1 j q1 j · · · j qm 2 K[t], then m = k and qi = pi for each i.
    [Show full text]
  • Toeplitz Nonnegative Realization of Spectra Via Companion Matrices Received July 25, 2019; Accepted November 26, 2019
    Spec. Matrices 2019; 7:230–245 Research Article Open Access Special Issue Dedicated to Charles R. Johnson Macarena Collao, Mario Salas, and Ricardo L. Soto* Toeplitz nonnegative realization of spectra via companion matrices https://doi.org/10.1515/spma-2019-0017 Received July 25, 2019; accepted November 26, 2019 Abstract: The nonnegative inverse eigenvalue problem (NIEP) is the problem of nding conditions for the exis- tence of an n × n entrywise nonnegative matrix A with prescribed spectrum Λ = fλ1, . , λng. If the problem has a solution, we say that Λ is realizable and that A is a realizing matrix. In this paper we consider the NIEP for a Toeplitz realizing matrix A, and as far as we know, this is the rst work which addresses the Toeplitz nonnegative realization of spectra. We show that nonnegative companion matrices are similar to nonnegative Toeplitz ones. We note that, as a consequence, a realizable list Λ = fλ1, . , λng of complex numbers in the left-half plane, that is, with Re λi ≤ 0, i = 2, . , n, is in particular realizable by a Toeplitz matrix. Moreover, we show how to construct symmetric nonnegative block Toeplitz matrices with prescribed spectrum and we explore the universal realizability of lists, which are realizable by this kind of matrices. We also propose a Matlab Toeplitz routine to compute a Toeplitz solution matrix. Keywords: Toeplitz nonnegative inverse eigenvalue problem, Unit Hessenberg Toeplitz matrix, Symmetric nonnegative block Toeplitz matrix, Universal realizability MSC: 15A29, 15A18 Dedicated with admiration and special thanks to Charles R. Johnson. 1 Introduction The nonnegative inverse eigenvalue problem (hereafter NIEP) is the problem of nding necessary and sucient conditions for a list Λ = fλ1, λ2, .
    [Show full text]
  • The Polar Decomposition of Block Companion Matrices
    The polar decomposition of block companion matrices Gregory Kalogeropoulos1 and Panayiotis Psarrakos2 September 28, 2005 Abstract m m¡1 Let L(¸) = In¸ + Am¡1¸ + ¢ ¢ ¢ + A1¸ + A0 be an n £ n monic matrix polynomial, and let CL be the corresponding block companion matrix. In this note, we extend a known result on scalar polynomials to obtain a formula for the polar decomposition of CL when the matrices A0 Pm¡1 ¤ and j=1 Aj Aj are nonsingular. Keywords: block companion matrix, matrix polynomial, eigenvalue, singular value, polar decomposition. AMS Subject Classifications: 15A18, 15A23, 65F30. 1 Introduction and notation Consider the monic matrix polynomial m m¡1 L(¸) = In¸ + Am¡1¸ + ¢ ¢ ¢ + A1¸ + A0; (1) n£n where Aj 2 C (j = 0; 1; : : : ; m ¡ 1; m ¸ 2), ¸ is a complex variable and In denotes the n £ n identity matrix. The study of matrix polynomials, especially with regard to their spectral analysis, has a long history and plays an important role in systems theory [1, 2, 3, 4]. A scalar ¸0 2 C is said to be an eigenvalue of n L(¸) if the system L(¸0)x = 0 has a nonzero solution x0 2 C . This solution x0 is known as an eigenvector of L(¸) corresponding to ¸0. The set of all eigenvalues of L(¸) is the spectrum of L(¸), namely, sp(L) = f¸ 2 C : detL(¸) = 0g; and contains no more than nm distinct (finite) elements. 1Department of Mathematics, University of Athens, Panepistimioupolis 15784, Athens, Greece (E-mail: [email protected]). 2Department of Mathematics, National Technical University, Zografou Campus 15780, Athens, Greece (E-mail: [email protected]).
    [Show full text]
  • The Weyr Characteristic
    Swarthmore College Works Mathematics & Statistics Faculty Works Mathematics & Statistics 12-1-1999 The Weyr Characteristic Helene Shapiro Swarthmore College, [email protected] Follow this and additional works at: https://works.swarthmore.edu/fac-math-stat Part of the Mathematics Commons Let us know how access to these works benefits ouy Recommended Citation Helene Shapiro. (1999). "The Weyr Characteristic". American Mathematical Monthly. Volume 106, Issue 10. 919-929. DOI: 10.2307/2589746 https://works.swarthmore.edu/fac-math-stat/28 This work is brought to you for free by Swarthmore College Libraries' Works. It has been accepted for inclusion in Mathematics & Statistics Faculty Works by an authorized administrator of Works. For more information, please contact [email protected]. The Weyr Characteristic Author(s): Helene Shapiro Source: The American Mathematical Monthly, Vol. 106, No. 10 (Dec., 1999), pp. 919-929 Published by: Mathematical Association of America Stable URL: http://www.jstor.org/stable/2589746 Accessed: 19-10-2016 15:44 UTC JSTOR is a not-for-profit service that helps scholars, researchers, and students discover, use, and build upon a wide range of content in a trusted digital archive. We use information technology and tools to increase productivity and facilitate new forms of scholarship. For more information about JSTOR, please contact [email protected]. Your use of the JSTOR archive indicates your acceptance of the Terms & Conditions of Use, available at http://about.jstor.org/terms Mathematical Association of America is collaborating with JSTOR to digitize, preserve and extend access to The American Mathematical Monthly This content downloaded from 130.58.65.20 on Wed, 19 Oct 2016 15:44:49 UTC All use subject to http://about.jstor.org/terms The Weyr Characteristic Helene Shapiro 1.
    [Show full text]
  • NON-SPARSE COMPANION MATRICES∗ 1. Introduction. The
    Electronic Journal of Linear Algebra, ISSN 1081-3810 A publication of the International Linear Algebra Society Volume 35, pp. 223-247, June 2019. NON-SPARSE COMPANION MATRICES∗ LOUIS DEAETTy , JONATHAN FISCHERz , COLIN GARNETTx , AND KEVIN N. VANDER MEULEN{ Abstract. Given a polynomial p(z), a companion matrix can be thought of as a simple template for placing the coefficients of p(z) in a matrix such that the characteristic polynomial is p(z). The Frobenius companion and the more recently-discovered Fiedler companion matrices are examples. Both the Frobenius and Fiedler companion matrices have the maximum possible number of zero entries, and in that sense are sparse. In this paper, companion matrices are explored that are not sparse. Some constructions of non-sparse companion matrices are provided, and properties that all companion matrices must exhibit are given. For example, it is shown that every companion matrix realization is non-derogatory. Bounds on the minimum number of zeros that must appear in a companion matrix, are also given. Key words. Companion matrix, Fiedler companion matrix, Non-derogatory matrix, Characteristic polynomial. AMS subject classifications. 15A18, 15B99, 11C20, 05C50. 1. Introduction. The Frobenius companion matrix to the polynomial n n−1 n−2 (1.1) p(z) = z + a1z + a2z + ··· + an−1z + an is the matrix 2 0 1 0 ··· 0 3 6 0 0 1 ··· 0 7 6 7 6 . 7 (1.2) F = 6 . .. 7 : 6 . 7 6 7 4 0 0 0 ··· 1 5 −an −an−1 · · · −a2 −a1 2 In general, a companion matrix to p(z) is an n×n matrix C = C(p) over R[a1; a2; : : : ; an] with n −n entries in R and the remaining entries variables −a1; −a2;:::; −an such that the characteristic polynomial of C is p(z).
    [Show full text]
  • Permanental Compounds and Permanents of (O,L)-Circulants*
    CORE Metadata, citation and similar papers at core.ac.uk Provided by Elsevier - Publisher Connector Permanental Compounds and Permanents of (O,l)-Circulants* Henryk Mint Department of Mathematics University of California Santa Barbara, Califmia 93106 Submitted by Richard A. Brualdi ABSTRACT It was shown by the author in a recent paper that a recurrence relation for permanents of (0, l)-circulants can be generated from the product of the characteristic polynomials of permanental compounds of the companion matrix of a polynomial associated with (O,l)-circulants of the given type. In the present paper general properties of permanental compounds of companion matrices are studied, and in particular of convertible companion matrices, i.e., matrices whose permanental com- pounds are equal to the determinantal compounds after changing the signs of some of their entries. These results are used to obtain formulas for the limit of the nth root of the permanent of the n X n (0, l)-circulant of a given type, as n tends to infinity. The root-squaring method is then used to evaluate this limit for a wide range of circulant types whose associated polynomials have convertible companion matrices. 1. INTRODUCTION tit Q,,, denote the set of increasing sequences of integers w = (+w2,..., w,), 1 < w1 < w2 < * 9 * < w, < t. If A = (ajj) is an s X t matrix, and (YE Qh,s, p E QkSt, then A[Lx~/~] denotes the h X k submatrix of A whose (i, j) entry is aa,Bj, i = 1,2 ,..., h, j = 1,2,.. ., k, and A(aIP) desig- nates the (s - h) X (t - k) submatrix obtained from A by deleting rows indexed (Y and columns indexed /I.
    [Show full text]
  • Matrix Theory
    Matrix Theory Xingzhi Zhan +VEHYEXI7XYHMIW MR1EXLIQEXMGW :SPYQI %QIVMGER1EXLIQEXMGEP7SGMIX] Matrix Theory https://doi.org/10.1090//gsm/147 Matrix Theory Xingzhi Zhan Graduate Studies in Mathematics Volume 147 American Mathematical Society Providence, Rhode Island EDITORIAL COMMITTEE David Cox (Chair) Daniel S. Freed Rafe Mazzeo Gigliola Staffilani 2010 Mathematics Subject Classification. Primary 15-01, 15A18, 15A21, 15A60, 15A83, 15A99, 15B35, 05B20, 47A63. For additional information and updates on this book, visit www.ams.org/bookpages/gsm-147 Library of Congress Cataloging-in-Publication Data Zhan, Xingzhi, 1965– Matrix theory / Xingzhi Zhan. pages cm — (Graduate studies in mathematics ; volume 147) Includes bibliographical references and index. ISBN 978-0-8218-9491-0 (alk. paper) 1. Matrices. 2. Algebras, Linear. I. Title. QA188.Z43 2013 512.9434—dc23 2013001353 Copying and reprinting. Individual readers of this publication, and nonprofit libraries acting for them, are permitted to make fair use of the material, such as to copy a chapter for use in teaching or research. Permission is granted to quote brief passages from this publication in reviews, provided the customary acknowledgment of the source is given. Republication, systematic copying, or multiple reproduction of any material in this publication is permitted only under license from the American Mathematical Society. Requests for such permission should be addressed to the Acquisitions Department, American Mathematical Society, 201 Charles Street, Providence, Rhode Island 02904-2294 USA. Requests can also be made by e-mail to [email protected]. c 2013 by the American Mathematical Society. All rights reserved. The American Mathematical Society retains all rights except those granted to the United States Government.
    [Show full text]
  • Computing Rational Forms of Integer Matrices
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Elsevier - Publisher Connector J. Symbolic Computation (2002) 34, 157–172 doi:10.1006/jsco.2002.0554 Available online at http://www.idealibrary.com on Computing Rational Forms of Integer Matrices MARK GIESBRECHT† AND ARNE STORJOHANN‡ School of Computer Science, University of Waterloo, Waterloo, Ontario, Canada N2L 3G1 n×n A new algorithm is presented for finding the Frobenius rational form F ∈ Z of any n×n 4 3 A ∈ Z which requires an expected number of O(n (log n + log kAk) + n (log n + log kAk)2) word operations using standard integer and matrix arithmetic (where kAk = maxij |Aij |). This substantially improves on the fastest previously known algorithms. The algorithm is probabilistic of the Las Vegas type: it assumes a source of random bits but always produces the correct answer. Las Vegas algorithms are also presented for computing a transformation matrix to the Frobenius form, and for computing the rational Jordan form of an integer matrix. c 2002 Elsevier Science Ltd. All rights reserved. 1. Introduction In this paper we present new algorithms for exactly computing the Frobenius and rational Jordan normal forms of an integer matrix which are substantially faster than those previously known. We show that the Frobenius form F ∈ Zn×n of any A ∈ Zn×n can be computed with an expected number of O(n4(log n+log kAk)+n3(log n+log kAk)2) word operations using standard integer and matrix arithmetic. Here and throughout this paper, kAk = maxij |Aij|.
    [Show full text]
  • New Classes of Matrix Decompositions
    NEW CLASSES OF MATRIX DECOMPOSITIONS KE YE Abstract. The idea of decomposing a matrix into a product of structured matrices such as tri- angular, orthogonal, diagonal matrices is a milestone of numerical computations. In this paper, we describe six new classes of matrix decompositions, extending our work in [8]. We prove that every n × n matrix is a product of finitely many bidiagonal, skew symmetric (when n is even), companion matrices and generalized Vandermonde matrices, respectively. We also prove that a generic n × n centrosymmetric matrix is a product of finitely many symmetric Toeplitz (resp. persymmetric Hankel) matrices. We determine an upper bound of the number of structured matrices needed to decompose a matrix for each case. 1. Introduction Matrix decomposition is an important technique in numerical computations. For example, we have classical matrix decompositions: (1) LU: a generic matrix can be decomposed as a product of an upper triangular matrix and a lower triangular matrix. (2) QR: every matrix can be decomposed as a product of an orthogonal matrix and an upper triangular matrix. (3) SVD: every matrix can be decomposed as a product of two orthogonal matrices and a diagonal matrix. These matrix decompositions play a central role in engineering and scientific problems related to matrix computations [1],[6]. For example, to solve a linear system Ax = b where A is an n×n matrix and b is a column vector of length n. We can first apply LU decomposition to A to obtain LUx = b, where L is a lower triangular matrix and U is an upper triangular matrix.
    [Show full text]
  • Commutators and Companion Matrices Over Rings of Stable Rank 1
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Elsevier - Publisher Connector Commutators and Companion Matrices over Rings of Stable Rank 1 L. N. Vaserstein and E. Wheland Department of Mathematics The Pennsylvania State University University Park, Pennsylvania 16802 Submitted by Robert C. Thompson ABSTRACT We consider the group GL,A of all invertible n by n matrices over a ring A satisfying the first Bass stable range condition. We prove that every matrix is similar to the product of a lower and upper triangular matrix, and that it is also the product of two matrices each similar to a companion matrix. We use this to show that, when n > 3 and A is commutative, every matrix in SL,A is the product of two commuta- tors. INTRODUCTION In various situations it is instructive to represent a matrix as a product of matrices of a special nature. A large survey of various results concerning factorization of matrices over a field and bounded operators on a Hilbert space is given in Wu [75]. G’lven a class of matrices (or operators), one studies products of elements from the class, and asks about the minimal number of factors in a factorization. Classes considered in [75] include the class of normal matrices, the class of involutions, the class of symmetric matrices, and many others. See also [14, 11, 15, 16, 711 for various matrix factorizations. A classical factorization result is that every orthogonal n by n matrix is the product of at most n reflections [2, Proposition 5, Chapter IX, $6, Section 41; see [I21 for further work on reflections.
    [Show full text]
  • A Polynomial in a of the Diagonalizable and Nilpotent Parts of A
    Western Oregon University Digital Commons@WOU Honors Senior Theses/Projects Student Scholarship 6-1-2017 A Polynomial in A of the Diagonalizable and Nilpotent Parts of A Kathryn Wilson Western Oregon University Follow this and additional works at: https://digitalcommons.wou.edu/honors_theses Recommended Citation Wilson, Kathryn, "A Polynomial in A of the Diagonalizable and Nilpotent Parts of A" (2017). Honors Senior Theses/Projects. 145. https://digitalcommons.wou.edu/honors_theses/145 This Undergraduate Honors Thesis/Project is brought to you for free and open access by the Student Scholarship at Digital Commons@WOU. It has been accepted for inclusion in Honors Senior Theses/Projects by an authorized administrator of Digital Commons@WOU. For more information, please contact [email protected], [email protected], [email protected]. A Polynomial in A of the Diagonalizable and Nilpotent Parts of A By Kathryn Wilson An Honors Thesis Submitted in Partial Fulfillment of the Requirements for Graduation from the Western Oregon University Honors Program Dr. Scott Beaver, Thesis Advisor Dr. Gavin Keulks, Honor Program Director June 2017 Acknowledgements I would first like to thank my advisor, Dr. Scott Beaver. He has been an amazing resources throughout this entire process. Your support was what made this project possible. I would also like to thank Dr. Gavin Keulks for his part in running the Honors Program and allowing me the opportunity to be part of it. Further I would like to thank the WOU Mathematics department for making my education here at Western immensely fulfilling, and instilling in me the quality of thinking that will create my success outside of school.
    [Show full text]