3·Singular Value Decomposition

Total Page:16

File Type:pdf, Size:1020Kb

3·Singular Value Decomposition i “book” — 2017/4/1 — 12:47 — page 91 — #93 i i i Embree – draft – 1 April 2017 3 Singular Value Decomposition · Thus far we have focused on matrix factorizations that reveal the eigenval- ues of a square matrix A n n,suchastheSchur factorization and the 2 ⇥ Jordan canonical form. Eigenvalue-based decompositions are ideal for ana- lyzing the behavior of dynamical systems likes x0(t)=Ax(t) or xk+1 = Axk. When it comes to solving linear systems of equations or tackling more general problems in data science, eigenvalue-based factorizations are often not so il- luminating. In this chapter we develop another decomposition that provides deep insight into the rank structure of a matrix, showing the way to solving all variety of linear equations and exposing optimal low-rank approximations. 3.1 Singular Value Decomposition The singular value decomposition (SVD) is remarkable factorization that writes a general rectangular matrix A m n in the form 2 ⇥ A = (unitary matrix) (diagonal matrix) (unitary matrix)⇤. ⇥ ⇥ From the unitary matrices we can extract bases for the four fundamental subspaces R(A), N(A), R(A⇤),andN(A⇤), and the diagonal matrix will reveal much about the rank structure of A. We will build up the SVD in a four-step process. For simplicity suppose that A m n with m n.(Ifm<n, apply the arguments below 2 ⇥ ≥ to A n m.) Note that A A n n is always Hermitian positive ⇤ 2 ⇥ ⇤ 2 ⇥ semidefinite. (Clearly (A⇤A)⇤ = A⇤(A⇤)⇤ = A⇤A,soA⇤A is Hermitian. For any x n,notethatx A Ax =(Ax) (Ax)= Ax 2 0,soA A is 2 ⇤ ⇤ ⇤ k k ≥ ⇤ positive semidefinite.) 91 i i i i i “book” — 2017/4/1 — 12:47 — page 92 — #94 i i i 92 Chapter 3. Singular Value Decomposition Step 1. Using the spectral decomposition of a Hermitian matrix discussed in Section 1.5, A A has n eigenpairs (λ , v ) n with orthonormal unit ⇤ { j j }j=1 eigenvectors (v v =1, v v =0when j = k). We are free to pick any j⇤ j j⇤ k 6 convenient indexing for these eigenpairs; label the eigenvalues in decreasing magnitude, λ λ λ 0. 1 ≥ 2 ≥···≥ n ≥ Step 2. Define s := Av . j k jk Note that s2 = Av 2 = v A Av = λ . Since the eigenvalues λ ,...,λ j k jk j⇤ ⇤ j j 1 n are decreasing in magnitude, so are the s values: s s s 0. j 1 ≥ 2 ≥···≥ n ≥ Step 3. Next, we will build a set of related orthonormal vectors in m. Suppose we have already constructed such vectors u1,...,uj 1. − 1 1 If s =0, then define u = s− Av ,sothat u = s− Av =1. j 6 j j j k jk j k jk If sj =0, then pick uj to be any unit vector such that uj span u1,...,uj 1 ?; 2 { − } 1 i.e., ensure uj⇤uk =0for all k<j. By construction, u u =0for j = k if s or s is zero. If both s and s j⇤ k 6 j k j k are nonzero, then 1 1 λk uj⇤uk = (Avj)⇤(Avk)= vj⇤A⇤Avk = vj⇤vk, sjsk sjsk sjsk where we used the fact that v is an eigenvector of A A.Nowifj = k, j ⇤ 6 then vj⇤vk =0, and hence uj⇤uk =0. On the other hand, j = k implies that 2 vj⇤vk =1,souj⇤uk = λj/sj =1. In conclusion, we have constructed a set of orthonormal vectors u n { j}j=1 with u m. j 2 Step 4. For all j =1,...,n, Avj = sjuj, regardless of whether sj =0or not. We can stack these n vector equations as columns of a single matrix equation, || | || | Av Av Av = s u s u s u . 2 1 2 ··· n 3 2 1 1 2 2 ··· n n 3 || | || | 4 5 4 5 1 If sj =0,thenλj =0,andsoA⇤A has a zero eigenvalue; i.e., this matrix is singular. Embree – draft – 1 April 2017 i i i i i “book” — 2017/4/1 — 12:47 — page 93 — #95 i i i 3.1. Singular Value Decomposition 93 Note that both matrices in this equation can be factored into the product of simpler matrices: s1 s || | || |2 2 3 A v1 v2 vn = u1 u2 un . 2 ··· 3 2 ··· 3 .. 6 7 || | || |6 s 7 4 5 4 5 6 n 7 4 5 Denote these matrices as AV = U⌃, where A m n, V n n, U 2 ⇥ 2 ⇥ 2 m n,and⌃ n n. ⇥ 2 ⇥ The (j, k) entry of V⇤V is simplyb b vj⇤vk,andsoV⇤V = I. Since Vbis a square matrix,b we have just proved that it is unitary, and hence, VV⇤ = I as well. We conclude that A = U⌃V⇤. This matrix factorization is known asb theb reduced singular value decomposi- tion or the economy-sized singular value decomposition (or, informally, the skinny SVD). It can be obtained via the MATLAB command [Uhat, Sighat, V] = svd(A,’econ’); The Reduced Singular Value Decomposition Theorem 3.1. Any matrix A m n with m n can be written as 2 ⇥ ≥ A = U⌃V⇤, m n n n where U ⇥ has orthonormalb columns,b V ⇥ is unitary, and 2 2 ⌃ = diag(s ,...,s ) n n has real nonnegative decreasing entries. 1 n 2 ⇥ The columnsb of U are left singular vectors,thecolumnsofV are right singularb vectors,andthevaluess1,...,sn are the singular values. While the matrix U has orthonormal columns, it is not a unitary matrix when m>n. In particular, we have U U = I n n,but ⇤ 2 ⇥ b m m UU⇤ b b ⇥ 2 cannot be the identity unless mb b= n. (To see this, note that UU⇤ is an orthogonal projection onto R(U) = span u ,...,u . Since dim(R(U)) = { 1 n} n, this projection cannot equal the m-by-m identity matrix whenbm>nb .) b b Embree – draft – 1 April 2017 i i i i i “book” — 2017/4/1 — 12:47 — page 94 — #96 i i i 94 Chapter 3. Singular Value Decomposition Though U is not unitary, it is subunitary. We can construct m n − additional column vectors to append to U to make it unitary. Here is the recipe: For jb= n +1,...,m,pick b uj span u1,...,uj 1 ? 2 { − } with uj⇤uj =1. Then define || | U = u u u . 2 1 2 ··· m 3 || | 4 5 Confirm that U U = UU = I m m, showing that U is unitary. ⇤ ⇤ 2 ⇥ We wish to replace the U in the reduced SVD with the unitary matrix U. To do so, we also need to replace ⌃ by some ⌃ in such a way that U⌃ = U⌃. The simplest approachb constructs ⌃ by appending zeros to the end of ⌃, thus ensuring there is no contributionb when the new entries of U multiplyb b against the new entries of ⌃: b ⌃ m n ⌃ = ⇥ . 0 2 b The factorization A = U⌃V⇤ is called the full singular value decomposition. The Full Singular Value Decomposition Theorem 3.2. Any matrix A m n can be written in the form 2 ⇥ A = U⌃V⇤, where U m m and V n n are unitary matrices and ⌃ m n 2 ⇥ 2 ⇥ 2 ⇥ is zero except for the main diagonal; The columns of U are left singular vectors,thecolumnsofV are right singular vectors,andthevaluess1,...,smin m,n are the singular values. { } A third version of the singular value decomposition is often very helpful. Start with the reduced SVD A = U⌃V⇤ in Theorem 3.1 Multiply U⌃ together to get b b b b v1⇤ n || . A = U⌃V⇤ = s u s u . = s u v⇤, 2 1 1 ··· n n 3 2 . 3 j j j v j=1 ||6 n⇤ 7 X b b 4 5 4 5 Embree – draft – 1 April 2017 i i i i i “book” — 2017/4/1 — 12:47 — page 95 — #97 i i i 3.1. Singular Value Decomposition 95 which renders A as the sum of outer products u v m n,weightedby j j⇤ 2 ⇥ nonnegative numbers sj. Let r denote the number of nonzero singular values, so that if r<n, then s = = s =0. r+1 ··· n Thus A can be written as r A = sjujvj⇤, (3.1) Xj=1 known as the dyadic form of the SVD. The Dyadic Form of the Singular Value Decomposition Theorem 3.3. For any matrix A m n,thereexistssomer 2 ⇥ 2 1,...,n such that { } r A = sjujvj⇤, Xj=1 where s s s > 0 and u ,...,u m are orthonormal, and 1 ≥ 2 ≥···≥ r 1 r 2 v ,...,v n are orthonormal. 1 r 2 Corollary 3.4. The rank of a matrix equals its number of nonzero singular values. The singular value decomposition also gives an immediate formula for the 2-norm of a matrix. Theorem 3.5. Let A m n have singular value decomposition (3.1). 2 ⇥ Then Ax A = max k k = s1. k k x=0 x 6 k k Proof. The proof follows from the construction of the SVD at the start of this chapter. Note that Ax 2 x A Ax A 2 = max k k = max ⇤ ⇤ , k k x=0 x 2 x=0 x x 6 k k 6 ⇤ and so A 2 is the maximum Rayleigh quotient of the Hermitian matrix k k A⇤A. By Theorem 2.2, this maximum value is the largest eigenvalue of Embree – draft – 1 April 2017 i i i i i “book” — 2017/4/1 — 12:47 — page 96 — #98 i i i 96 Chapter 3.
Recommended publications
  • The Weyr Characteristic
    Swarthmore College Works Mathematics & Statistics Faculty Works Mathematics & Statistics 12-1-1999 The Weyr Characteristic Helene Shapiro Swarthmore College, [email protected] Follow this and additional works at: https://works.swarthmore.edu/fac-math-stat Part of the Mathematics Commons Let us know how access to these works benefits ouy Recommended Citation Helene Shapiro. (1999). "The Weyr Characteristic". American Mathematical Monthly. Volume 106, Issue 10. 919-929. DOI: 10.2307/2589746 https://works.swarthmore.edu/fac-math-stat/28 This work is brought to you for free by Swarthmore College Libraries' Works. It has been accepted for inclusion in Mathematics & Statistics Faculty Works by an authorized administrator of Works. For more information, please contact [email protected]. The Weyr Characteristic Author(s): Helene Shapiro Source: The American Mathematical Monthly, Vol. 106, No. 10 (Dec., 1999), pp. 919-929 Published by: Mathematical Association of America Stable URL: http://www.jstor.org/stable/2589746 Accessed: 19-10-2016 15:44 UTC JSTOR is a not-for-profit service that helps scholars, researchers, and students discover, use, and build upon a wide range of content in a trusted digital archive. We use information technology and tools to increase productivity and facilitate new forms of scholarship. For more information about JSTOR, please contact [email protected]. Your use of the JSTOR archive indicates your acceptance of the Terms & Conditions of Use, available at http://about.jstor.org/terms Mathematical Association of America is collaborating with JSTOR to digitize, preserve and extend access to The American Mathematical Monthly This content downloaded from 130.58.65.20 on Wed, 19 Oct 2016 15:44:49 UTC All use subject to http://about.jstor.org/terms The Weyr Characteristic Helene Shapiro 1.
    [Show full text]
  • A Polynomial in a of the Diagonalizable and Nilpotent Parts of A
    Western Oregon University Digital Commons@WOU Honors Senior Theses/Projects Student Scholarship 6-1-2017 A Polynomial in A of the Diagonalizable and Nilpotent Parts of A Kathryn Wilson Western Oregon University Follow this and additional works at: https://digitalcommons.wou.edu/honors_theses Recommended Citation Wilson, Kathryn, "A Polynomial in A of the Diagonalizable and Nilpotent Parts of A" (2017). Honors Senior Theses/Projects. 145. https://digitalcommons.wou.edu/honors_theses/145 This Undergraduate Honors Thesis/Project is brought to you for free and open access by the Student Scholarship at Digital Commons@WOU. It has been accepted for inclusion in Honors Senior Theses/Projects by an authorized administrator of Digital Commons@WOU. For more information, please contact [email protected], [email protected], [email protected]. A Polynomial in A of the Diagonalizable and Nilpotent Parts of A By Kathryn Wilson An Honors Thesis Submitted in Partial Fulfillment of the Requirements for Graduation from the Western Oregon University Honors Program Dr. Scott Beaver, Thesis Advisor Dr. Gavin Keulks, Honor Program Director June 2017 Acknowledgements I would first like to thank my advisor, Dr. Scott Beaver. He has been an amazing resources throughout this entire process. Your support was what made this project possible. I would also like to thank Dr. Gavin Keulks for his part in running the Honors Program and allowing me the opportunity to be part of it. Further I would like to thank the WOU Mathematics department for making my education here at Western immensely fulfilling, and instilling in me the quality of thinking that will create my success outside of school.
    [Show full text]
  • Canonical Matrices for Linear Matrix Problems, Preprint 99-070 of SFB 343, Bielefeld University, 1999, 43 P
    Canonical matrices for linear matrix problems Vladimir V. Sergeichuk Institute of Mathematics Tereshchenkivska 3, Kiev, Ukraine [email protected] Abstract We consider a large class of matrix problems, which includes the problem of classifying arbitrary systems of linear mappings. For every matrix problem from this class, we construct Belitski˘ı’s algorithm for reducing a matrix to a canonical form, which is the generalization of the Jordan normal form, and study the set Cmn of indecomposable canonical m × n matrices. Considering Cmn as a subset in the affine space of m-by-n matrices, we prove that either Cmn consists of a finite number of points and straight lines for every m × n, or Cmn contains a 2-dimensional plane for a certain m × n. AMS classification: 15A21; 16G60. Keywords: Canonical forms; Canonical matrices; Reduction; Clas- sification; Tame and wild matrix problems. All matrices are considered over an algebraically closed field k; km×n denotes the set of m-by-n matrices over k. The article consists of three sections. arXiv:0709.2485v1 [math.RT] 16 Sep 2007 In Section 1 we present Belitski˘ı’salgorithm [2] (see also [3]) in a form, which is convenient for linear algebra. In particular, the algorithm permits to reduce pairs of n-by-n matrices to a canonical form by transformations of simultaneous similarity: (A, B) 7→ (S−1AS,S−1BS); another solution of this classical problem was given by Friedland [15]. This section uses rudimentary linear algebra (except for the proof of Theorem 1.1) and may be interested for the general reader.
    [Show full text]
  • The Jordan Canonical Form
    The Jordan canonical form Attila M´at´e Brooklyn College of the City University of New York November 14, 2014 Contents 1 Preliminaries 2 1.1 Rank-nullitytheorem. ....... 2 1.2 Existence of eigenvalues and eigenvectors . ............. 3 2 Existence of a Jordan decomposition 3 3 Uniqueness of the Jordan decomposition 6 4 Theminimalpolynomialofalineartransformation 7 4.1 Existenceoftheminimalpolynomial . .......... 7 4.2 The minimal polynomial for algebraically closed fields . ................ 8 4.3 The characteristic polynomial and the Cayley–Hamilton theorem ........... 9 4.4 Findingtheminimalpolynomial . ........ 9 5 Consequences for matrices 11 5.1 Representation of vector spaces and linear transformations............... 11 5.2 Similaritytransformations . .......... 12 5.3 Directsumsofmatrices ............................ ...... 12 5.4 Jordanblockmatrices ............................. ...... 12 5.5 TheJordancanonicalformofamatrix . ......... 13 5.6 The characteristic polynomial as a determinant . .............. 13 6 Anexampleforfindingtheminimalpolynomialofamatrix 13 7 Theexamplecontinued: findingtheJordancanonicalform 14 7.1 ThefirsttwoJordansubspaces . ....... 14 7.2 ThenexttoJordansubspaces. ....... 16 7.3 TheJordancanonicalform . ...... 18 8 Matrices and field extensions 20 8.1 Afurtherdecompositiontheorem . ......... 20 1 1 Preliminaries 1.1 Rank-nullity theorem Let U and V be vector spaces over a field F , and let T : U V be a linear transformation. Then the domain of T , denoted as dom(T ), is the set U itself. The range→ of T is the set ra(T )= T u : u U , and it is a subspace of V ; one often writes TU instead of ra(T ). The kernel, or null{ space, of∈T }is the set ker(T )= u U : T u = 0 ; the kernel is a subspace of U. A key theorem is the following, often called the Rank-Nullity{ ∈ Theorem:} Theorem 1.
    [Show full text]
  • The Canonical Contact Form
    The Canonical Contact Form Peter J. Olvery School of Mathematics University of Minnesota Minneapolis, MN 55455 U.S.A. [email protected] http://www.math.umn.edu/»olver Abstract. The structure group and the involutive di®erential system that character- ize the pseudo-group of contact transformations on a jet space are determined. 1. Introduction. The canonical form on the coframe bundle over a smooth manifold originally arose as the natural generalization of the canonical form on the cotangent bundle, which plays an essential role in Hamiltonian mechanics, [19; xIII.7]. The coframe bundle F ¤M ! M forms a principal GL(m) bundle over the m-dimensional manifold M. The canonical form on the coframe bundle serves to characterize the di®eomorphism pseudo-group of the manifold, or, more correctly, its lift to the coframe bundle. Indeed, the invariance of the canonical form forms an involutive di®erential system, whose general solution, guaranteed by the Cartan{KÄahler Theorem, is the lifted di®eomorphism pseudo-group. Kobayashi, [11], introduces a vector-valued canonical form on the higher order frame bundles over y Supported in part by NSF Grant DMS 01{03944. February 21, 2005 1 the manifold. He demonstrates that the components of the canonical form constitute an involutive di®erential system that characterizes the higher order lifts of the di®eomorphism group. The geometrical study of di®erential equations relies on the jet space ¯rst introduced by Ehresmann, [6]. In the jet bundle framework, the pseudo-group of contact transforma- tions, [13, 16], assumes the role of the di®eomorphism pseudo-group.
    [Show full text]
  • Notes on Math 321
    Notes on Math 321 September 2, 2007 ii °c 2006 Guowu Meng. All rights reserved. Contents Preface v 1 Curves 1 1.1 Basic De¯nitions ...................................... 1 1.2 Frenet Trihedron ...................................... 3 1.3 Local canonical form .................................... 4 1.4 Exercises .......................................... 5 2 Regular Surfaces 7 2.1 Digression: Calculus .................................... 8 2.1.1 Chain Rule ..................................... 8 2.1.2 Inverse Function Theorems ............................ 8 2.2 Basic De¯nitions ...................................... 9 2.2.1 Proof of theorem 2.2.1 ............................... 10 2.2.2 Proof of theorem 2.2.3 ............................... 11 2.3 Existence of tangent planes ................................ 11 2.4 Local Canonical Form ................................... 12 2.5 Exercises .......................................... 13 3 Gauss Map 15 3.1 Smooth maps between regular surface .......................... 15 3.2 Orientation, Gauss map and curvatures ......................... 15 3.3 Digression: Linear Algebra ................................ 17 3.3.1 One-form ...................................... 17 3.3.2 Two-form ...................................... 17 3.3.3 The area form on an oriented Euclidean plane .................. 19 3.4 Fundamental Forms .................................... 20 3.4.1 The total di®erentials ............................... 20 3.4.2 The ¯rst fundamental two-form .......................... 22 3.4.3 The second
    [Show full text]
  • Arxiv:1801.05390V2 [Quant-Ph]
    Gauge fixing, canonical forms and optimal truncations in tensor networks with closed loops Glen Evenbly1 1D´epartement de Physique and Institut Quantique, Universit´ede Sherbrooke, Qu´ebec, Canada ∗ (Dated: May 21, 2018) We describe an approach to fix the gauge degrees of freedom in tensor networks, including those with closed loops, which allows a canonical form for arbitrary tensor networks to be realised. Ad- ditionally, a measure for the internal correlations present in a tensor network is proposed, which quantifies the extent of resonances around closed loops in the network. Finally we describe an al- gorithm for the optimal truncation of an internal index from a tensor network, based upon proper removal of the redundant internal correlations. These results, which offer a unified theoretical framework for the manipulation of tensor networks with closed loops, can be applied to improve existing tensor network methods for the study of many-body systems and may also constitute key algorithmic components of sophisticated new tensor methods. PACS numbers: 05.30.-d, 02.70.-c, 03.67.Mn, 75.10.Jm I. INTRODUCTION velopment of tensor network algorithms for quantum sys- tems in D > 1 dimensions, such as those based on PEPS. 1–3 In this manuscript we present, for an arbitrary tensor Tensor network methods have proven to be excep- networks including those with closed loops (also called tionally useful in the study of quantum many-body sys- cyclic networks), (i) a method of fixing the gauge degrees tem and, more recently, have also found a diverse range of freedom, related to a generalization of the Schmidt of applications in areas such as quantum chemistry4,5, 6–8 9–12 decomposition, (ii) a means of quantifying the extent of machine learning , and holography .
    [Show full text]
  • 0.1 Rational Canonical Forms
    Draft. Do not cite or quote. We have already seen that it is useful and simpler to study linear systems using matrices. But matrices are themselves cumbersome, as they are stuffed with many entries, and it turns out that it's best to focus on the fact that we can view an m × n matrix A as a function n m n T : R ! R ; namely, T (x) = Ax for every column vector x 2 R . The function T , called a linear transformation, has two important properties: T (x + y) = T (x) + T (y), for all x; y; for every number α and vector x, we have T (αx) = αT (x). As we shall see in the course, the matrix A is a way of studying T by choosing a coordinate n system in R ; choosing another coordinate system gives another matrix describing T . Just as in calculus when studying conic sections, choosing a good coordinate system can greatly simplify things: we can put a conic section defined by Ax2 + Bxy + Cy2 + Dx + Ey + F into normal form, and we can easily read off properties of the conic (is it a parabola, hyperbola, or ellipse; what are its foci, etc) from the normal form. The important object of study for us is a linear transformation T , and any choice of coordinate system produces a matrix describing it. Is there a \best" matrix describing T analogous to a normal form for a conic? The answer is yes, and normal forms here are called canonical forms. Now coordinates arise from bases of a vector space, and the following discussion is often called change of basis.
    [Show full text]
  • Formalisation of the Computation of the Echelon Form of a Matrix in Isabelle/HOL
    Under consideration for publication in Formal Aspects of Computing Formalisation of the Computation of the Echelon Form of a Matrix in Isabelle/HOL Jes´usAransay and Jose Divas´on Abstract. In this contribution we present a formalised algorithm in the Isabelle/HOL proof assistant to compute echelon forms, and, as a consequence, characteristic polynomials of matrices. We have proved its correctness over B´ezoutdomains, but its executability is only guaranteed over Euclidean domains, such as the integer ring and the univariate polynomials over a field. This is possible since the algorithm has been parameterised by a (possibly non-computable) operation that returns the B´ezoutcoefficients of a pair of elements of a ring. The echelon form is also used to compute determinants and inverses of matrices. As a by-product, some algebraic structures have been implemented (principal ideal domains, B´ezoutdomains, etc.). In order to improve performance, the algorithm has been refined to immutable arrays inside of Isabelle and code can be generated to functional languages as well. Keywords: Theorem proving; Isabelle/HOL; Linear Algebra; Verified Code Generation 1. Introduction A classical mathematical problem is the transformation of a matrix over a ring into a canonical form (i.e., a unique matrix that is equivalent in some sense to the original matrix, and that provides a explicit way to distinguish matrices that are not equivalent), which have many applications in computational Linear Algebra. These canonical forms contain fundamental information of the original matrix, such as the determinant and the rank. Examples of canonical forms are the Hermite normal form (usually applied to integer matrices), the Smith normal form and the reduced row echelon form.
    [Show full text]
  • Minimal Polynomial and Jordan Form
    Minimal Polynomial and Jordan Form Tom Leinster The idea of these notes is to provide a summary of some of the results you need for this course, as well as a different perspective from the lectures. Minimal Polynomial Let V be a vector space over some field k, and let α : V - V be a linear map (an `endomorphism of V '). Given any polynomial p with coefficients in k, there is an endomorphism p(α) of V , and we say that p is an annihilating polynomial for α if p(α) = 0. Our first major goal is to see that for any α, the annihilating polynomials can easily be classified: they're precisely the multiples of a certain polynomial mα. We reach this goal on the next page. Let us say that a polynomial m is a minimal polynomial for α (note: `a', not `the') if: m(α) = 0 • if p is any non-zero polynomial with p(α) = 0, then deg(m) deg(p) • ≤ the leading coefficient of m is 1 (i.e. m is monic). • (We don't assign a degree to the zero polynomial. Note that the last condition implies m = 0.) We'd like6 to be able to talk about the minimal polynomial of α, and the following result legitimizes this: Proposition 1 There is precisely one minimal polynomial for α. Sketch Proof Existence: First prove that there exists a non-zero annihilating polynomial, as in your notes (using the fact that the space of endomorphisms on V is finite-dimensional). There is then a non-zero annihilating polynomial 1 M of least degree, and if c is the leading coefficient of M then c M is a minimal polynomial.
    [Show full text]
  • A General Theory of Canonical Forms
    TRANSACTIONS OF THE AMERICAN MATHEMATICAL SOCIETY Volume 300, Number 2, April 1987 A GENERAL THEORY OF CANONICAL FORMS RICHARD S. PALAIS AND CHUU-LIAN TERNG ABSTRACT. If G is a compact Lie group and M a Riemannian G-manifold with principal orbits of codimension k then a section or canonical form for M is a closed, smooth k-dimensional submanifold of M which meets all orbits of M orthogonally. We discuss some of the remarkable properties of G-manifolds that admit sections, develop methods for constructing sections, and consider several applications. O. Introduction. Let G be a compact Lie group acting isometrically on a Riemannian manifold M. Then the image S of a small ball in the normal plane v{Gx)x under the exponential map is a smooth, local Gx-slice, which in general cannot be extended to a global slice for M. A section E for M is defined to be a closed, smooth submanifold of M which meets every orbit of M orthogonally. A good example to keep in mind is perhaps the most important of all canonical form theorems; namely for M we take the Euclidean space of symmetric k x k matrices with inner product (A, B) = tr(AB), and for G the orthogonal group O{k) acting on M by conjugation. Then the space E of diagonal matrices is a section. Moreover the symmetric group Sk acts on E by permuting the diagonal entries and the orbit spaces M/G and E/Sk are isomorphic as stratified sets. Quite generally it is good intuition to think of a section E as representing a "canonical form" for elements of M; hence our title.
    [Show full text]
  • Canonical Forms Some Questions to Be Explored by High School Investigators William J
    MME 529 June 2017 Canonical Forms Some questions to be explored by high school investigators William J. Martin, WPI Here are some exercises based on various ideas of \canonical form" in mathematics. Perhaps a few can be used with students. I think we subconsciously use canonical forms in many parts of mathematics, mainly as a sort of language tool. In order to communicate with one another, we must agree to write things in the same way, as we agree on spelling conventions. Rational numbers are written with relatively prime numerator and denominator, with the denominator always pos- itive; algebraic numbers never have square roots in their denominators; integers in modular arithmetic are always reduced to their smallest non-negative remainder; etc. But the importance of canonical forms can be lost if we simply view them as notational conventions. The more important point is that canonical forms give us algorithms to de- termine if two mathematical objects are the same. If two students have fractions such as −105=35 and 51=(−17), one way to have them determine if the numbers are equal is to ask each student to reduce their fraction: they will both arrive at −3=1 and will immediately see that they are equal. We also do this if we are comparing f(x) = x3−13x+12 against g(x) = (x−3)(x−1)(x+4) or h(x) = (x + 4)(x2 + 2x − 3). Expanding the three polynomials and writing them each in 3 2 the canonical form a3x + a2x + a1x + a0, we may quickly check which two are equal.
    [Show full text]