Minimal Polynomial and Jordan Form
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
3·Singular Value Decomposition
i “book” — 2017/4/1 — 12:47 — page 91 — #93 i i i Embree – draft – 1 April 2017 3 Singular Value Decomposition · Thus far we have focused on matrix factorizations that reveal the eigenval- ues of a square matrix A n n,suchastheSchur factorization and the 2 ⇥ Jordan canonical form. Eigenvalue-based decompositions are ideal for ana- lyzing the behavior of dynamical systems likes x0(t)=Ax(t) or xk+1 = Axk. When it comes to solving linear systems of equations or tackling more general problems in data science, eigenvalue-based factorizations are often not so il- luminating. In this chapter we develop another decomposition that provides deep insight into the rank structure of a matrix, showing the way to solving all variety of linear equations and exposing optimal low-rank approximations. 3.1 Singular Value Decomposition The singular value decomposition (SVD) is remarkable factorization that writes a general rectangular matrix A m n in the form 2 ⇥ A = (unitary matrix) (diagonal matrix) (unitary matrix)⇤. ⇥ ⇥ From the unitary matrices we can extract bases for the four fundamental subspaces R(A), N(A), R(A⇤),andN(A⇤), and the diagonal matrix will reveal much about the rank structure of A. We will build up the SVD in a four-step process. For simplicity suppose that A m n with m n.(Ifm<n, apply the arguments below 2 ⇥ ≥ to A n m.) Note that A A n n is always Hermitian positive ⇤ 2 ⇥ ⇤ 2 ⇥ semidefinite. (Clearly (A⇤A)⇤ = A⇤(A⇤)⇤ = A⇤A,soA⇤A is Hermitian. For any x n,notethatx A Ax =(Ax) (Ax)= Ax 2 0,soA A is 2 ⇤ ⇤ ⇤ k k ≥ ⇤ positive semidefinite.) 91 i i i i i “book” — 2017/4/1 — 12:47 — page 92 — #94 i i i 92 Chapter 3. -
Estimations of the Trace of Powers of Positive Self-Adjoint Operators by Extrapolation of the Moments∗
Electronic Transactions on Numerical Analysis. ETNA Volume 39, pp. 144-155, 2012. Kent State University Copyright 2012, Kent State University. http://etna.math.kent.edu ISSN 1068-9613. ESTIMATIONS OF THE TRACE OF POWERS OF POSITIVE SELF-ADJOINT OPERATORS BY EXTRAPOLATION OF THE MOMENTS∗ CLAUDE BREZINSKI†, PARASKEVI FIKA‡, AND MARILENA MITROULI‡ Abstract. Let A be a positive self-adjoint linear operator on a real separable Hilbert space H. Our aim is to build estimates of the trace of Aq, for q ∈ R. These estimates are obtained by extrapolation of the moments of A. Applications of the matrix case are discussed, and numerical results are given. Key words. Trace, positive self-adjoint linear operator, symmetric matrix, matrix powers, matrix moments, extrapolation. AMS subject classifications. 65F15, 65F30, 65B05, 65C05, 65J10, 15A18, 15A45. 1. Introduction. Let A be a positive self-adjoint linear operator from H to H, where H is a real separable Hilbert space with inner product denoted by (·, ·). Our aim is to build estimates of the trace of Aq, for q ∈ R. These estimates are obtained by extrapolation of the integer moments (z, Anz) of A, for n ∈ N. A similar procedure was first introduced in [3] for estimating the Euclidean norm of the error when solving a system of linear equations, which corresponds to q = −2. The case q = −1, which leads to estimates of the trace of the inverse of a matrix, was studied in [4]; on this problem, see [10]. Let us mention that, when only positive powers of A are used, the Hilbert space H could be infinite dimensional, while, for negative powers of A, it is always assumed to be a finite dimensional one, and, obviously, A is also assumed to be invertible. -
5 the Dirac Equation and Spinors
5 The Dirac Equation and Spinors In this section we develop the appropriate wavefunctions for fundamental fermions and bosons. 5.1 Notation Review The three dimension differential operator is : ∂ ∂ ∂ = , , (5.1) ∂x ∂y ∂z We can generalise this to four dimensions ∂µ: 1 ∂ ∂ ∂ ∂ ∂ = , , , (5.2) µ c ∂t ∂x ∂y ∂z 5.2 The Schr¨odinger Equation First consider a classical non-relativistic particle of mass m in a potential U. The energy-momentum relationship is: p2 E = + U (5.3) 2m we can substitute the differential operators: ∂ Eˆ i pˆ i (5.4) → ∂t →− to obtain the non-relativistic Schr¨odinger Equation (with = 1): ∂ψ 1 i = 2 + U ψ (5.5) ∂t −2m For U = 0, the free particle solutions are: iEt ψ(x, t) e− ψ(x) (5.6) ∝ and the probability density ρ and current j are given by: 2 i ρ = ψ(x) j = ψ∗ ψ ψ ψ∗ (5.7) | | −2m − with conservation of probability giving the continuity equation: ∂ρ + j =0, (5.8) ∂t · Or in Covariant notation: µ µ ∂µj = 0 with j =(ρ,j) (5.9) The Schr¨odinger equation is 1st order in ∂/∂t but second order in ∂/∂x. However, as we are going to be dealing with relativistic particles, space and time should be treated equally. 25 5.3 The Klein-Gordon Equation For a relativistic particle the energy-momentum relationship is: p p = p pµ = E2 p 2 = m2 (5.10) · µ − | | Substituting the equation (5.4), leads to the relativistic Klein-Gordon equation: ∂2 + 2 ψ = m2ψ (5.11) −∂t2 The free particle solutions are plane waves: ip x i(Et p x) ψ e− · = e− − · (5.12) ∝ The Klein-Gordon equation successfully describes spin 0 particles in relativistic quan- tum field theory. -
A Some Basic Rules of Tensor Calculus
A Some Basic Rules of Tensor Calculus The tensor calculus is a powerful tool for the description of the fundamentals in con- tinuum mechanics and the derivation of the governing equations for applied prob- lems. In general, there are two possibilities for the representation of the tensors and the tensorial equations: – the direct (symbolic) notation and – the index (component) notation The direct notation operates with scalars, vectors and tensors as physical objects defined in the three dimensional space. A vector (first rank tensor) a is considered as a directed line segment rather than a triple of numbers (coordinates). A second rank tensor A is any finite sum of ordered vector pairs A = a b + ... +c d. The scalars, vectors and tensors are handled as invariant (independent⊗ from the choice⊗ of the coordinate system) objects. This is the reason for the use of the direct notation in the modern literature of mechanics and rheology, e.g. [29, 32, 49, 123, 131, 199, 246, 313, 334] among others. The index notation deals with components or coordinates of vectors and tensors. For a selected basis, e.g. gi, i = 1, 2, 3 one can write a = aig , A = aibj + ... + cidj g g i i ⊗ j Here the Einstein’s summation convention is used: in one expression the twice re- peated indices are summed up from 1 to 3, e.g. 3 3 k k ik ik a gk ∑ a gk, A bk ∑ A bk ≡ k=1 ≡ k=1 In the above examples k is a so-called dummy index. Within the index notation the basic operations with tensors are defined with respect to their coordinates, e. -
18.700 JORDAN NORMAL FORM NOTES These Are Some Supplementary Notes on How to Find the Jordan Normal Form of a Small Matrix. Firs
18.700 JORDAN NORMAL FORM NOTES These are some supplementary notes on how to find the Jordan normal form of a small matrix. First we recall some of the facts from lecture, next we give the general algorithm for finding the Jordan normal form of a linear operator, and then we will see how this works for small matrices. 1. Facts Throughout we will work over the field C of complex numbers, but if you like you may replace this with any other algebraically closed field. Suppose that V is a C-vector space of dimension n and suppose that T : V → V is a C-linear operator. Then the characteristic polynomial of T factors into a product of linear terms, and the irreducible factorization has the form m1 m2 mr cT (X) = (X − λ1) (X − λ2) ... (X − λr) , (1) for some distinct numbers λ1, . , λr ∈ C and with each mi an integer m1 ≥ 1 such that m1 + ··· + mr = n. Recall that for each eigenvalue λi, the eigenspace Eλi is the kernel of T − λiIV . We generalized this by defining for each integer k = 1, 2,... the vector subspace k k E(X−λi) = ker(T − λiIV ) . (2) It is clear that we have inclusions 2 e Eλi = EX−λi ⊂ E(X−λi) ⊂ · · · ⊂ E(X−λi) ⊂ .... (3) k k+1 Since dim(V ) = n, it cannot happen that each dim(E(X−λi) ) < dim(E(X−λi) ), for each e e +1 k = 1, . , n. Therefore there is some least integer ei ≤ n such that E(X−λi) i = E(X−λi) i . -
(VI.E) Jordan Normal Form
(VI.E) Jordan Normal Form Set V = Cn and let T : V ! V be any linear transformation, with distinct eigenvalues s1,..., sm. In the last lecture we showed that V decomposes into stable eigenspaces for T : s s V = W1 ⊕ · · · ⊕ Wm = ker (T − s1I) ⊕ · · · ⊕ ker (T − smI). Let B = fB1,..., Bmg be a basis for V subordinate to this direct sum and set B = [T j ] , so that k Wk Bk [T]B = diagfB1,..., Bmg. Each Bk has only sk as eigenvalue. In the event that A = [T]eˆ is s diagonalizable, or equivalently ker (T − skI) = ker(T − skI) for all k , B is an eigenbasis and [T]B is a diagonal matrix diagf s1,..., s1 ;...; sm,..., sm g. | {z } | {z } d1=dim W1 dm=dim Wm Otherwise we must perform further surgery on the Bk ’s separately, in order to transform the blocks Bk (and so the entire matrix for T ) into the “simplest possible” form. The attentive reader will have noticed above that I have written T − skI in place of skI − T . This is a strategic move: when deal- ing with characteristic polynomials it is far more convenient to write det(lI − A) to produce a monic polynomial. On the other hand, as you’ll see now, it is better to work on the individual Wk with the nilpotent transformation T j − s I =: N . Wk k k Decomposition of the Stable Eigenspaces (Take 1). Let’s briefly omit subscripts and consider T : W ! W with one eigenvalue s , dim W = d , B a basis for W and [T]B = B. -
Trace Inequalities for Matrices
Bull. Aust. Math. Soc. 87 (2013), 139–148 doi:10.1017/S0004972712000627 TRACE INEQUALITIES FOR MATRICES KHALID SHEBRAWI ˛ and HUSSIEN ALBADAWI (Received 23 February 2012; accepted 20 June 2012) Abstract Trace inequalities for sums and products of matrices are presented. Relations between the given inequalities and earlier results are discussed. Among other inequalities, it is shown that if A and B are positive semidefinite matrices then tr(AB)k ≤ minfkAkk tr Bk; kBkk tr Akg for any positive integer k. 2010 Mathematics subject classification: primary 15A18; secondary 15A42, 15A45. Keywords and phrases: trace inequalities, eigenvalues, singular values. 1. Introduction Let Mn(C) be the algebra of all n × n matrices over the complex number field. The singular values of A 2 Mn(C), denoted by s1(A); s2(A);:::; sn(A), are the eigenvalues ∗ 1=2 of the matrix jAj = (A A) arranged in such a way that s1(A) ≥ s2(A) ≥ · · · ≥ sn(A). 2 ∗ ∗ Note that si (A) = λi(A A) = λi(AA ), so for a positive semidefinite matrix A, we have si(A) = λi(A)(i = 1; 2;:::; n). The trace functional of A 2 Mn(C), denoted by tr A or tr(A), is defined to be the sum of the entries on the main diagonal of A and it is well known that the trace of a Pn matrix A is equal to the sum of its eigenvalues, that is, tr A = j=1 λ j(A). Two principal properties of the trace are that it is a linear functional and, for A; B 2 Mn(C), we have tr(AB) = tr(BA). -
Your PRINTED Name Is: Please Circle Your Recitation
18.06 Professor Edelman Quiz 3 December 3, 2012 Grading 1 Your PRINTED name is: 2 3 4 Please circle your recitation: 1 T 9 2-132 Andrey Grinshpun 2-349 3-7578 agrinshp 2 T 10 2-132 Rosalie Belanger-Rioux 2-331 3-5029 robr 3 T 10 2-146 Andrey Grinshpun 2-349 3-7578 agrinshp 4 T 11 2-132 Rosalie Belanger-Rioux 2-331 3-5029 robr 5 T 12 2-132 Georoy Horel 2-490 3-4094 ghorel 6 T 1 2-132 Tiankai Liu 2-491 3-4091 tiankai 7 T 2 2-132 Tiankai Liu 2-491 3-4091 tiankai 1 (16 pts.) a) (4 pts.) Suppose C is n × n and positive denite. If A is n × m and M = AT CA is not positive denite, nd the smallest eigenvalue of M: (Explain briey.) Solution. The smallest eigenvalue of M is 0. The problem only asks for brief explanations, but to help students understand the material better, I will give lengthy ones. First of all, note that M T = AT CT A = AT CA = M, so M is symmetric. That implies that all the eigenvalues of M are real. (Otherwise, the question wouldn't even make sense; what would the smallest of a set of complex numbers mean?) Since we are assuming that M is not positive denite, at least one of its eigenvalues must be nonpositive. So, to solve the problem, we just have to explain why M cannot have any negative eigenvalues. The explanation is that M is positive semidenite. -
The Weyr Characteristic
Swarthmore College Works Mathematics & Statistics Faculty Works Mathematics & Statistics 12-1-1999 The Weyr Characteristic Helene Shapiro Swarthmore College, [email protected] Follow this and additional works at: https://works.swarthmore.edu/fac-math-stat Part of the Mathematics Commons Let us know how access to these works benefits ouy Recommended Citation Helene Shapiro. (1999). "The Weyr Characteristic". American Mathematical Monthly. Volume 106, Issue 10. 919-929. DOI: 10.2307/2589746 https://works.swarthmore.edu/fac-math-stat/28 This work is brought to you for free by Swarthmore College Libraries' Works. It has been accepted for inclusion in Mathematics & Statistics Faculty Works by an authorized administrator of Works. For more information, please contact [email protected]. The Weyr Characteristic Author(s): Helene Shapiro Source: The American Mathematical Monthly, Vol. 106, No. 10 (Dec., 1999), pp. 919-929 Published by: Mathematical Association of America Stable URL: http://www.jstor.org/stable/2589746 Accessed: 19-10-2016 15:44 UTC JSTOR is a not-for-profit service that helps scholars, researchers, and students discover, use, and build upon a wide range of content in a trusted digital archive. We use information technology and tools to increase productivity and facilitate new forms of scholarship. For more information about JSTOR, please contact [email protected]. Your use of the JSTOR archive indicates your acceptance of the Terms & Conditions of Use, available at http://about.jstor.org/terms Mathematical Association of America is collaborating with JSTOR to digitize, preserve and extend access to The American Mathematical Monthly This content downloaded from 130.58.65.20 on Wed, 19 Oct 2016 15:44:49 UTC All use subject to http://about.jstor.org/terms The Weyr Characteristic Helene Shapiro 1. -
Lecture 28: Eigenvalues Allowing Complex Eigenvalues Is Really a Blessing
Math 19b: Linear Algebra with Probability Oliver Knill, Spring 2011 cos(t) sin(t) 2 2 For a rotation A = the characteristic polynomial is λ − 2 cos(α)+1 − sin(t) cos(t) which has the roots cos(α) ± i sin(α)= eiα. Lecture 28: Eigenvalues Allowing complex eigenvalues is really a blessing. The structure is very simple: Fundamental theorem of algebra: For a n × n matrix A, the characteristic We have seen that det(A) = 0 if and only if A is invertible. polynomial has exactly n roots. There are therefore exactly n eigenvalues of A if we count them with multiplicity. The polynomial fA(λ) = det(A − λIn) is called the characteristic polynomial of 1 n n−1 A. Proof One only has to show a polynomial p(z)= z + an−1z + ··· + a1z + a0 always has a root z0 We can then factor out p(z)=(z − z0)g(z) where g(z) is a polynomial of degree (n − 1) and The eigenvalues of A are the roots of the characteristic polynomial. use induction in n. Assume now that in contrary the polynomial p has no root. Cauchy’s integral theorem then tells dz 2πi Proof. If Av = λv,then v is in the kernel of A − λIn. Consequently, A − λIn is not invertible and = =0 . (1) z =r | | | zp(z) p(0) det(A − λIn)=0 . On the other hand, for all r, 2 1 dz 1 2π 1 For the matrix A = , the characteristic polynomial is | | ≤ 2πrmax|z|=r = . (2) z =r 4 −1 | | | zp(z) |zp(z)| min|z|=rp(z) The right hand side goes to 0 for r →∞ because 2 − λ 1 2 det(A − λI2) = det( )= λ − λ − 6 . -
Advanced Linear Algebra (MA251) Lecture Notes Contents
Algebra I – Advanced Linear Algebra (MA251) Lecture Notes Derek Holt and Dmitriy Rumynin year 2009 (revised at the end) Contents 1 Review of Some Linear Algebra 3 1.1 The matrix of a linear map with respect to a fixed basis . ........ 3 1.2 Changeofbasis................................... 4 2 The Jordan Canonical Form 4 2.1 Introduction.................................... 4 2.2 TheCayley-Hamiltontheorem . ... 6 2.3 Theminimalpolynomial. 7 2.4 JordanchainsandJordanblocks . .... 9 2.5 Jordan bases and the Jordan canonical form . ....... 10 2.6 The JCF when n =2and3 ............................ 11 2.7 Thegeneralcase .................................. 14 2.8 Examples ...................................... 15 2.9 Proof of Theorem 2.9 (non-examinable) . ...... 16 2.10 Applications to difference equations . ........ 17 2.11 Functions of matrices and applications to differential equations . 19 3 Bilinear Maps and Quadratic Forms 21 3.1 Bilinearmaps:definitions . 21 3.2 Bilinearmaps:changeofbasis . 22 3.3 Quadraticforms: introduction. ...... 22 3.4 Quadraticforms: definitions. ..... 25 3.5 Change of variable under the general linear group . .......... 26 3.6 Change of variable under the orthogonal group . ........ 29 3.7 Applications of quadratic forms to geometry . ......... 33 3.7.1 Reduction of the general second degree equation . ....... 33 3.7.2 The case n =2............................... 34 3.7.3 The case n =3............................... 34 1 3.8 Unitary, hermitian and normal matrices . ....... 35 3.9 Applications to quantum mechanics . ..... 41 4 Finitely Generated Abelian Groups 44 4.1 Definitions...................................... 44 4.2 Subgroups,cosetsandquotientgroups . ....... 45 4.3 Homomorphisms and the first isomorphism theorem . ....... 48 4.4 Freeabeliangroups............................... 50 4.5 Unimodular elementary row and column operations and the Smith normal formforintegralmatrices . -
A SHORT PROOF of the EXISTENCE of JORDAN NORMAL FORM Let V Be a Finite-Dimensional Complex Vector Space and Let T : V → V Be A
A SHORT PROOF OF THE EXISTENCE OF JORDAN NORMAL FORM MARK WILDON Let V be a finite-dimensional complex vector space and let T : V → V be a linear map. A fundamental theorem in linear algebra asserts that there is a basis of V in which T is represented by a matrix in Jordan normal form J1 0 ... 0 0 J2 ... 0 . . .. . 0 0 ...Jk where each Ji is a matrix of the form λ 1 ... 0 0 0 λ . 0 0 . . .. 0 0 . λ 1 0 0 ... 0 λ for some λ ∈ C. We shall assume that the usual reduction to the case where some power of T is the zero map has been made. (See [1, §58] for a characteristically clear account of this step.) After this reduction, it is sufficient to prove the following theorem. Theorem 1. If T : V → V is a linear transformation of a finite-dimensional vector space such that T m = 0 for some m ≥ 1, then there is a basis of V of the form a1−1 ak−1 u1, T u1,...,T u1, . , uk, T uk,...,T uk a where T i ui = 0 for 1 ≤ i ≤ k. At this point all the proofs the author has seen (even Halmos’ in [1, §57]) become unnecessarily long-winded. In this note we present a simple proof which leads to a straightforward algorithm for finding the required basis. Date: December 2007. 1 2 MARK WILDON Proof. We work by induction on dim V . For the inductive step we may assume that dim V ≥ 1.