Diagonally Scaled Permutations and Circulant Matrices
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Diagonalizing a Matrix
Diagonalizing a Matrix Definition 1. We say that two square matrices A and B are similar provided there exists an invertible matrix P so that . 2. We say a matrix A is diagonalizable if it is similar to a diagonal matrix. Example 1. The matrices and are similar matrices since . We conclude that is diagonalizable. 2. The matrices and are similar matrices since . After we have developed some additional theory, we will be able to conclude that the matrices and are not diagonalizable. Theorem Suppose A, B and C are square matrices. (1) A is similar to A. (2) If A is similar to B, then B is similar to A. (3) If A is similar to B and if B is similar to C, then A is similar to C. Proof of (3) Since A is similar to B, there exists an invertible matrix P so that . Also, since B is similar to C, there exists an invertible matrix R so that . Now, and so A is similar to C. Thus, “A is similar to B” is an equivalence relation. Theorem If A is similar to B, then A and B have the same eigenvalues. Proof Since A is similar to B, there exists an invertible matrix P so that . Now, Since A and B have the same characteristic equation, they have the same eigenvalues. > Example Find the eigenvalues for . Solution Since is similar to the diagonal matrix , they have the same eigenvalues. Because the eigenvalues of an upper (or lower) triangular matrix are the entries on the main diagonal, we see that the eigenvalues for , and, hence, are . -
Simultaneous Diagonalization and SVD of Commuting Matrices
Simultaneous Diagonalization and SVD of Commuting Matrices Ronald P. Nordgren1 Brown School of Engineering, Rice University Abstract. We present a matrix version of a known method of constructing common eigenvectors of two diagonalizable commuting matrices, thus enabling their simultane- ous diagonalization. The matrices may have simple eigenvalues of multiplicity greater than one. The singular value decomposition (SVD) of a class of commuting matrices also is treated. The effect of row/column permutation is examined. Examples are given. 1 Introduction It is well known that if two diagonalizable matrices have the same eigenvectors, then they commute. The converse also is true and a construction for the common eigen- vectors (enabling simultaneous diagonalization) is known. If one of the matrices has distinct eigenvalues (multiplicity one), it is easy to show that its eigenvectors pertain to both commuting matrices. The case of matrices with simple eigenvalues of mul- tiplicity greater than one requires a more complicated construction of their common eigenvectors. Here, we present a matrix version of a construction procedure given by Horn and Johnson [1, Theorem 1.3.12] and in a video by Sadun [4]. The eigenvector construction procedure also is applied to the singular value decomposition of a class of commuting matrices that includes the case where at least one of the matrices is real and symmetric. In addition, we consider row/column permutation of the commuting matrices. Three examples illustrate the eigenvector construction procedure. 2 Eigenvector Construction Let A and B be diagonalizable square matrices that commute, i.e. AB = BA (1) and their Jordan canonical forms read −1 −1 A = SADASA , B = SBDB SB . -
MATH 2370, Practice Problems
MATH 2370, Practice Problems Kiumars Kaveh Problem: Prove that an n × n complex matrix A is diagonalizable if and only if there is a basis consisting of eigenvectors of A. Problem: Let A : V ! W be a one-to-one linear map between two finite dimensional vector spaces V and W . Show that the dual map A0 : W 0 ! V 0 is surjective. Problem: Determine if the curve 2 2 2 f(x; y) 2 R j x + y + xy = 10g is an ellipse or hyperbola or union of two lines. Problem: Show that if a nilpotent matrix is diagonalizable then it is the zero matrix. Problem: Let P be a permutation matrix. Show that P is diagonalizable. Show that if λ is an eigenvalue of P then for some integer m > 0 we have λm = 1 (i.e. λ is an m-th root of unity). Hint: Note that P m = I for some integer m > 0. Problem: Show that if λ is an eigenvector of an orthogonal matrix A then jλj = 1. n Problem: Take a vector v 2 R and let H be the hyperplane orthogonal n n to v. Let R : R ! R be the reflection with respect to a hyperplane H. Prove that R is a diagonalizable linear map. Problem: Prove that if λ1; λ2 are distinct eigenvalues of a complex matrix A then the intersection of the generalized eigenspaces Eλ1 and Eλ2 is zero (this is part of the Spectral Theorem). 1 Problem: Let H = (hij) be a 2 × 2 Hermitian matrix. Use the Min- imax Principle to show that if λ1 ≤ λ2 are the eigenvalues of H then λ1 ≤ h11 ≤ λ2. -
The Algebra Generated by Three Commuting Matrices
THE ALGEBRA GENERATED BY THREE COMMUTING MATRICES B.A. SETHURAMAN Abstract. We present a survey of an open problem concerning the dimension of the algebra generated by three commuting matrices. This article concerns a problem in algebra that is completely elementary to state, yet, has proven tantalizingly difficult and is as yet unsolved. Consider C[A; B; C] , the C-subalgebra of the n × n matrices Mn(C) generated by three commuting matrices A, B, and C. Thus, C[A; B; C] consists of all C- linear combinations of \monomials" AiBjCk, where i, j, and k range from 0 to infinity. Note that C[A; B; C] and Mn(C) are naturally vector-spaces over C; moreover, C[A; B; C] is a subspace of Mn(C). The problem, quite simply, is this: Is the dimension of C[A; B; C] as a C vector space bounded above by n? 2 Note that the dimension of C[A; B; C] is at most n , because the dimen- 2 sion of Mn(C) is n . Asking for the dimension of C[A; B; C] to be bounded above by n when A, B, and C commute is to put considerable restrictions on C[A; B; C]: this is to require that C[A; B; C] occupy only a small portion of the ambient Mn(C) space in which it sits. Actually, the dimension of C[A; B; C] is already bounded above by some- thing slightly smaller than n2, thanks to a classical theorem of Schur ([16]), who showed that the maximum possible dimension of a commutative C- 2 subalgebra of Mn(C) is 1 + bn =4c. -
Arxiv:1705.10957V2 [Math.AC] 14 Dec 2017 Ideal Called Is and Commutator X Y Vrafield a Over Let Matrix Eae Leri Es Is E Sdfierlvn Notions
NEARLY COMMUTING MATRICES ZHIBEK KADYRSIZOVA Abstract. We prove that the algebraic set of pairs of matrices with a diag- onal commutator over a field of positive prime characteristic, its irreducible components, and their intersection are F -pure when the size of matrices is equal to 3. Furthermore, we show that this algebraic set is reduced and the intersection of its irreducible components is irreducible in any characteristic for pairs of matrices of any size. In addition, we discuss various conjectures on the singularities of these algebraic sets and the system of parameters on the corresponding coordinate rings. Keywords: Frobenius, singularities, F -purity, commuting matrices 1. Introduction and preliminaries In this paper we study algebraic sets of pairs of matrices such that their commutator is either nonzero diagonal or zero. We also consider some other related algebraic sets. First let us define relevant notions. Let X =(x ) and Y =(y ) be n×n matrices of indeterminates arXiv:1705.10957v2 [math.AC] 14 Dec 2017 ij 1≤i,j≤n ij 1≤i,j≤n over a field K. Let R = K[X,Y ] be the polynomial ring in {xij,yij}1≤i,j≤n and let I denote the ideal generated by the off-diagonal entries of the commutator matrix XY − YX and J denote the ideal generated by the entries of XY − YX. The ideal I defines the algebraic set of pairs of matrices with a diagonal commutator and is called the algebraic set of nearly commuting matrices. The ideal J defines the algebraic set of pairs of commuting matrices. -
3.3 Diagonalization
3.3 Diagonalization −4 1 1 1 Let A = 0 1. Then 0 1 and 0 1 are eigenvectors of A, with corresponding @ 4 −4 A @ 2 A @ −2 A eigenvalues −2 and −6 respectively (check). This means −4 1 1 1 −4 1 1 1 0 1 0 1 = −2 0 1 ; 0 1 0 1 = −6 0 1 : @ 4 −4 A @ 2 A @ 2 A @ 4 −4 A @ −2 A @ −2 A Thus −4 1 1 1 1 1 −2 −6 0 1 0 1 = 0−2 0 1 − 6 0 11 = 0 1 @ 4 −4 A @ 2 −2 A @ @ −2 A @ −2 AA @ −4 12 A We have −4 1 1 1 1 1 −2 0 0 1 0 1 = 0 1 0 1 @ 4 −4 A @ 2 −2 A @ 2 −2 A @ 0 −6 A 1 1 (Think about this). Thus AE = ED where E = 0 1 has the eigenvectors of A as @ 2 −2 A −2 0 columns and D = 0 1 is the diagonal matrix having the eigenvalues of A on the @ 0 −6 A main diagonal, in the order in which their corresponding eigenvectors appear as columns of E. Definition 3.3.1 A n × n matrix is A diagonal if all of its non-zero entries are located on its main diagonal, i.e. if Aij = 0 whenever i =6 j. Diagonal matrices are particularly easy to handle computationally. If A and B are diagonal n × n matrices then the product AB is obtained from A and B by simply multiplying entries in corresponding positions along the diagonal, and AB = BA. -
Problems in Abstract Algebra
STUDENT MATHEMATICAL LIBRARY Volume 82 Problems in Abstract Algebra A. R. Wadsworth 10.1090/stml/082 STUDENT MATHEMATICAL LIBRARY Volume 82 Problems in Abstract Algebra A. R. Wadsworth American Mathematical Society Providence, Rhode Island Editorial Board Satyan L. Devadoss John Stillwell (Chair) Erica Flapan Serge Tabachnikov 2010 Mathematics Subject Classification. Primary 00A07, 12-01, 13-01, 15-01, 20-01. For additional information and updates on this book, visit www.ams.org/bookpages/stml-82 Library of Congress Cataloging-in-Publication Data Names: Wadsworth, Adrian R., 1947– Title: Problems in abstract algebra / A. R. Wadsworth. Description: Providence, Rhode Island: American Mathematical Society, [2017] | Series: Student mathematical library; volume 82 | Includes bibliographical references and index. Identifiers: LCCN 2016057500 | ISBN 9781470435837 (alk. paper) Subjects: LCSH: Algebra, Abstract – Textbooks. | AMS: General – General and miscellaneous specific topics – Problem books. msc | Field theory and polyno- mials – Instructional exposition (textbooks, tutorial papers, etc.). msc | Com- mutative algebra – Instructional exposition (textbooks, tutorial papers, etc.). msc | Linear and multilinear algebra; matrix theory – Instructional exposition (textbooks, tutorial papers, etc.). msc | Group theory and generalizations – Instructional exposition (textbooks, tutorial papers, etc.). msc Classification: LCC QA162 .W33 2017 | DDC 512/.02–dc23 LC record available at https://lccn.loc.gov/2016057500 Copying and reprinting. Individual readers of this publication, and nonprofit libraries acting for them, are permitted to make fair use of the material, such as to copy select pages for use in teaching or research. Permission is granted to quote brief passages from this publication in reviews, provided the customary acknowledgment of the source is given. Republication, systematic copying, or multiple reproduction of any material in this publication is permitted only under license from the American Mathematical Society. -
R'kj.Oti-1). (3) the Object of the Present Article Is to Make This Estimate Effective
TRANSACTIONS OF THE AMERICAN MATHEMATICAL SOCIETY Volume 259, Number 2, June 1980 EFFECTIVE p-ADIC BOUNDS FOR SOLUTIONS OF HOMOGENEOUS LINEAR DIFFERENTIAL EQUATIONS BY B. DWORK AND P. ROBBA Dedicated to K. Iwasawa Abstract. We consider a finite set of power series in one variable with coefficients in a field of characteristic zero having a chosen nonarchimedean valuation. We study the growth of these series near the boundary of their common "open" disk of convergence. Our results are definitive when the wronskian is bounded. The main application involves local solutions of ordinary linear differential equations with analytic coefficients. The effective determination of the common radius of conver- gence remains open (and is not treated here). Let K be an algebraically closed field of characteristic zero complete under a nonarchimedean valuation with residue class field of characteristic p. Let D = d/dx L = D"+Cn_lD'-l+ ■ ■■ +C0 (1) be a linear differential operator with coefficients meromorphic in some neighbor- hood of the origin. Let u = a0 + a,jc + . (2) be a power series solution of L which converges in an open (/>-adic) disk of radius r. Our object is to describe the asymptotic behavior of \a,\rs as s —*oo. In a series of articles we have shown that subject to certain restrictions we may conclude that r'KJ.Oti-1). (3) The object of the present article is to make this estimate effective. At the same time we greatly simplify, and generalize, our best previous results [12] for the noneffective form. Our previous work was based on the notion of a generic disk together with a condition for reducibility of differential operators with unbounded solutions [4, Theorem 4]. -
Lecture 2: Spectral Theorems
Lecture 2: Spectral Theorems This lecture introduces normal matrices. The spectral theorem will inform us that normal matrices are exactly the unitarily diagonalizable matrices. As a consequence, we will deduce the classical spectral theorem for Hermitian matrices. The case of commuting families of matrices will also be studied. All of this corresponds to section 2.5 of the textbook. 1 Normal matrices Definition 1. A matrix A 2 Mn is called a normal matrix if AA∗ = A∗A: Observation: The set of normal matrices includes all the Hermitian matrices (A∗ = A), the skew-Hermitian matrices (A∗ = −A), and the unitary matrices (AA∗ = A∗A = I). It also " # " # 1 −1 1 1 contains other matrices, e.g. , but not all matrices, e.g. 1 1 0 1 Here is an alternate characterization of normal matrices. Theorem 2. A matrix A 2 Mn is normal iff ∗ n kAxk2 = kA xk2 for all x 2 C : n Proof. If A is normal, then for any x 2 C , 2 ∗ ∗ ∗ ∗ ∗ 2 kAxk2 = hAx; Axi = hx; A Axi = hx; AA xi = hA x; A xi = kA xk2: ∗ n n Conversely, suppose that kAxk = kA xk for all x 2 C . For any x; y 2 C and for λ 2 C with jλj = 1 chosen so that <(λhx; (A∗A − AA∗)yi) = jhx; (A∗A − AA∗)yij, we expand both sides of 2 ∗ 2 kA(λx + y)k2 = kA (λx + y)k2 to obtain 2 2 ∗ 2 ∗ 2 ∗ ∗ kAxk2 + kAyk2 + 2<(λhAx; Ayi) = kA xk2 + kA yk2 + 2<(λhA x; A yi): 2 ∗ 2 2 ∗ 2 Using the facts that kAxk2 = kA xk2 and kAyk2 = kA yk2, we derive 0 = <(λhAx; Ayi − λhA∗x; A∗yi) = <(λhx; A∗Ayi − λhx; AA∗yi) = <(λhx; (A∗A − AA∗)yi) = jhx; (A∗A − AA∗)yij: n ∗ ∗ n Since this is true for any x 2 C , we deduce (A A − AA )y = 0, which holds for any y 2 C , meaning that A∗A − AA∗ = 0, as desired. -
Diagonalizable Matrix - Wikipedia, the Free Encyclopedia
Diagonalizable matrix - Wikipedia, the free encyclopedia http://en.wikipedia.org/wiki/Matrix_diagonalization Diagonalizable matrix From Wikipedia, the free encyclopedia (Redirected from Matrix diagonalization) In linear algebra, a square matrix A is called diagonalizable if it is similar to a diagonal matrix, i.e., if there exists an invertible matrix P such that P −1AP is a diagonal matrix. If V is a finite-dimensional vector space, then a linear map T : V → V is called diagonalizable if there exists a basis of V with respect to which T is represented by a diagonal matrix. Diagonalization is the process of finding a corresponding diagonal matrix for a diagonalizable matrix or linear map.[1] A square matrix which is not diagonalizable is called defective. Diagonalizable matrices and maps are of interest because diagonal matrices are especially easy to handle: their eigenvalues and eigenvectors are known and one can raise a diagonal matrix to a power by simply raising the diagonal entries to that same power. Geometrically, a diagonalizable matrix is an inhomogeneous dilation (or anisotropic scaling) — it scales the space, as does a homogeneous dilation, but by a different factor in each direction, determined by the scale factors on each axis (diagonal entries). Contents 1 Characterisation 2 Diagonalization 3 Simultaneous diagonalization 4 Examples 4.1 Diagonalizable matrices 4.2 Matrices that are not diagonalizable 4.3 How to diagonalize a matrix 4.3.1 Alternative Method 5 An application 5.1 Particular application 6 Quantum mechanical application 7 See also 8 Notes 9 References 10 External links Characterisation The fundamental fact about diagonalizable maps and matrices is expressed by the following: An n-by-n matrix A over the field F is diagonalizable if and only if the sum of the dimensions of its eigenspaces is equal to n, which is the case if and only if there exists a basis of Fn consisting of eigenvectors of A. -
EIGENVALUES and EIGENVECTORS 1. Diagonalizable Linear Transformations and Matrices Recall, a Matrix, D, Is Diagonal If It Is
EIGENVALUES AND EIGENVECTORS 1. Diagonalizable linear transformations and matrices Recall, a matrix, D, is diagonal if it is square and the only non-zero entries are on the diagonal. This is equivalent to D~ei = λi~ei where here ~ei are the standard n n vector and the λi are the diagonal entries. A linear transformation, T : R ! R , is n diagonalizable if there is a basis B of R so that [T ]B is diagonal. This means [T ] is n×n similar to the diagonal matrix [T ]B. Similarly, a matrix A 2 R is diagonalizable if it is similar to some diagonal matrix D. To diagonalize a linear transformation is to find a basis B so that [T ]B is diagonal. To diagonalize a square matrix is to find an invertible S so that S−1AS = D is diagonal. Fix a matrix A 2 Rn×n We say a vector ~v 2 Rn is an eigenvector if (1) ~v 6= 0. (2) A~v = λ~v for some scalar λ 2 R. The scalar λ is the eigenvalue associated to ~v or just an eigenvalue of A. Geo- metrically, A~v is parallel to ~v and the eigenvalue, λ. counts the stretching factor. Another way to think about this is that the line L := span(~v) is left invariant by multiplication by A. n An eigenbasis of A is a basis, B = (~v1; : : : ;~vn) of R so that each ~vi is an eigenvector of A. Theorem 1.1. The matrix A is diagonalizable if and only if there is an eigenbasis of A. -
Arxiv:1508.00183V2 [Math.RA] 6 Aug 2015 Atclrb Ed Yasemigroup a by field
A THEOREM OF KAPLANSKY REVISITED HEYDAR RADJAVI AND BAMDAD R. YAHAGHI Abstract. We present a new and simple proof of a theorem due to Kaplansky which unifies theorems of Kolchin and Levitzki on triangularizability of semigroups of matrices. We also give two different extensions of the theorem. As a consequence, we prove the counterpart of Kolchin’s Theorem for finite groups of unipotent matrices over division rings. We also show that the counterpart of Kolchin’s Theorem over division rings of characteristic zero implies that of Kaplansky’s Theorem over such division rings. 1. Introduction The purpose of this short note is three-fold. We give a simple proof of Kaplansky’s Theorem on the (simultaneous) triangularizability of semi- groups whose members all have singleton spectra. Our proof, although not independent of Kaplansky’s, avoids the deeper group theoretical aspects present in it and in other existing proofs we are aware of. Also, this proof can be adjusted to give an affirmative answer to Kolchin’s Problem for finite groups of unipotent matrices over division rings. In particular, it follows from this proof that the counterpart of Kolchin’s Theorem over division rings of characteristic zero implies that of Ka- plansky’s Theorem over such division rings. We also present extensions of Kaplansky’s Theorem in two different directions. M arXiv:1508.00183v2 [math.RA] 6 Aug 2015 Let us fix some notation. Let ∆ be a division ring and n(∆) the algebra of all n × n matrices over ∆. The division ring ∆ could in particular be a field. By a semigroup S ⊆ Mn(∆), we mean a set of matrices closed under multiplication.