Computation of the Jordan Normal Form of a Matrix

Total Page:16

File Type:pdf, Size:1020Kb

Load more

COMPUTATION OF THE JORDAN NORMAL FORM OF A MATRIX USING VERSAL DEFORMATIONS ALEXEI A. MAILYBAEV Abstract. Numerical problem of nding multiple eigenvalues of real and complex nonsymmet- ric matrices is considered. Using versal deformation theory, explicit formulae for approximations of matrices having nonderogatory multiple eigenvalues in the vicinity of a given matrix A are de- rived. These formulae provide lo cal information on the geometry of a set of matrices having a given nonderogatory Jordan structure, distance to the nearest matrix with a multiple eigenvalue of given multiplicity,values of multiple eigenvalues, and corresp onding Jordan chains (generalized eigenvec- tors). The formulae use only eigenvalues and eigenvectors of the initial matrix A,whichtypically has only simple eigenvalues. Both the case of matrix families (matrices smo othly dep endentonseveral parameters) and the case of real or complex matrix spaces are studied. Several examples showing simplicity and eciency of the suggested metho d are given. Key words. Jordan normal form, multiple eigenvalue, generalized eigenvector, matrix family, versal deformation, bifurcation diagram AMS sub ject classi cations. 65F15, 15A18, 15A21 1. Intro duction. Computation of the Jordan normal form (JNF) of a square nonsymmetric matrix A is needed in many branches of mathematics, mechanics, and physics. Due to its great imp ortance this problem was intensively studied during the last century. It is well known that a generic (typical) matrix has only simple eigenvalues, which means that its JNF is a diagonal matrix [1,2]. The problem of nding the JNF of a generic matrix has b een well studied b oth from theoretical and practical (numerical) sides; see [24]. Nevertheless, manyinteresting and imp or- tant phenomena, asso ciated with qualitativechanges in the dynamics of mechanical systems [17,18, 22], stability optimization [3, 14, 20], and eigenvalue b ehavior under matrix p erturbations [4, 21, 23], are connected with degenerate matrices, i.e., matrices having multiple eigenvalues. In the presence of multiple eigenvalues the numerical problem of calculation of the JNF b ecomes very complicated due to its instability, since arbitrarily small p ertur- bations (caused, for example, by round-o errors) destroy the nongeneric structure. This means that we should consider a matrix given with some accuracy,thus, substi- tuting the matrix A by its neighb or in the matrix space. Suchformulation leads to several imp ortant problems. The rst problem left op en by Wilkinson [25, 26] is to nd a distance from a given matrix A to the nearest nondiagonalizable (degenerate) b matrix A. This problem can b e reformulated in the extended form: to nd a distance (with resp ect to a given norm) from a matrix A to the b nearest matrix A having a given Jordan structure; b to nd a matrix A with the least generic Jordan structure in a given neigh- b orho o d of the matrix A (the least generic Jordan structure means that a set of matrices with this Jordan structure has the highest co dimension). Here we assume that two matrices have the same Jordan structure, if their JNFs di er only byvalues of di erent eigenvalues. Co dimensions of sets of matrices having the same Jordan structure were studied by Arnold [1, 2]. According to Arnold, matrices b A with the least generic Jordan structure have the most p owerful in uence on the Institute of Mechanics, Moscow State Lomonosov University,Michurinsky pr. 1, 117192 Moscow, Russia ([email protected]). 1 2 ALEXEI A. MAILYBAEV eigenvalue b ehavior in the neighb orho o d of the matrix A. Therefore, the ab ove stated problems represent the key p oint in the numerical analysis of multiple eigenvalues. Several authors addressed the problem of nding b ounds on the distance to the nearest degenerate matrix; see [5,6,9,15,25,26] and more references therein. A nu- merical metho d for nding the JNF of a degenerate matrix given with some accuracy was prop osed in [12,13]. This metho d is based on making a numb er of successive transformations of the matrix and deleting small elements whenever appropriate. Nev- ertheless, this pro cedure do esn't provide the least generic JNF, and it is restricted to the use of the Frob enius matrix norm (for example, it can not b e used in the case, when di erententries of a matrix are given with di erent accuracy). Another metho d based on singular value decomp ositions was given in [19]. This metho d is simpler, but it has the same prop erties. More references of JNF computation metho ds can b e found in [8]. It is understo o d now that the problem of calculation of the nondiagonal JNF needs the singularity theory approach based on the use of versal deformations (lo cally generic families of matrices). Versal deformations were intro duced by Arnold [1, 2] in order to stabilize the JNF transformation, but no numerical metho ds were suggested. Fairgrieve[9] succeeded in using the idea of versality for construction of a stable numerical pro cedure, which calculates matrices of a given Jordan structure. Though this metho d do es not provide the nearest matrix of a given Jordan structure, it shows the usefulness of the singularity theory for the numerical eigenvalue problem. In [7] versal deformations were used for improvement of an existing numerical program in a similar problem of nding the Kronecker normal form of a matrix p encil. Application of the singularity theory to computation of a distance from a matrix A to the nearest matrix with a double eigenvalue [15] showed that this numerical problem is very complex even in the case of a double eigenvalue. This pap er is intended to prove that the versal deformation theory can b e an indep endentpowerful to ol for analysis of multiple eigenvalues. It is shown that us- ing versal deformations wecanavoid a numb er of transformations and singular value decomp ositions (used in most algorithms). Instead of this, explicit approximate for- mulae describing the geometry of a set of matrices with given Jordan structure in the vicinityofa given matrix A are derived. These formulae use only information on eigenvalues and eigenvectors of the matrix A. Since the matrix A is typically nondegenerate, i.e., has only simple eigenvalues, its eigenvalues and eigenvectors can b e obtained by means of standart co des. As a result, we obtain approximations of a distance from A to the nearest matrix having a given Jordan structure, its multiple eigenvalues, and the Jordan chains (generalized eigenvectors). The metho d do es not dep end on the norm used in the matrix space. Moreover, it allows analyzing multiple eigenvalues of matrices dep endent on several parameters (matrix families), where we can not vary matrix elements indep endently.Approximations derived in the pap er 2 have the accuracy O (" ), where " is the distance from A to a matrix we are lo ok- ing for. All these prop erties make the suggested approach useful and attractive for applications. In this pap er the case of a nonderogatory JNF (every eigenvalue has one corre- sp onding Jordan blo ck) is considered. Derogatory Jordan structures require further investigation and new ideas. Diculties app earing in the derogatory case are asso ci- ated with the lack of uniqueness of the transformation to a versal deformation and will b e discussed in the conclusion. The pap er is organized as follows. In section 2 some concepts of the singularity COMPUTATION OF THE JORDAN NORMAL FORM 3 theory are intro duced and the general idea of the pap er is describ ed. Sections 3 and 4 give approximate formulae for matrices having a prescrib ed nonderogatory Jordan structure in the neighb orho o d of a given generic (nondegenerate) complex or real matrix A. In section 5 the same problem is studied in the case, when the initial matrix A is degenerate. Conclusion discusses extension of the metho d to the derogatory case and p ossibilities for its application to sp eci c typ es of matrices or matrix p encils. In the pap er matrices are denoted by b old capital letters, vectors take the form of b old small letters, and scalars are represented by small italic characters. 2. Bifurcation diagram. Let us consider an m m complex (real) nonsym- metric matrix A holomorphically (smo othly) dep endentonavector of complex (real) parameters p =(p ;:::;p ). The space of all m m matrices can b e considered as a 1 n sp ecial case, where all entries of A are indep endent parameters. Matrices A and B are said to have the same Jordan structure if there exists one- to-one corresp ondence b etween di erent eigenvalues of A and B such that sizes of Jordan blo cks in the JNFs of A and B are equal for corresp onding eigenvalues (JNFs of the matrices A and B di er only byvalues of their eigenvalues). For example, among the following matrices 1 1 0 1 0 0 2 1 0 3 0 0 5 1 0 A A @ A @ @ (2.1) ; C = 0 2 0 ; B = 0 2 1 A = 0 5 0 0 0 2 0 0 2 0 0 2 the matrices A and B have the same Jordan structure, while the matrix C has a di erent Jordan structure. In the generic case a set of values of the vector p corresp onding to the matrices A(p)having the same Jordan structure represents a smo oth submanifold of the parameter space [1, 2]. This submanifold is called a stratum (also called a colored orbit ofamatrix).Abifurcation diagram of the matrix family A(p) is a partition (strati cation) of the parameter space according to the Jordan structure of the matrix A(p)[1, 2].
Recommended publications
  • Random Matrix Theory and Covariance Estimation

    Random Matrix Theory and Covariance Estimation

    Introduction Random matrix theory Estimating correlations Comparison with Barra Conclusion Appendix Random Matrix Theory and Covariance Estimation Jim Gatheral New York, October 3, 2008 Introduction Random matrix theory Estimating correlations Comparison with Barra Conclusion Appendix Motivation Sophisticated optimal liquidation portfolio algorithms that balance risk against impact cost involve inverting the covariance matrix. Eigenvalues of the covariance matrix that are small (or even zero) correspond to portfolios of stocks that have nonzero returns but extremely low or vanishing risk; such portfolios are invariably related to estimation errors resulting from insuffient data. One of the approaches used to eliminate the problem of small eigenvalues in the estimated covariance matrix is the so-called random matrix technique. We would like to understand: the basis of random matrix theory. (RMT) how to apply RMT to the estimation of covariance matrices. whether the resulting covariance matrix performs better than (for example) the Barra covariance matrix. Introduction Random matrix theory Estimating correlations Comparison with Barra Conclusion Appendix Outline 1 Random matrix theory Random matrix examples Wigner’s semicircle law The Marˇcenko-Pastur density The Tracy-Widom law Impact of fat tails 2 Estimating correlations Uncertainty in correlation estimates. Example with SPX stocks A recipe for filtering the sample correlation matrix 3 Comparison with Barra Comparison of eigenvectors The minimum variance portfolio Comparison of weights In-sample and out-of-sample performance 4 Conclusions 5 Appendix with a sketch of Wigner’s original proof Introduction Random matrix theory Estimating correlations Comparison with Barra Conclusion Appendix Example 1: Normal random symmetric matrix Generate a 5,000 x 5,000 random symmetric matrix with entries aij N(0, 1).
  • Rotation Matrix Sampling Scheme for Multidimensional Probability Distribution Transfer

    Rotation Matrix Sampling Scheme for Multidimensional Probability Distribution Transfer

    ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Volume III-7, 2016 XXIII ISPRS Congress, 12–19 July 2016, Prague, Czech Republic ROTATION MATRIX SAMPLING SCHEME FOR MULTIDIMENSIONAL PROBABILITY DISTRIBUTION TRANSFER P. Srestasathiern ∗, S. Lawawirojwong, R. Suwantong, and P. Phuthong Geo-Informatics and Space Technology Development Agency (GISTDA), Laksi, Bangkok, Thailand (panu,siam,rata,pattrawuth)@gistda.or.th Commission VII,WG VII/4 KEY WORDS: Histogram specification, Distribution transfer, Color transfer, Data Gaussianization ABSTRACT: This paper address the problem of rotation matrix sampling used for multidimensional probability distribution transfer. The distri- bution transfer has many applications in remote sensing and image processing such as color adjustment for image mosaicing, image classification, and change detection. The sampling begins with generating a set of random orthogonal matrix samples by Householder transformation technique. The advantage of using the Householder transformation for generating the set of orthogonal matrices is the uniform distribution of the orthogonal matrix samples. The obtained orthogonal matrices are then converted to proper rotation matrices. The performance of using the proposed rotation matrix sampling scheme was tested against the uniform rotation angle sampling. The applications of the proposed method were also demonstrated using two applications i.e., image to image probability distribution transfer and data Gaussianization. 1. INTRODUCTION is based on the Monge’s transport problem which is to find mini- mal displacement for mass transportation. The solution from this In remote sensing, the analysis of multi-temporal data is widely linear approach can be used as initial solution for non-linear prob- used in many applications such as urban expansion monitoring, ability distribution transfer methods.
  • New Algorithms in the Frobenius Matrix Algebra for Polynomial Root-Finding

    New Algorithms in the Frobenius Matrix Algebra for Polynomial Root-Finding

    City University of New York (CUNY) CUNY Academic Works Computer Science Technical Reports CUNY Academic Works 2014 TR-2014006: New Algorithms in the Frobenius Matrix Algebra for Polynomial Root-Finding Victor Y. Pan Ai-Long Zheng How does access to this work benefit ou?y Let us know! More information about this work at: https://academicworks.cuny.edu/gc_cs_tr/397 Discover additional works at: https://academicworks.cuny.edu This work is made publicly available by the City University of New York (CUNY). Contact: [email protected] New Algoirthms in the Frobenius Matrix Algebra for Polynomial Root-finding ∗ Victor Y. Pan[1,2],[a] and Ai-Long Zheng[2],[b] Supported by NSF Grant CCF-1116736 and PSC CUNY Award 64512–0042 [1] Department of Mathematics and Computer Science Lehman College of the City University of New York Bronx, NY 10468 USA [2] Ph.D. Programs in Mathematics and Computer Science The Graduate Center of the City University of New York New York, NY 10036 USA [a] [email protected] http://comet.lehman.cuny.edu/vpan/ [b] [email protected] Abstract In 1996 Cardinal applied fast algorithms in Frobenius matrix algebra to complex root-finding for univariate polynomials, but he resorted to some numerically unsafe techniques of symbolic manipulation with polynomials at the final stages of his algorithms. We extend his work to complete the computations by operating with matrices at the final stage as well and also to adjust them to real polynomial root-finding. Our analysis and experiments show efficiency of the resulting algorithms. 2000 Math.
  • The Distribution of Eigenvalues of Randomized Permutation Matrices Tome 63, No 3 (2013), P

    The Distribution of Eigenvalues of Randomized Permutation Matrices Tome 63, No 3 (2013), P

    R AN IE N R A U L E O S F D T E U L T I ’ I T N S ANNALES DE L’INSTITUT FOURIER Joseph NAJNUDEL & Ashkan NIKEGHBALI The distribution of eigenvalues of randomized permutation matrices Tome 63, no 3 (2013), p. 773-838. <http://aif.cedram.org/item?id=AIF_2013__63_3_773_0> © Association des Annales de l’institut Fourier, 2013, tous droits réservés. L’accès aux articles de la revue « Annales de l’institut Fourier » (http://aif.cedram.org/), implique l’accord avec les conditions générales d’utilisation (http://aif.cedram.org/legal/). Toute re- production en tout ou partie de cet article sous quelque forme que ce soit pour tout usage autre que l’utilisation à fin strictement per- sonnelle du copiste est constitutive d’une infraction pénale. Toute copie ou impression de ce fichier doit contenir la présente mention de copyright. cedram Article mis en ligne dans le cadre du Centre de diffusion des revues académiques de mathématiques http://www.cedram.org/ Ann. Inst. Fourier, Grenoble 63, 3 (2013) 773-838 THE DISTRIBUTION OF EIGENVALUES OF RANDOMIZED PERMUTATION MATRICES by Joseph NAJNUDEL & Ashkan NIKEGHBALI Abstract. — In this article we study in detail a family of random matrix ensembles which are obtained from random permutations matrices (chosen at ran- dom according to the Ewens measure of parameter θ > 0) by replacing the entries equal to one by more general non-vanishing complex random variables. For these ensembles, in contrast with more classical models as the Gaussian Unitary En- semble, or the Circular Unitary Ensemble, the eigenvalues can be very explicitly computed by using the cycle structure of the permutations.
  • Subspace-Preserving Sparsification of Matrices with Minimal Perturbation

    Subspace-Preserving Sparsification of Matrices with Minimal Perturbation

    Subspace-preserving sparsification of matrices with minimal perturbation to the near null-space. Part I: Basics Chetan Jhurani Tech-X Corporation 5621 Arapahoe Ave Boulder, Colorado 80303, U.S.A. Abstract This is the first of two papers to describe a matrix sparsification algorithm that takes a general real or complex matrix as input and produces a sparse output matrix of the same size. The non-zero entries in the output are chosen to minimize changes to the singular values and singular vectors corresponding to the near null-space of the input. The output matrix is constrained to preserve left and right null-spaces exactly. The sparsity pattern of the output matrix is automatically determined or can be given as input. If the input matrix belongs to a common matrix subspace, we prove that the computed sparse matrix belongs to the same subspace. This works with- out imposing explicit constraints pertaining to the subspace. This property holds for the subspaces of Hermitian, complex-symmetric, Hamiltonian, cir- culant, centrosymmetric, and persymmetric matrices, and for each of the skew counterparts. Applications of our method include computation of reusable sparse pre- conditioning matrices for reliable and efficient solution of high-order finite element systems. The second paper in this series [1] describes our open- arXiv:1304.7049v1 [math.NA] 26 Apr 2013 source implementation, and presents further technical details. Keywords: Sparsification, Spectral equivalence, Matrix structure, Convex optimization, Moore-Penrose pseudoinverse Email address: [email protected] (Chetan Jhurani) Preprint submitted to Computers and Mathematics with Applications October 10, 2018 1. Introduction We present and analyze a matrix-valued optimization problem formulated to sparsify matrices while preserving the matrix null-spaces and certain spe- cial structural properties.
  • 18.700 JORDAN NORMAL FORM NOTES These Are Some Supplementary Notes on How to Find the Jordan Normal Form of a Small Matrix. Firs

    18.700 JORDAN NORMAL FORM NOTES These Are Some Supplementary Notes on How to Find the Jordan Normal Form of a Small Matrix. Firs

    18.700 JORDAN NORMAL FORM NOTES These are some supplementary notes on how to find the Jordan normal form of a small matrix. First we recall some of the facts from lecture, next we give the general algorithm for finding the Jordan normal form of a linear operator, and then we will see how this works for small matrices. 1. Facts Throughout we will work over the field C of complex numbers, but if you like you may replace this with any other algebraically closed field. Suppose that V is a C-vector space of dimension n and suppose that T : V → V is a C-linear operator. Then the characteristic polynomial of T factors into a product of linear terms, and the irreducible factorization has the form m1 m2 mr cT (X) = (X − λ1) (X − λ2) ... (X − λr) , (1) for some distinct numbers λ1, . , λr ∈ C and with each mi an integer m1 ≥ 1 such that m1 + ··· + mr = n. Recall that for each eigenvalue λi, the eigenspace Eλi is the kernel of T − λiIV . We generalized this by defining for each integer k = 1, 2,... the vector subspace k k E(X−λi) = ker(T − λiIV ) . (2) It is clear that we have inclusions 2 e Eλi = EX−λi ⊂ E(X−λi) ⊂ · · · ⊂ E(X−λi) ⊂ .... (3) k k+1 Since dim(V ) = n, it cannot happen that each dim(E(X−λi) ) < dim(E(X−λi) ), for each e e +1 k = 1, . , n. Therefore there is some least integer ei ≤ n such that E(X−λi) i = E(X−λi) i .
  • (VI.E) Jordan Normal Form

    (VI.E) Jordan Normal Form

    (VI.E) Jordan Normal Form Set V = Cn and let T : V ! V be any linear transformation, with distinct eigenvalues s1,..., sm. In the last lecture we showed that V decomposes into stable eigenspaces for T : s s V = W1 ⊕ · · · ⊕ Wm = ker (T − s1I) ⊕ · · · ⊕ ker (T − smI). Let B = fB1,..., Bmg be a basis for V subordinate to this direct sum and set B = [T j ] , so that k Wk Bk [T]B = diagfB1,..., Bmg. Each Bk has only sk as eigenvalue. In the event that A = [T]eˆ is s diagonalizable, or equivalently ker (T − skI) = ker(T − skI) for all k , B is an eigenbasis and [T]B is a diagonal matrix diagf s1,..., s1 ;...; sm,..., sm g. | {z } | {z } d1=dim W1 dm=dim Wm Otherwise we must perform further surgery on the Bk ’s separately, in order to transform the blocks Bk (and so the entire matrix for T ) into the “simplest possible” form. The attentive reader will have noticed above that I have written T − skI in place of skI − T . This is a strategic move: when deal- ing with characteristic polynomials it is far more convenient to write det(lI − A) to produce a monic polynomial. On the other hand, as you’ll see now, it is better to work on the individual Wk with the nilpotent transformation T j − s I =: N . Wk k k Decomposition of the Stable Eigenspaces (Take 1). Let’s briefly omit subscripts and consider T : W ! W with one eigenvalue s , dim W = d , B a basis for W and [T]B = B.
  • New Results on the Spectral Density of Random Matrices

    New Results on the Spectral Density of Random Matrices

    New Results on the Spectral Density of Random Matrices Timothy Rogers Department of Mathematics, King’s College London, July 2010 A thesis submitted for the degree of Doctor of Philosophy: Applied Mathematics Abstract This thesis presents some new results and novel techniques in the studyof thespectral density of random matrices. Spectral density is a central object of interest in random matrix theory, capturing the large-scale statistical behaviour of eigenvalues of random matrices. For certain ran- dom matrix ensembles the spectral density will converge, as the matrix dimension grows, to a well-known limit. Examples of this include Wigner’s famous semi-circle law and Girko’s elliptic law. Apart from these, and a few other cases, little else is known. Two factors in particular can complicate the analysis enormously - the intro- duction of sparsity (that is, many entries being zero) and the breaking of Hermitian symmetry. The results presented in this thesis extend the class of random matrix en- sembles for which the limiting spectral density can be computed to include various sparse and non-Hermitian ensembles. Sparse random matrices are studied through a mathematical analogy with statistical mechanics. Employing the cavity method, a technique from the field of disordered systems, a system of equations is derived which characterises the spectral density of a given sparse matrix. Analytical and numerical solutions to these equations are ex- plored, as well as the ensemble average for various sparse random matrix ensembles. For the case of non-Hermitian random matrices, a result is presented giving a method 1 for exactly calculating the spectral density of matrices under a particular type of per- turbation.
  • Random Rotation Ensembles

    Random Rotation Ensembles

    Journal of Machine Learning Research 2 (2015) 1-15 Submitted 03/15; Published ??/15 Random Rotation Ensembles Rico Blaser ([email protected]) Piotr Fryzlewicz ([email protected]) Department of Statistics London School of Economics Houghton Street London, WC2A 2AE, UK Editor: TBD Abstract In machine learning, ensemble methods combine the predictions of multiple base learners to construct more accurate aggregate predictions. Established supervised learning algo- rithms inject randomness into the construction of the individual base learners in an effort to promote diversity within the resulting ensembles. An undesirable side effect of this ap- proach is that it generally also reduces the accuracy of the base learners. In this paper, we introduce a method that is simple to implement yet general and effective in improv- ing ensemble diversity with only modest impact on the accuracy of the individual base learners. By randomly rotating the feature space prior to inducing the base learners, we achieve favorable aggregate predictions on standard data sets compared to state of the art ensemble methods, most notably for tree-based ensembles, which are particularly sensitive to rotation. Keywords: Feature Rotation, Ensemble Diversity, Smooth Decision Boundary 1. Introduction Modern statistical learning algorithms combine the predictions of multiple base learners to form ensembles, which typically achieve better aggregate predictive performance than the individual base learners (Rokach, 2010). This approach has proven to be effective in practice and some ensemble methods rank among the most accurate general-purpose supervised learning algorithms currently available. For example, a large-scale empirical study (Caruana and Niculescu-Mizil, 2006) of supervised learning algorithms found that decision tree ensembles consistently outperformed traditional single-predictor models on a representative set of binary classification tasks.
  • Moments of Random Matrices and Hypergeometric Orthogonal Polynomials

    Moments of Random Matrices and Hypergeometric Orthogonal Polynomials

    Commun. Math. Phys. 369, 1091–1145 (2019) Communications in Digital Object Identifier (DOI) https://doi.org/10.1007/s00220-019-03323-9 Mathematical Physics Moments of Random Matrices and Hypergeometric Orthogonal Polynomials Fabio Deelan Cunden1,FrancescoMezzadri2,NeilO’Connell1 , Nick Simm3 1 School of Mathematics and Statistics, University College Dublin, Belfield, Dublin 4, Ireland. E-mail: [email protected] 2 School of Mathematics, University of Bristol, University Walk, Bristol, BS8 1TW, UK 3 Mathematics Department, University of Sussex, Brighton, BN1 9RH, UK Received: 9 June 2018 / Accepted: 10 November 2018 Published online: 6 February 2019 – © The Author(s) 2019 Abstract: We establish a new connection between moments of n n random matri- × ces Xn and hypergeometric orthogonal polynomials. Specifically, we consider moments s E Tr Xn− as a function of the complex variable s C, whose analytic structure we describe completely. We discover several remarkable∈ features, including a reflection symmetry (or functional equation), zeros on a critical line in the complex plane, and orthogonality relations. An application of the theory resolves part of an integrality con- jecture of Cunden et al. (J Math Phys 57:111901, 2016)onthetime-delaymatrixof chaotic cavities. In each of the classical ensembles of random matrix theory (Gaus- sian, Laguerre, Jacobi) we characterise the moments in terms of the Askey scheme of hypergeometric orthogonal polynomials. We also calculate the leading order n asymptotics of the moments and discuss their symmetries and zeroes. We discuss aspects→∞ of these phenomena beyond the random matrix setting, including the Mellin transform of products and Wronskians of pairs of classical orthogonal polynomials.
  • Matrices and Linear Algebra

    Matrices and Linear Algebra

    Chapter 26 Matrices and linear algebra ap,mat Contents 26.1 Matrix algebra........................................... 26.2 26.1.1 Determinant (s,mat,det)................................... 26.2 26.1.2 Eigenvalues and eigenvectors (s,mat,eig).......................... 26.2 26.1.3 Hermitian and symmetric matrices............................. 26.3 26.1.4 Matrix trace (s,mat,trace).................................. 26.3 26.1.5 Inversion formulas (s,mat,mil)............................... 26.3 26.1.6 Kronecker products (s,mat,kron).............................. 26.4 26.2 Positive-definite matrices (s,mat,pd) ................................ 26.4 26.2.1 Properties of positive-definite matrices........................... 26.5 26.2.2 Partial order......................................... 26.5 26.2.3 Diagonal dominance.................................... 26.5 26.2.4 Diagonal majorizers..................................... 26.5 26.2.5 Simultaneous diagonalization................................ 26.6 26.3 Vector norms (s,mat,vnorm) .................................... 26.6 26.3.1 Examples of vector norms.................................. 26.6 26.3.2 Inequalities......................................... 26.7 26.3.3 Properties.......................................... 26.7 26.4 Inner products (s,mat,inprod) ................................... 26.7 26.4.1 Examples.......................................... 26.7 26.4.2 Properties.......................................... 26.8 26.5 Matrix norms (s,mat,mnorm) .................................... 26.8 26.5.1
  • A Logarithmic Minimization Property of the Unitary Polar Factor in The

    A Logarithmic Minimization Property of the Unitary Polar Factor in The

    A logarithmic minimization property of the unitary polar factor in the spectral norm and the Frobenius matrix norm Patrizio Neff∗ Yuji Nakatsukasa† Andreas Fischle‡ October 30, 2018 Abstract The unitary polar factor Q = Up in the polar decomposition of Z = Up H is the minimizer over unitary matrices Q for both Log(Q∗Z) 2 and its Hermitian part k k sym (Log(Q∗Z)) 2 over both R and C for any given invertible matrix Z Cn×n and k * k ∈ any matrix logarithm Log, not necessarily the principal logarithm log. We prove this for the spectral matrix norm for any n and for the Frobenius matrix norm in two and three dimensions. The result shows that the unitary polar factor is the nearest orthogonal matrix to Z not only in the normwise sense, but also in a geodesic distance. The derivation is based on Bhatia’s generalization of Bernstein’s trace inequality for the matrix exponential and a new sum of squared logarithms inequality. Our result generalizes the fact for scalars that for any complex logarithm and for all z C 0 ∈ \ { } −iϑ 2 2 −iϑ 2 2 min LogC(e z) = log z , min Re LogC(e z) = log z . ϑ∈(−π,π] | | | | || ϑ∈(−π,π] | | | | || Keywords: unitary polar factor, matrix logarithm, matrix exponential, Hermitian part, minimization, unitarily invariant norm, polar decomposition, sum of squared logarithms in- equality, optimality, matrix Lie-group, geodesic distance arXiv:1302.3235v4 [math.CA] 2 Jan 2014 AMS 2010 subject classification: 15A16, 15A18, 15A24, 15A44, 15A45, 26Dxx ∗Corresponding author, Head of Lehrstuhl f¨ur Nichtlineare Analysis und Modellierung, Fakult¨at f¨ur Mathematik, Universit¨at Duisburg-Essen, Campus Essen, Thea-Leymann Str.