The Test Matrix Toolbox for Matlab (Version 3.0)
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Parametrizations of K-Nonnegative Matrices
Parametrizations of k-Nonnegative Matrices Anna Brosowsky, Neeraja Kulkarni, Alex Mason, Joe Suk, Ewin Tang∗ October 2, 2017 Abstract Totally nonnegative (positive) matrices are matrices whose minors are all nonnegative (positive). We generalize the notion of total nonnegativity, as follows. A k-nonnegative (resp. k-positive) matrix has all minors of size k or less nonnegative (resp. positive). We give a generating set for the semigroup of k-nonnegative matrices, as well as relations for certain special cases, i.e. the k = n − 1 and k = n − 2 unitriangular cases. In the above two cases, we find that the set of k-nonnegative matrices can be partitioned into cells, analogous to the Bruhat cells of totally nonnegative matrices, based on their factorizations into generators. We will show that these cells, like the Bruhat cells, are homeomorphic to open balls, and we prove some results about the topological structure of the closure of these cells, and in fact, in the latter case, the cells form a Bruhat-like CW complex. We also give a family of minimal k-positivity tests which form sub-cluster algebras of the total positivity test cluster algebra. We describe ways to jump between these tests, and give an alternate description of some tests as double wiring diagrams. 1 Introduction A totally nonnegative (respectively totally positive) matrix is a matrix whose minors are all nonnegative (respectively positive). Total positivity and nonnegativity are well-studied phenomena and arise in areas such as planar networks, combinatorics, dynamics, statistics and probability. The study of total positivity and total nonnegativity admit many varied applications, some of which are explored in “Totally Nonnegative Matrices” by Fallat and Johnson [5]. -
Criss-Cross Type Algorithms for Computing the Real Pseudospectral Abscissa
CRISS-CROSS TYPE ALGORITHMS FOR COMPUTING THE REAL PSEUDOSPECTRAL ABSCISSA DING LU∗ AND BART VANDEREYCKEN∗ Abstract. The real "-pseudospectrum of a real matrix A consists of the eigenvalues of all real matrices that are "-close to A. The closeness is commonly measured in spectral or Frobenius norm. The real "-pseudospectral abscissa, which is the largest real part of these eigenvalues for a prescribed value ", measures the structured robust stability of A. In this paper, we introduce a criss-cross type algorithm to compute the real "-pseudospectral abscissa for the spectral norm. Our algorithm is based on a superset characterization of the real pseudospectrum where each criss and cross search involves solving linear eigenvalue problems and singular value optimization problems. The new algorithm is proved to be globally convergent, and observed to be locally linearly convergent. Moreover, we propose a subspace projection framework in which we combine the criss-cross algorithm with subspace projection techniques to solve large-scale problems. The subspace acceleration is proved to be locally superlinearly convergent. The robustness and efficiency of the proposed algorithms are demonstrated on numerical examples. Key words. eigenvalue problem, real pseudospectra, spectral abscissa, subspace methods, criss- cross methods, robust stability AMS subject classifications. 15A18, 93B35, 30E10, 65F15 1. Introduction. The real "-pseudospectrum of a matrix A 2 Rn×n is defined as R n×n (1) Λ" (A) = fλ 2 C : λ 2 Λ(A + E) with E 2 R ; kEk ≤ "g; where Λ(A) denotes the eigenvalues of a matrix A and k · k is the spectral norm. It is also known as the real unstructured spectral value set (see, e.g., [9, 13, 11]) and it can be regarded as a generalization of the more standard complex pseudospectrum (see, e.g., [23, 24]). -
QR Decomposition: History and Its Applications
Mathematics & Statistics Auburn University, Alabama, USA QR History Dec 17, 2010 Asymptotic result QR iteration QR decomposition: History and its EE Applications Home Page Title Page Tin-Yau Tam èèèUUUÎÎÎ JJ II J I Page 1 of 37 Æâ§w Go Back fÆêÆÆÆ Full Screen Close email: [email protected] Website: www.auburn.edu/∼tamtiny Quit 1. QR decomposition Recall the QR decomposition of A ∈ GLn(C): QR A = QR History Asymptotic result QR iteration where Q ∈ GLn(C) is unitary and R ∈ GLn(C) is upper ∆ with positive EE diagonal entries. Such decomposition is unique. Set Home Page a(A) := diag (r11, . , rnn) Title Page where A is written in column form JJ II J I A = (a1| · · · |an) Page 2 of 37 Go Back Geometric interpretation of a(A): Full Screen rii is the distance (w.r.t. 2-norm) between ai and span {a1, . , ai−1}, Close i = 2, . , n. Quit Example: 12 −51 4 6/7 −69/175 −58/175 14 21 −14 6 167 −68 = 3/7 158/175 6/175 0 175 −70 . QR −4 24 −41 −2/7 6/35 −33/35 0 0 35 History Asymptotic result QR iteration EE • QR decomposition is the matrix version of the Gram-Schmidt orthonor- Home Page malization process. Title Page JJ II • QR decomposition can be extended to rectangular matrices, i.e., if A ∈ J I m×n with m ≥ n (tall matrix) and full rank, then C Page 3 of 37 A = QR Go Back Full Screen where Q ∈ Cm×n has orthonormal columns and R ∈ Cn×n is upper ∆ Close with positive “diagonal” entries. -
High-Performance Parallel Computations Using Python As High-Level Language
Aachen Institute for Advanced Study in Computational Engineering Science Preprint: AICES-2010/08-01 23/August/2010 High-Performance Parallel Computations using Python as High-Level Language Stefano Masini and Paolo Bientinesi Financial support from the Deutsche Forschungsgemeinschaft (German Research Association) through grant GSC 111 is gratefully acknowledged. ©Stefano Masini and Paolo Bientinesi 2010. All rights reserved List of AICES technical reports: http://www.aices.rwth-aachen.de/preprints High-Performance Parallel Computations using Python as High-Level Language Stefano Masini⋆ and Paolo Bientinesi⋆⋆ RWTH Aachen, AICES, Aachen, Germany Abstract. High-performance and parallel computations have always rep- resented a challenge in terms of code optimization and memory usage, and have typically been tackled with languages that allow a low-level management of resources, like Fortran, C and C++. Nowadays, most of the implementation effort goes into constructing the bookkeeping logic that binds together functionalities taken from standard libraries. Because of the increasing complexity of this kind of codes, it becomes more and more necessary to keep it well organized through proper software en- gineering practices. Indeed, in the presence of chaotic implementations, reasoning about correctness is difficult, even when limited to specific aspects like concurrency; moreover, due to the lack in flexibility of the code, making substantial changes for experimentation becomes a grand challenge. Since the bookkeeping logic only accounts for a tiny fraction of the total execution time, we believe that for such a task it can be afforded to introduce an overhead due to a high-level language. We consider Python as a preliminary candidate with the intent of improving code readability, flexibility and, in turn, the level of confidence with respect to correctness. -
Stirling Matrix Via Pascal Matrix
LinearAlgebraanditsApplications329(2001)49–59 www.elsevier.com/locate/laa StirlingmatrixviaPascalmatrix Gi-SangCheon a,∗,Jin-SooKim b aDepartmentofMathematics,DaejinUniversity,Pocheon487-711,SouthKorea bDepartmentofMathematics,SungkyunkwanUniversity,Suwon440-746,SouthKorea Received11May2000;accepted17November2000 SubmittedbyR.A.Brualdi Abstract ThePascal-typematricesobtainedfromtheStirlingnumbersofthefirstkinds(n,k)and ofthesecondkindS(n,k)arestudied,respectively.Itisshownthatthesematricescanbe factorizedbythePascalmatrices.AlsotheLDU-factorizationofaVandermondematrixof theformVn(x,x+1,...,x+n−1)foranyrealnumberxisobtained.Furthermore,some well-knowncombinatorialidentitiesareobtainedfromthematrixrepresentationoftheStirling numbers,andthesematricesaregeneralizedinoneortwovariables.©2001ElsevierScience Inc.Allrightsreserved. AMSclassification:05A19;05A10 Keywords:Pascalmatrix;Stirlingnumber;Stirlingmatrix 1.Introduction Forintegersnandkwithnk0,theStirlingnumbersofthefirstkinds(n,k) andofthesecondkindS(n,k)canbedefinedasthecoefficientsinthefollowing expansionofavariablex(see[3,pp.271–279]): n n−k k [x]n = (−1) s(n,k)x k=0 and ∗ Correspondingauthor. E-mailaddresses:[email protected](G.-S.Cheon),[email protected](J.-S.Kim). 0024-3795/01/$-seefrontmatter2001ElsevierScienceInc.Allrightsreserved. PII:S0024-3795(01)00234-8 50 G.-S. Cheon, J.-S. Kim / Linear Algebra and its Applications 329 (2001) 49–59 n n x = S(n,k)[x]k, (1.1) k=0 where x(x − 1) ···(x − n + 1) if n 1, [x] = (1.2) n 1ifn = 0. It is known that for an n, k 0, the s(n,k), S(n,k) and [n]k satisfy the following Pascal-type recurrence relations: s(n,k) = s(n − 1,k− 1) + (n − 1)s(n − 1,k), S(n,k) = S(n − 1,k− 1) + kS(n − 1,k), (1.3) [n]k =[n − 1]k + k[n − 1]k−1, where s(n,0) = s(0,k)= S(n,0) = S(0,k)=[0]k = 0ands(0, 0) = S(0, 0) = 1, and moreover the S(n,k) satisfies the following formula known as ‘vertical’ recur- rence relation: n− 1 n − 1 S(n,k) = S(l,k − 1). -
Reproducing Kernel Hilbert Space Compactification of Unitary Evolution
Reproducing kernel Hilbert space compactification of unitary evolution groups Suddhasattwa Dasa, Dimitrios Giannakisa,∗, Joanna Slawinskab aCourant Institute of Mathematical Sciences, New York University, New York, NY 10012, USA bFinnish Center for Artificial Intelligence, Department of Computer Science, University of Helsinki, Helsinki, Finland Abstract A framework for coherent pattern extraction and prediction of observables of measure-preserving, ergodic dynamical systems with both atomic and continuous spectral components is developed. This framework is based on an approximation of the generator of the system by a compact operator Wτ on a reproducing kernel Hilbert space (RKHS). A key element of this approach is that Wτ is skew-adjoint (unlike regularization approaches based on the addition of diffusion), and thus can be characterized by a unique projection- valued measure, discrete by compactness, and an associated orthonormal basis of eigenfunctions. These eigenfunctions can be ordered in terms of a Dirichlet energy on the RKHS, and provide a notion of coherent observables under the dynamics akin to the Koopman eigenfunctions associated with the atomic part of the spectrum. In addition, the regularized generator has a well-defined Borel functional calculus allowing tWτ the construction of a unitary evolution group fe gt2R on the RKHS, which approximates the unitary Koopman evolution group of the original system. We establish convergence results for the spectrum and Borel functional calculus of the regularized generator to those of the original system in the limit τ ! 0+. Convergence results are also established for a data-driven formulation, where these operators are approximated using finite-rank operators obtained from observed time series. An advantage of working in spaces of observables with an RKHS structure is that one can perform pointwise evaluation and interpolation through bounded linear operators, which is not possible in Lp spaces. -
Iterative Refinement for Symmetric Eigenvalue Decomposition II
Japan Journal of Industrial and Applied Mathematics (2019) 36:435–459 https://doi.org/10.1007/s13160-019-00348-4 ORIGINAL PAPER Iterative refinement for symmetric eigenvalue decomposition II: clustered eigenvalues Takeshi Ogita1 · Kensuke Aishima2 Received: 22 October 2018 / Revised: 5 February 2019 / Published online: 22 February 2019 © The Author(s) 2019 Abstract We are concerned with accurate eigenvalue decomposition of a real symmetric matrix A. In the previous paper (Ogita and Aishima in Jpn J Ind Appl Math 35(3): 1007–1035, 2018), we proposed an efficient refinement algorithm for improving the accuracy of all eigenvectors, which converges quadratically if a sufficiently accurate initial guess is given. However, since the accuracy of eigenvectors depends on the eigenvalue gap, it is difficult to provide such an initial guess to the algorithm in the case where A has clustered eigenvalues. To overcome this problem, we propose a novel algorithm that can refine approximate eigenvectors corresponding to clustered eigenvalues on the basis of the algorithm proposed in the previous paper. Numerical results are presented showing excellent performance of the proposed algorithm in terms of convergence rate and overall computational cost and illustrating an application to a quantum materials simulation. Keywords Accurate numerical algorithm · Iterative refinement · Symmetric eigenvalue decomposition · Clustered eigenvalues Mathematics Subject Classification 65F15 · 15A18 · 15A23 This study was partially supported by CREST, JST and JSPS KAKENHI Grant numbers 16H03917, 25790096. B Takeshi Ogita [email protected] Kensuke Aishima [email protected] 1 Division of Mathematical Sciences, School of Arts and Sciences, Tokyo Woman’s Christian University, 2-6-1 Zempukuji, Suginami-ku, Tokyo 167-8585, Japan 2 Faculty of Computer and Information Sciences, Hosei University, 3-7-2 Kajino-cho, Koganei-shi, Tokyo 184-8584, Japan 123 436 T. -
Pascal Matrices Alan Edelman and Gilbert Strang Department of Mathematics, Massachusetts Institute of Technology [email protected] and [email protected]
Pascal Matrices Alan Edelman and Gilbert Strang Department of Mathematics, Massachusetts Institute of Technology [email protected] and [email protected] Every polynomial of degree n has n roots; every continuous function on [0, 1] attains its maximum; every real symmetric matrix has a complete set of orthonormal eigenvectors. “General theorems” are a big part of the mathematics we know. We can hardly resist the urge to generalize further! Remove hypotheses, make the theorem tighter and more difficult, include more functions, move into Hilbert space,. It’s in our nature. The other extreme in mathematics might be called the “particular case”. One specific function or group or matrix becomes special. It obeys the general rules, like everyone else. At the same time it has some little twist that connects familiar objects in a neat way. This paper is about an extremely particular case. The familiar object is Pascal’s triangle. The little twist begins by putting that triangle of binomial coefficients into a matrix. Three different matrices—symmetric, lower triangular, and upper triangular—can hold Pascal’s triangle in a convenient way. Truncation produces n by n matrices Sn and Ln and Un—the pattern is visible for n = 4: 1 1 1 1 1 1 1 1 1 1 2 3 4 1 1 1 2 3 S4 = L4 = U4 = . 1 3 6 10 1 2 1 1 3 1 4 10 20 1 3 3 1 1 We mention first a very specific fact: The determinant of every Sn is 1. (If we emphasized det Ln = 1 and det Un = 1, you would write to the Editor. -
Arxiv:1904.01037V3 [Math.GR]
AN EFFECTIVE LIE–KOLCHIN THEOREM FOR QUASI-UNIPOTENT MATRICES THOMAS KOBERDA, FENG LUO, AND HONGBIN SUN Abstract. We establish an effective version of the classical Lie–Kolchin Theo- rem. Namely, let A, B P GLmpCq be quasi–unipotent matrices such that the Jordan Canonical Form of B consists of a single block, and suppose that for all k ě 0 the matrix ABk is also quasi–unipotent. Then A and B have a common eigenvector. In particular, xA, Byă GLmpCq is a solvable subgroup. We give applications of this result to the representation theory of mapping class groups of orientable surfaces. 1. Introduction Let V be a finite dimensional vector space over an algebraically closed field. In this paper, we study the structure of certain subgroups of GLpVq which contain “sufficiently many” elements of a relatively simple form. We are motivated by the representation theory of the mapping class group of a surface of hyperbolic type. If S be an orientable surface of genus 2 or more, the mapping class group ModpS q is the group of homotopy classes of orientation preserving homeomor- phisms of S . The group ModpS q is generated by certain mapping classes known as Dehn twists, which are defined for essential simple closed curves of S . Here, an essential simple closed curve is a free homotopy class of embedded copies of S 1 in S which is homotopically essential, in that the homotopy class represents a nontrivial conjugacy class in π1pS q which is not the homotopy class of a boundary component or a puncture of S . -
(Hessenberg) Eigenvalue-Eigenmatrix Relations∗
(HESSENBERG) EIGENVALUE-EIGENMATRIX RELATIONS∗ JENS-PETER M. ZEMKE† Abstract. Explicit relations between eigenvalues, eigenmatrix entries and matrix elements are derived. First, a general, theoretical result based on the Taylor expansion of the adjugate of zI − A on the one hand and explicit knowledge of the Jordan decomposition on the other hand is proven. This result forms the basis for several, more practical and enlightening results tailored to non-derogatory, diagonalizable and normal matrices, respectively. Finally, inherent properties of (upper) Hessenberg, resp. tridiagonal matrix structure are utilized to construct computable relations between eigenvalues, eigenvector components, eigenvalues of principal submatrices and products of lower diagonal elements. Key words. Algebraic eigenvalue problem, eigenvalue-eigenmatrix relations, Jordan normal form, adjugate, principal submatrices, Hessenberg matrices, eigenvector components AMS subject classifications. 15A18 (primary), 15A24, 15A15, 15A57 1. Introduction. Eigenvalues and eigenvectors are defined using the relations Av = vλ and V −1AV = J. (1.1) We speak of a partial eigenvalue problem, when for a given matrix A ∈ Cn×n we seek scalar λ ∈ C and a corresponding nonzero vector v ∈ Cn. The scalar λ is called the eigenvalue and the corresponding vector v is called the eigenvector. We speak of the full or algebraic eigenvalue problem, when for a given matrix A ∈ Cn×n we seek its Jordan normal form J ∈ Cn×n and a corresponding (not necessarily unique) eigenmatrix V ∈ Cn×n. Apart from these constitutional relations, for some classes of structured matrices several more intriguing relations between components of eigenvectors, matrix entries and eigenvalues are known. For example, consider the so-called Jacobi matrices. -
Etna the Block Hessenberg Process for Matrix
Electronic Transactions on Numerical Analysis. Volume 46, pp. 460–473, 2017. ETNA Kent State University and c Copyright 2017, Kent State University. Johann Radon Institute (RICAM) ISSN 1068–9613. THE BLOCK HESSENBERG PROCESS FOR MATRIX EQUATIONS∗ M. ADDAMy, M. HEYOUNIy, AND H. SADOKz Abstract. In the present paper, we first introduce a block variant of the Hessenberg process and discuss its properties. Then, we show how to apply the block Hessenberg process in order to solve linear systems with multiple right-hand sides. More precisely, we define the block CMRH method for solving linear systems that share the same coefficient matrix. We also show how to apply this process for solving discrete Sylvester matrix equations. Finally, numerical comparisons are provided in order to compare the proposed new algorithms with other existing methods. Key words. Block Krylov subspace methods, Hessenberg process, Arnoldi process, CMRH, GMRES, low-rank matrix equations. AMS subject classifications. 65F10, 65F30 1. Introduction. In this work, we are first interested in solving s systems of linear equations with the same coefficient matrix and different right-hand sides of the form (1.1) A x(i) = y(i); 1 ≤ i ≤ s; where A is a large and sparse n × n real matrix, y(i) is a real column vector of length n, and s n. Such linear systems arise in numerous applications in computational science and engineering such as wave propagation phenomena, quantum chromodynamics, and dynamics of structures [5, 9, 36, 39]. When n is small, it is well known that the solution of (1.1) can be computed by a direct method such as LU or Cholesky factorization. -
Matrices That Are Similar to Their Inverses
116 THE MATHEMATICAL GAZETTE Matrices that are similar to their inverses GRIGORE CÃLUGÃREANU 1. Introduction In a group G, an element which is conjugate with its inverse is called real, i.e. the element and its inverse belong to the same conjugacy class. An element is called an involution if it is of order 2. With these notions it is easy to formulate the following questions. 1) Which are the (finite) groups all of whose elements are real ? 2) Which are the (finite) groups such that the identity and involutions are the only real elements ? 3) Which are the (finite) groups in which the real elements form a subgroup closed under multiplication? According to specialists, these (general) questions cannot be solved in any reasonable way. For example, there are numerous families of groups all of whose elements are real, like the symmetric groups Sn. There are many solvable groups whose elements are all real, and one can prove that any finite solvable group occurs as a subgroup of a solvable group whose elements are all real. As for question 2, note that in any Abelian group (conjugations are all the identity function), the only real elements are the identity and the involutions, and they form a subgroup. There are non-abelian examples as well, like a Suzuki 2-group. Question 3 is similar to questions 1 and 2. Therefore the abstract study of reality questions in finite groups is unlikely to have a good outcome. This may explain why in the existing bibliography there are only specific studies (see [1, 2, 3, 4]).