Extreme Points of the Vandermonde Determinant in Numerical
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
1111: Linear Algebra I
1111: Linear Algebra I Dr. Vladimir Dotsenko (Vlad) Lecture 11 Dr. Vladimir Dotsenko (Vlad) 1111: Linear Algebra I Lecture 11 1 / 13 Previously on. Theorem. Let A be an n × n-matrix, and b a vector with n entries. The following statements are equivalent: (a) the homogeneous system Ax = 0 has only the trivial solution x = 0; (b) the reduced row echelon form of A is In; (c) det(A) 6= 0; (d) the matrix A is invertible; (e) the system Ax = b has exactly one solution. A very important consequence (finite dimensional Fredholm alternative): For an n × n-matrix A, the system Ax = b either has exactly one solution for every b, or has infinitely many solutions for some choices of b and no solutions for some other choices. In particular, to prove that Ax = b has solutions for every b, it is enough to prove that Ax = 0 has only the trivial solution. Dr. Vladimir Dotsenko (Vlad) 1111: Linear Algebra I Lecture 11 2 / 13 An example for the Fredholm alternative Let us consider the following question: Given some numbers in the first row, the last row, the first column, and the last column of an n × n-matrix, is it possible to fill the numbers in all the remaining slots in a way that each of them is the average of its 4 neighbours? This is the \discrete Dirichlet problem", a finite grid approximation to many foundational questions of mathematical physics. Dr. Vladimir Dotsenko (Vlad) 1111: Linear Algebra I Lecture 11 3 / 13 An example for the Fredholm alternative For instance, for n = 4 we may face the following problem: find a; b; c; d to put in the matrix 0 4 3 0 1:51 B 1 a b -1C B C @0:5 c d 2 A 2:1 4 2 1 so that 1 a = 4 (3 + 1 + b + c); 1 8b = 4 (a + 0 - 1 + d); >c = 1 (a + 0:5 + d + 4); > 4 < 1 d = 4 (b + c + 2 + 2): > > This is a system with 4:> equations and 4 unknowns. -
Extremely Accurate Solutions Using Block Decomposition and Extended Precision for Solving Very Ill-Conditioned Equations
Extremely accurate solutions using block decomposition and extended precision for solving very ill-conditioned equations †Kansa, E.J.1, Skala V.2, and Holoborodko, P.3 1Convergent Solutions, Livermore, CA 94550 USA 2Computer Science & Engineering Dept., Faculty of Applied Sciences, University of West Bohemia, University 8, CZ 301 00 Plzen, Czech Republic 3Advanpix LLC, Maison Takashima Bldg. 2F, Daimachi 10-15-201, Yokohama, Japan 221-0834 †Corresponding author: [email protected] Abstract Many authors have complained about the ill-conditioning associated the numerical solution of partial differential equations (PDEs) and integral equations (IEs) using as the continuously differentiable Gaussian and multiquadric continuously differentiable (C ) radial basis functions (RBFs). Unlike finite elements, finite difference, or finite volume methods that lave compact local support that give rise to sparse equations, the C -RBFs with simple collocation methods give rise to full, asymmetric systems of equations. Since C RBFs have adjustable constent or variable shape parameters, the resulting systems of equations that are solve on single or double precision computers can suffer from “ill-conditioning”. Condition numbers can be either the absolute or relative condition number, but in the context of linear equations, the absolute condition number will be understood. Results will be presented that demonstrates the combination of Block Gaussian elimination, arbitrary arithmetic precision, and iterative refinement can give remarkably accurate numerical salutations to large Hilbert and van der Monde equation systems. 1. Introduction An accurate definition of the condition number, , of the matrix, A, is the ratio of the largest to smallest absolute value of the singular values, { i}, obtained from the singular value decomposition (SVD) method, see [1]: (A) = maxjjminjj (1) This definition of condition number will be used in this study. -
Polynomials and Hankel Matrices
View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Elsevier - Publisher Connector Polynomials and Hankel Matrices Miroslav Fiedler Czechoslovak Academy of Sciences Institute of Mathematics iitnci 25 115 67 Praha 1, Czechoslovakia Submitted by V. Ptak ABSTRACT Compatibility of a Hankel n X n matrix W and a polynomial f of degree m, m < n, is defined. If m = n, compatibility means that HC’ = CfH where Cf is the companion matrix of f With a suitable generalization of Cr, this theorem is gener- alized to the case that m < n. INTRODUCTION By a Hankel matrix [S] we shall mean a square complex matrix which has, if of order n, the form ( ai +k), i, k = 0,. , n - 1. If H = (~y~+~) is a singular n X n Hankel matrix, the H-polynomial (Pi of H was defined 131 as the greatest common divisor of the determinants of all (r + 1) x (r + 1) submatrices~of the matrix where r is the rank of H. In other words, (Pi is that polynomial for which the nX(n+l)matrix I, 0 0 0 %fb) 0 i 0 0 0 1 LINEAR ALGEBRA AND ITS APPLICATIONS 66:235-248(1985) 235 ‘F’Elsevier Science Publishing Co., Inc., 1985 52 Vanderbilt Ave., New York, NY 10017 0024.3795/85/$3.30 236 MIROSLAV FIEDLER is the Smith normal form [6] of H,. It has also been shown [3] that qN is a (nonzero) polynomial of degree at most T. It is known [4] that to a nonsingular n X n Hankel matrix H = ((Y~+~)a linear pencil of polynomials of degree at most n can be assigned as follows: f(x) = fo + f,x + . -
Inverse Problems for Hankel and Toeplitz Matrices
View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Elsevier - Publisher Connector Inverse Problems for Hankel and Toeplitz Matrices Georg Heinig Vniversitiit Leipzig Sektion Mathematik Leipzig 7010, Germany Submitted by Peter Lancaster ABSTRACT The problem is considered how to construct a Toeplitz or Hankel matrix A from one or two equations of the form Au = g. The general solution is described explicitly. Special cases are inverse spectral problems for Hankel and Toeplitz matrices and problems from the theory of orthogonal polynomials. INTRODUCTION When we speak of inverse problems we have in mind the following type of problems: Given vectors uj E C”, gj E C”’ (j = l,.. .,T), we ask for an m x n matrix A of a certain matrix class such that Auj = gj (j=l,...,?-). (0.1) In the present paper we deal with inverse problems with the additional condition that A is a Hankel matrix [ si +j] or a Toeplitz matrix [ ci _j]. Inverse problems for Hankel and Toeplitz matices occur, for example, in the theory of orthogonal polynomials when a measure p on the real line or the unit circle is wanted such that given polynomials are orthogonal with respect to this measure. The moment matrix of p is just the solution of a certain inverse problem and is Hankel (in the real line case) or Toeplitz (in the unit circle case); here the gj are unit vectors. LINEAR ALGEBRA AND ITS APPLICATIONS 165:1-23 (1992) 1 0 Elsevier Science Publishing Co., Inc., 1992 655 Avenue of tbe Americas, New York, NY 10010 0024-3795/92/$5.00 2 GEORG HEINIG Inverse problems for Toeplitz matrices were considered for the first time in the paper [lo]of M. -
A Study of Convergence of the PMARC Matrices Applicable to WICS Calculations
A Study of Convergence of the PMARC Matrices applicable to WICS Calculations /- by Amitabha Ghosh Department of Mechanical Engineering Rochester Institute of Technology Rochester, NY 14623 Final Report NASA Cooperative Agreement No: NCC 2-937 Presented to NASA Ames Research Center Moffett Field, CA 94035 August 31, 1997 Table of Contents Abstract ............................................ 3 Introduction .......................................... 3 Solution of Linear Systems ................................... 4 Direct Solvers .......................................... 5 Gaussian Elimination: .................................. 5 Gauss-Jordan Elimination: ................................ 5 L-U Decompostion: ................................... 6 Iterative Solvers ........................................ 6 Jacobi Method: ..................................... 7 Gauss-Seidel Method: .............................. .... 7 Successive Over-relaxation Method: .......................... 8 Conjugate Gradient Method: .............................. 9 Recent Developments: ......................... ....... 9 Computational Efficiency .................................... 10 Graphical Interpretation of Residual Correction Schemes .................... 11 Defective Matrices ............................. , ......... 15 Results and Discussion ............ ......................... 16 Concluding Remarks ...................................... 19 Acknowledgements ....................................... 19 References .......................................... -
Paper We Shall Assume That All Logs Are Base 2
Maximally Recoverable Codes: the Bounded Case Venkata Gandikota ∗ Elena Grigorescu y Clayton Thomas z Minshen Zhu x Abstract Modern distributed storage systems employ Maximally Recoverable codes that aim to bal- ance failure recovery capabilities with encoding/decoding efficiency tradeoffs. Recent works of Gopalan et al [SODA 2017] and Kane et al [FOCS 2017] show that the alphabet size of grid-like topologies of practical interest must be large, a feature that hampers decoding efficiency. To bypass such shortcomings, in this work we initiate the study of a weaker version of recoverability, where instead of being able to correct all correctable erasure patterns (as is the case for maximal recoverability), we only require to correct all erasure patterns of bounded size. The study of this notion reduces to a variant of a combinatorial problem studied in the literature, which is interesting in its own right. We study the alphabet size of codes withstanding all erasure patterns of small (constant) size. We believe the questions we propose are relevant to both real storage systems and combinatorial analysis, and merit further study. 1 Introduction Modern distributed storage systems need to address the challenge of storing large amounts of data, with small overhead, while providing reliable recovery in the face of failures. Unlike in com- munication settings, if a failure occurs in a storage system, it is typically easy to detect where it occurs (e.g., rack or data center failure are obvious to the system). Hence, recent trends in practical systems adopt erasure coding schemes with fast encoding and decoding capabilities [BHH13, HSX+12, SLR+14]. -
Applied Linear Algebra
APPLIED LINEAR ALGEBRA Giorgio Picci November 24, 2015 1 Contents 1 LINEAR VECTOR SPACES AND LINEAR MAPS 10 1.1 Linear Maps and Matrices . 11 1.2 Inverse of a Linear Map . 12 1.3 Inner products and norms . 13 1.4 Inner products in coordinate spaces (1) . 14 1.5 Inner products in coordinate spaces (2) . 15 1.6 Adjoints. 16 1.7 Subspaces . 18 1.8 Image and kernel of a linear map . 19 1.9 Invariant subspaces in Rn ........................... 22 1.10 Invariant subspaces and block-diagonalization . 23 2 1.11 Eigenvalues and Eigenvectors . 24 2 SYMMETRIC MATRICES 25 2.1 Generalizations: Normal, Hermitian and Unitary matrices . 26 2.2 Change of Basis . 27 2.3 Similarity . 29 2.4 Similarity again . 30 2.5 Problems . 31 2.6 Skew-Hermitian matrices (1) . 32 2.7 Skew-Symmetric matrices (2) . 33 2.8 Square roots of positive semidefinite matrices . 36 2.9 Projections in Rn ............................... 38 2.10 Projections on general inner product spaces . 40 3 2.11 Gramians. 41 2.12 Example: Polynomial vector spaces . 42 3 LINEAR LEAST SQUARES PROBLEMS 43 3.1 Weighted Least Squares . 44 3.2 Solution by the Orthogonality Principle . 46 3.3 Matrix least-Squares Problems . 48 3.4 A problem from subspace identification . 50 3.5 Relation with Left- and Right- Inverses . 51 3.6 The Pseudoinverse . 54 3.7 The Euclidean pseudoinverse . 63 3.8 The Pseudoinverse and Orthogonal Projections . 64 3.9 Linear equations . 66 4 3.10 Unfeasible linear equations and Least Squares . 68 3.11 The Singular value decomposition (SVD) . -
Seminar VII for the Course GROUP THEORY in PHYSICS Micael Flohr
Seminar VII for the course GROUP THEORY IN PHYSICS Mic~ael Flohr The classical Lie groups 25. January 2005 MATRIX LIE GROUPS Most Lie groups one ever encouters in physics are realized as matrix Lie groups and thus as subgroups of GL(n, R) or GL(n, C). This is the group of invertibel n × n matrices with coefficients in R or C, respectively. This is a Lie group, since it forms as an open subset of the vector space of n × n matrices a manifold. Matrix multiplication is certainly a differentiable map, as is taking the inverse via Cramer’s rule. The only condition defining the open 2 subset is that the determinat must not be zero, which implies that dimKGL(n, K) = n is the same as the one of the vector space Mn(K). However, GL(n, R) is not connected, because we cannot move continuously from a matrix with determinant less than zero to one with determinant larger than zero. It is worth mentioning that gl(n, K) is the vector space of all n × n matrices over the field K, equipped with the standard commutator as Lie bracket. We can describe most other Lie groups as subgroups of GL(n, K) for either K = R or K = C. There are two ways to do so. Firstly, one can give restricting equations to the coefficients of the matrices. Secondly, one can find subgroups of the automorphisms of V =∼ Kn, which conserve a given structure on Kn. In the following, we give some examples for this: SL(n, K). -
Inertia of the Matrix [(Pi + Pj) ]
isid/ms/2013/12 October 20, 2013 http://www.isid.ac.in/estatmath/eprints r Inertia of the matrix [(pi + pj) ] Rajendra Bhatia and Tanvi Jain Indian Statistical Institute, Delhi Centre 7, SJSS Marg, New Delhi{110 016, India r INERTIA OF THE MATRIX [(pi + pj) ] RAJENDRA BHATIA* AND TANVI JAIN** Abstract. Let p1; : : : ; pn be positive real numbers. It is well r known that for every r < 0 the matrix [(pi + pj) ] is positive def- inite. Our main theorem gives a count of the number of positive and negative eigenvalues of this matrix when r > 0: Connections with some other matrices that arise in Loewner's theory of oper- ator monotone functions and in the theory of spline interpolation are discussed. 1. Introduction Let p1; p2; : : : ; pn be distinct positive real numbers. The n×n matrix 1 C = [ ] is known as the Cauchy matrix. The special case pi = i pi+pj 1 gives the Hilbert matrix H = [ i+j ]: Both matrices have been studied by several authors in diverse contexts and are much used as test matrices in numerical analysis. The Cauchy matrix is known to be positive definite. It possessesh ai ◦r 1 stronger property: for each r > 0 the entrywise power C = r (pi+pj ) is positive definite. (See [4] for a proof.) The object of this paper is to study positivity properties of the related family of matrices r Pr = [(pi + pj) ]; r ≥ 0: (1) The inertia of a Hermitian matrix A is the triple In(A) = (π(A); ζ(A); ν(A)) ; in which π(A); ζ(A) and ν(A) stand for the number of positive, zero, and negative eigenvalues of A; respectively. -
Mathematicians Fleeing from Nazi Germany
Mathematicians Fleeing from Nazi Germany Mathematicians Fleeing from Nazi Germany Individual Fates and Global Impact Reinhard Siegmund-Schultze princeton university press princeton and oxford Copyright 2009 © by Princeton University Press Published by Princeton University Press, 41 William Street, Princeton, New Jersey 08540 In the United Kingdom: Princeton University Press, 6 Oxford Street, Woodstock, Oxfordshire OX20 1TW All Rights Reserved Library of Congress Cataloging-in-Publication Data Siegmund-Schultze, R. (Reinhard) Mathematicians fleeing from Nazi Germany: individual fates and global impact / Reinhard Siegmund-Schultze. p. cm. Includes bibliographical references and index. ISBN 978-0-691-12593-0 (cloth) — ISBN 978-0-691-14041-4 (pbk.) 1. Mathematicians—Germany—History—20th century. 2. Mathematicians— United States—History—20th century. 3. Mathematicians—Germany—Biography. 4. Mathematicians—United States—Biography. 5. World War, 1939–1945— Refuges—Germany. 6. Germany—Emigration and immigration—History—1933–1945. 7. Germans—United States—History—20th century. 8. Immigrants—United States—History—20th century. 9. Mathematics—Germany—History—20th century. 10. Mathematics—United States—History—20th century. I. Title. QA27.G4S53 2008 510.09'04—dc22 2008048855 British Library Cataloging-in-Publication Data is available This book has been composed in Sabon Printed on acid-free paper. ∞ press.princeton.edu Printed in the United States of America 10 987654321 Contents List of Figures and Tables xiii Preface xvii Chapter 1 The Terms “German-Speaking Mathematician,” “Forced,” and“Voluntary Emigration” 1 Chapter 2 The Notion of “Mathematician” Plus Quantitative Figures on Persecution 13 Chapter 3 Early Emigration 30 3.1. The Push-Factor 32 3.2. The Pull-Factor 36 3.D. -
Math 126 Lecture 4. Basic Facts in Representation Theory
Math 126 Lecture 4. Basic facts in representation theory. Notice. Definition of a representation of a group. The theory of group representations is the creation of Frobenius: Georg Frobenius lived from 1849 to 1917 Frobenius combined results from the theory of algebraic equations, geometry, and number theory, which led him to the study of abstract groups, the representation theory of groups and the character theory of groups. Find out more at: http://www-history.mcs.st-andrews.ac.uk/history/ Mathematicians/Frobenius.html Matrix form of a representation. Equivalence of two representations. Invariant subspaces. Irreducible representations. One dimensional representations. Representations of cyclic groups. Direct sums. Tensor product. Unitary representations. Averaging over the group. Maschke’s theorem. Heinrich Maschke 1853 - 1908 Schur’s lemma. Issai Schur Biography of Schur. Issai Schur Born: 10 Jan 1875 in Mogilyov, Mogilyov province, Russian Empire (now Belarus) Died: 10 Jan 1941 in Tel Aviv, Palestine (now Israel) Although Issai Schur was born in Mogilyov on the Dnieper, he spoke German without a trace of an accent, and nobody even guessed that it was not his first language. He went to Latvia at the age of 13 and there he attended the Gymnasium in Libau, now called Liepaja. In 1894 Schur entered the University of Berlin to read mathematics and physics. Frobenius was one of his teachers and he was to greatly influence Schur and later to direct his doctoral studies. Frobenius and Burnside had been the two main founders of the theory of representations of groups as groups of matrices. This theory proved a very powerful tool in the study of groups and Schur was to learn the foundations of this subject from Frobenius. -
Hermitian, Symmetric and Symplectic Random Ensembles
Annals of Mathematics, 153 (2001), 149–189 Hermitian, symmetric and symplectic random ensembles: PDEs for the distribution of the spectrum By M. Adler and P. van Moerbeke* Abstract Given the Hermitian, symmetric and symplectic ensembles, it is shown that the probability that the spectrum belongs to one or several intervals sat- isfies a nonlinear PDE. This is done for the three classical ensembles: Gaussian, Laguerre and Jacobi. For the Hermitian ensemble, the PDE (in the boundary points of the intervals) is related to the Toda lattice and the KP equation, whereas for the symmetric and symplectic ensembles the PDE is an inductive equation, related to the so-called Pfaff-KP equation and the Pfaff lattice. The method consists of inserting time-variables in the integral and showing that this integral satisfies integrable lattice equations and Virasoro constraints. Contents 0. Introduction 0.1. Hermitian, symmetric and symplectic Gaussian ensembles 0.2. Hermitian, symmetric and symplectic Laguerre ensembles 0.3. Hermitian, symmetric and symplectic Jacobi ensembles 0.4. ODEs, when E has one boundary point 1. Beta-integrals 1.1. Virasoro constraints for β-integrals 1.2. Proof: β-integrals as fixed points of vertex operators 1.3. Examples arXiv:math-ph/0009001v2 14 Aug 2001 2. Matrix integrals and associated integrable systems 2.1. Hermitian matrix integrals and the Toda lattice 2.2. Symmetric/symplectic matrix integrals and the Pfaff lattice 3. Expressing t-partials in terms of boundary-partials 3.1. Gaussian and Laguerre ensembles 3.2. Jacobi ensemble 3.3. Evaluating the matrix integrals on the full range ∗The support of a National Science Foundation grant #DMS-98-4-50790 is gratefully acknowl- edged.