Matrix Depot Documentation Release 0.1.0
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Extremely Accurate Solutions Using Block Decomposition and Extended Precision for Solving Very Ill-Conditioned Equations
Extremely accurate solutions using block decomposition and extended precision for solving very ill-conditioned equations †Kansa, E.J.1, Skala V.2, and Holoborodko, P.3 1Convergent Solutions, Livermore, CA 94550 USA 2Computer Science & Engineering Dept., Faculty of Applied Sciences, University of West Bohemia, University 8, CZ 301 00 Plzen, Czech Republic 3Advanpix LLC, Maison Takashima Bldg. 2F, Daimachi 10-15-201, Yokohama, Japan 221-0834 †Corresponding author: [email protected] Abstract Many authors have complained about the ill-conditioning associated the numerical solution of partial differential equations (PDEs) and integral equations (IEs) using as the continuously differentiable Gaussian and multiquadric continuously differentiable (C ) radial basis functions (RBFs). Unlike finite elements, finite difference, or finite volume methods that lave compact local support that give rise to sparse equations, the C -RBFs with simple collocation methods give rise to full, asymmetric systems of equations. Since C RBFs have adjustable constent or variable shape parameters, the resulting systems of equations that are solve on single or double precision computers can suffer from “ill-conditioning”. Condition numbers can be either the absolute or relative condition number, but in the context of linear equations, the absolute condition number will be understood. Results will be presented that demonstrates the combination of Block Gaussian elimination, arbitrary arithmetic precision, and iterative refinement can give remarkably accurate numerical salutations to large Hilbert and van der Monde equation systems. 1. Introduction An accurate definition of the condition number, , of the matrix, A, is the ratio of the largest to smallest absolute value of the singular values, { i}, obtained from the singular value decomposition (SVD) method, see [1]: (A) = maxjjminjj (1) This definition of condition number will be used in this study. -
High-Performance Parallel Computations Using Python As High-Level Language
Aachen Institute for Advanced Study in Computational Engineering Science Preprint: AICES-2010/08-01 23/August/2010 High-Performance Parallel Computations using Python as High-Level Language Stefano Masini and Paolo Bientinesi Financial support from the Deutsche Forschungsgemeinschaft (German Research Association) through grant GSC 111 is gratefully acknowledged. ©Stefano Masini and Paolo Bientinesi 2010. All rights reserved List of AICES technical reports: http://www.aices.rwth-aachen.de/preprints High-Performance Parallel Computations using Python as High-Level Language Stefano Masini⋆ and Paolo Bientinesi⋆⋆ RWTH Aachen, AICES, Aachen, Germany Abstract. High-performance and parallel computations have always rep- resented a challenge in terms of code optimization and memory usage, and have typically been tackled with languages that allow a low-level management of resources, like Fortran, C and C++. Nowadays, most of the implementation effort goes into constructing the bookkeeping logic that binds together functionalities taken from standard libraries. Because of the increasing complexity of this kind of codes, it becomes more and more necessary to keep it well organized through proper software en- gineering practices. Indeed, in the presence of chaotic implementations, reasoning about correctness is difficult, even when limited to specific aspects like concurrency; moreover, due to the lack in flexibility of the code, making substantial changes for experimentation becomes a grand challenge. Since the bookkeeping logic only accounts for a tiny fraction of the total execution time, we believe that for such a task it can be afforded to introduce an overhead due to a high-level language. We consider Python as a preliminary candidate with the intent of improving code readability, flexibility and, in turn, the level of confidence with respect to correctness. -
Stirling Matrix Via Pascal Matrix
LinearAlgebraanditsApplications329(2001)49–59 www.elsevier.com/locate/laa StirlingmatrixviaPascalmatrix Gi-SangCheon a,∗,Jin-SooKim b aDepartmentofMathematics,DaejinUniversity,Pocheon487-711,SouthKorea bDepartmentofMathematics,SungkyunkwanUniversity,Suwon440-746,SouthKorea Received11May2000;accepted17November2000 SubmittedbyR.A.Brualdi Abstract ThePascal-typematricesobtainedfromtheStirlingnumbersofthefirstkinds(n,k)and ofthesecondkindS(n,k)arestudied,respectively.Itisshownthatthesematricescanbe factorizedbythePascalmatrices.AlsotheLDU-factorizationofaVandermondematrixof theformVn(x,x+1,...,x+n−1)foranyrealnumberxisobtained.Furthermore,some well-knowncombinatorialidentitiesareobtainedfromthematrixrepresentationoftheStirling numbers,andthesematricesaregeneralizedinoneortwovariables.©2001ElsevierScience Inc.Allrightsreserved. AMSclassification:05A19;05A10 Keywords:Pascalmatrix;Stirlingnumber;Stirlingmatrix 1.Introduction Forintegersnandkwithnk0,theStirlingnumbersofthefirstkinds(n,k) andofthesecondkindS(n,k)canbedefinedasthecoefficientsinthefollowing expansionofavariablex(see[3,pp.271–279]): n n−k k [x]n = (−1) s(n,k)x k=0 and ∗ Correspondingauthor. E-mailaddresses:[email protected](G.-S.Cheon),[email protected](J.-S.Kim). 0024-3795/01/$-seefrontmatter2001ElsevierScienceInc.Allrightsreserved. PII:S0024-3795(01)00234-8 50 G.-S. Cheon, J.-S. Kim / Linear Algebra and its Applications 329 (2001) 49–59 n n x = S(n,k)[x]k, (1.1) k=0 where x(x − 1) ···(x − n + 1) if n 1, [x] = (1.2) n 1ifn = 0. It is known that for an n, k 0, the s(n,k), S(n,k) and [n]k satisfy the following Pascal-type recurrence relations: s(n,k) = s(n − 1,k− 1) + (n − 1)s(n − 1,k), S(n,k) = S(n − 1,k− 1) + kS(n − 1,k), (1.3) [n]k =[n − 1]k + k[n − 1]k−1, where s(n,0) = s(0,k)= S(n,0) = S(0,k)=[0]k = 0ands(0, 0) = S(0, 0) = 1, and moreover the S(n,k) satisfies the following formula known as ‘vertical’ recur- rence relation: n− 1 n − 1 S(n,k) = S(l,k − 1). -
A Study of Convergence of the PMARC Matrices Applicable to WICS Calculations
A Study of Convergence of the PMARC Matrices applicable to WICS Calculations /- by Amitabha Ghosh Department of Mechanical Engineering Rochester Institute of Technology Rochester, NY 14623 Final Report NASA Cooperative Agreement No: NCC 2-937 Presented to NASA Ames Research Center Moffett Field, CA 94035 August 31, 1997 Table of Contents Abstract ............................................ 3 Introduction .......................................... 3 Solution of Linear Systems ................................... 4 Direct Solvers .......................................... 5 Gaussian Elimination: .................................. 5 Gauss-Jordan Elimination: ................................ 5 L-U Decompostion: ................................... 6 Iterative Solvers ........................................ 6 Jacobi Method: ..................................... 7 Gauss-Seidel Method: .............................. .... 7 Successive Over-relaxation Method: .......................... 8 Conjugate Gradient Method: .............................. 9 Recent Developments: ......................... ....... 9 Computational Efficiency .................................... 10 Graphical Interpretation of Residual Correction Schemes .................... 11 Defective Matrices ............................. , ......... 15 Results and Discussion ............ ......................... 16 Concluding Remarks ...................................... 19 Acknowledgements ....................................... 19 References .......................................... -
Chapter 7 Block Designs
Chapter 7 Block Designs One simile that solitary shines In the dry desert of a thousand lines. Epilogue to the Satires, ALEXANDER POPE The collection of all subsets of cardinality k of a set with v elements (k < v) has the ³ ´ v¡t property that any subset of t elements, with 0 · t · k, is contained in precisely k¡t subsets of size k. The subsets of size k provide therefore a nice covering for the subsets of a lesser cardinality. Observe that the number of subsets of size k that contain a subset of size t depends only on v; k; and t and not on the specific subset of size t in question. This is the essential defining feature of the structures that we wish to study. The example we just described inspires general interest in producing similar coverings ³ ´ v without using all the k subsets of size k but rather as small a number of them as possi- 1 2 CHAPTER 7. BLOCK DESIGNS ble. The coverings that result are often elegant geometrical configurations, of which the projective and affine planes are examples. These latter configurations form nice coverings only for the subsets of cardinality 2, that is, any two elements are in the same number of these special subsets of size k which we call blocks (or, in certain instances, lines). A collection of subsets of cardinality k, called blocks, with the property that every subset of size t (t · k) is contained in the same number (say ¸) of blocks is called a t-design. We supply the reader with constructions for t-designs with t as high as 5. -
Applied Linear Algebra
APPLIED LINEAR ALGEBRA Giorgio Picci November 24, 2015 1 Contents 1 LINEAR VECTOR SPACES AND LINEAR MAPS 10 1.1 Linear Maps and Matrices . 11 1.2 Inverse of a Linear Map . 12 1.3 Inner products and norms . 13 1.4 Inner products in coordinate spaces (1) . 14 1.5 Inner products in coordinate spaces (2) . 15 1.6 Adjoints. 16 1.7 Subspaces . 18 1.8 Image and kernel of a linear map . 19 1.9 Invariant subspaces in Rn ........................... 22 1.10 Invariant subspaces and block-diagonalization . 23 2 1.11 Eigenvalues and Eigenvectors . 24 2 SYMMETRIC MATRICES 25 2.1 Generalizations: Normal, Hermitian and Unitary matrices . 26 2.2 Change of Basis . 27 2.3 Similarity . 29 2.4 Similarity again . 30 2.5 Problems . 31 2.6 Skew-Hermitian matrices (1) . 32 2.7 Skew-Symmetric matrices (2) . 33 2.8 Square roots of positive semidefinite matrices . 36 2.9 Projections in Rn ............................... 38 2.10 Projections on general inner product spaces . 40 3 2.11 Gramians. 41 2.12 Example: Polynomial vector spaces . 42 3 LINEAR LEAST SQUARES PROBLEMS 43 3.1 Weighted Least Squares . 44 3.2 Solution by the Orthogonality Principle . 46 3.3 Matrix least-Squares Problems . 48 3.4 A problem from subspace identification . 50 3.5 Relation with Left- and Right- Inverses . 51 3.6 The Pseudoinverse . 54 3.7 The Euclidean pseudoinverse . 63 3.8 The Pseudoinverse and Orthogonal Projections . 64 3.9 Linear equations . 66 4 3.10 Unfeasible linear equations and Least Squares . 68 3.11 The Singular value decomposition (SVD) . -
Iterative Refinement for Symmetric Eigenvalue Decomposition II
Japan Journal of Industrial and Applied Mathematics (2019) 36:435–459 https://doi.org/10.1007/s13160-019-00348-4 ORIGINAL PAPER Iterative refinement for symmetric eigenvalue decomposition II: clustered eigenvalues Takeshi Ogita1 · Kensuke Aishima2 Received: 22 October 2018 / Revised: 5 February 2019 / Published online: 22 February 2019 © The Author(s) 2019 Abstract We are concerned with accurate eigenvalue decomposition of a real symmetric matrix A. In the previous paper (Ogita and Aishima in Jpn J Ind Appl Math 35(3): 1007–1035, 2018), we proposed an efficient refinement algorithm for improving the accuracy of all eigenvectors, which converges quadratically if a sufficiently accurate initial guess is given. However, since the accuracy of eigenvectors depends on the eigenvalue gap, it is difficult to provide such an initial guess to the algorithm in the case where A has clustered eigenvalues. To overcome this problem, we propose a novel algorithm that can refine approximate eigenvectors corresponding to clustered eigenvalues on the basis of the algorithm proposed in the previous paper. Numerical results are presented showing excellent performance of the proposed algorithm in terms of convergence rate and overall computational cost and illustrating an application to a quantum materials simulation. Keywords Accurate numerical algorithm · Iterative refinement · Symmetric eigenvalue decomposition · Clustered eigenvalues Mathematics Subject Classification 65F15 · 15A18 · 15A23 This study was partially supported by CREST, JST and JSPS KAKENHI Grant numbers 16H03917, 25790096. B Takeshi Ogita [email protected] Kensuke Aishima [email protected] 1 Division of Mathematical Sciences, School of Arts and Sciences, Tokyo Woman’s Christian University, 2-6-1 Zempukuji, Suginami-ku, Tokyo 167-8585, Japan 2 Faculty of Computer and Information Sciences, Hosei University, 3-7-2 Kajino-cho, Koganei-shi, Tokyo 184-8584, Japan 123 436 T. -
Pascal Matrices Alan Edelman and Gilbert Strang Department of Mathematics, Massachusetts Institute of Technology [email protected] and [email protected]
Pascal Matrices Alan Edelman and Gilbert Strang Department of Mathematics, Massachusetts Institute of Technology [email protected] and [email protected] Every polynomial of degree n has n roots; every continuous function on [0, 1] attains its maximum; every real symmetric matrix has a complete set of orthonormal eigenvectors. “General theorems” are a big part of the mathematics we know. We can hardly resist the urge to generalize further! Remove hypotheses, make the theorem tighter and more difficult, include more functions, move into Hilbert space,. It’s in our nature. The other extreme in mathematics might be called the “particular case”. One specific function or group or matrix becomes special. It obeys the general rules, like everyone else. At the same time it has some little twist that connects familiar objects in a neat way. This paper is about an extremely particular case. The familiar object is Pascal’s triangle. The little twist begins by putting that triangle of binomial coefficients into a matrix. Three different matrices—symmetric, lower triangular, and upper triangular—can hold Pascal’s triangle in a convenient way. Truncation produces n by n matrices Sn and Ln and Un—the pattern is visible for n = 4: 1 1 1 1 1 1 1 1 1 1 2 3 4 1 1 1 2 3 S4 = L4 = U4 = . 1 3 6 10 1 2 1 1 3 1 4 10 20 1 3 3 1 1 We mention first a very specific fact: The determinant of every Sn is 1. (If we emphasized det Ln = 1 and det Un = 1, you would write to the Editor. -
Arxiv:1904.01037V3 [Math.GR]
AN EFFECTIVE LIE–KOLCHIN THEOREM FOR QUASI-UNIPOTENT MATRICES THOMAS KOBERDA, FENG LUO, AND HONGBIN SUN Abstract. We establish an effective version of the classical Lie–Kolchin Theo- rem. Namely, let A, B P GLmpCq be quasi–unipotent matrices such that the Jordan Canonical Form of B consists of a single block, and suppose that for all k ě 0 the matrix ABk is also quasi–unipotent. Then A and B have a common eigenvector. In particular, xA, Byă GLmpCq is a solvable subgroup. We give applications of this result to the representation theory of mapping class groups of orientable surfaces. 1. Introduction Let V be a finite dimensional vector space over an algebraically closed field. In this paper, we study the structure of certain subgroups of GLpVq which contain “sufficiently many” elements of a relatively simple form. We are motivated by the representation theory of the mapping class group of a surface of hyperbolic type. If S be an orientable surface of genus 2 or more, the mapping class group ModpS q is the group of homotopy classes of orientation preserving homeomor- phisms of S . The group ModpS q is generated by certain mapping classes known as Dehn twists, which are defined for essential simple closed curves of S . Here, an essential simple closed curve is a free homotopy class of embedded copies of S 1 in S which is homotopically essential, in that the homotopy class represents a nontrivial conjugacy class in π1pS q which is not the homotopy class of a boundary component or a puncture of S . -
Matrices That Are Similar to Their Inverses
116 THE MATHEMATICAL GAZETTE Matrices that are similar to their inverses GRIGORE CÃLUGÃREANU 1. Introduction In a group G, an element which is conjugate with its inverse is called real, i.e. the element and its inverse belong to the same conjugacy class. An element is called an involution if it is of order 2. With these notions it is easy to formulate the following questions. 1) Which are the (finite) groups all of whose elements are real ? 2) Which are the (finite) groups such that the identity and involutions are the only real elements ? 3) Which are the (finite) groups in which the real elements form a subgroup closed under multiplication? According to specialists, these (general) questions cannot be solved in any reasonable way. For example, there are numerous families of groups all of whose elements are real, like the symmetric groups Sn. There are many solvable groups whose elements are all real, and one can prove that any finite solvable group occurs as a subgroup of a solvable group whose elements are all real. As for question 2, note that in any Abelian group (conjugations are all the identity function), the only real elements are the identity and the involutions, and they form a subgroup. There are non-abelian examples as well, like a Suzuki 2-group. Question 3 is similar to questions 1 and 2. Therefore the abstract study of reality questions in finite groups is unlikely to have a good outcome. This may explain why in the existing bibliography there are only specific studies (see [1, 2, 3, 4]). -
Centro-Invertible Matrices Linear Algebra and Its Applications, 434 (2011) Pp144-151
References • R.S. Wikramaratna, The centro-invertible matrix:a new type of matrix arising in pseudo-random number generation, Centro-invertible Matrices Linear Algebra and its Applications, 434 (2011) pp144-151. [doi:10.1016/j.laa.2010.08.011]. Roy S Wikramaratna, RPS Energy [email protected] • R.S. Wikramaratna, Theoretical and empirical convergence results for additive congruential random number generators, Reading University (Conference in honour of J. Comput. Appl. Math., 233 (2010) 2302-2311. Nancy Nichols' 70th birthday ) [doi: 10.1016/j.cam.2009.10.015]. 2-3 July 2012 Career Background Some definitions … • Worked at Institute of Hydrology, 1977-1984 • I is the k by k identity matrix – Groundwater modelling research and consultancy • J is the k by k matrix with ones on anti-diagonal and zeroes – P/t MSc at Reading 1980-82 (Numerical Solution of PDEs) elsewhere • Worked at Winfrith, Dorset since 1984 – Pre-multiplication by J turns a matrix ‘upside down’, reversing order of terms in each column – UKAEA (1984 – 1995), AEA Technology (1995 – 2002), ECL Technology (2002 – 2005) and RPS Energy (2005 onwards) – Post-multiplication by J reverses order of terms in each row – Oil reservoir engineering, porous medium flow simulation and 0 0 0 1 simulator development 0 0 1 0 – Consultancy to Oil Industry and to Government J = = ()j 0 1 0 0 pq • Personal research interests in development and application of numerical methods to solve engineering 1 0 0 0 j =1 if p + q = k +1 problems, and in mathematical and numerical analysis -
Numeric-Symbolic Exact Rational Linear System Solver∗
Numeric-Symbolic Exact Rational Linear System Solver∗ B. David Saunders, David Harlan Wood, and Bryan S. Youse Dept. of Computer and Information Sciences, University of Delaware Newark, Delaware, USA [email protected], [email protected], [email protected] ABSTRACT 1. INTRODUCTION An iterative refinement approach is taken to rational linear We address the problem of solving Ax = b for a vector x 2 n m×n m system solving. Such methods produce, for each entry of the Q ; given A 2 Q and b 2 Q : We will restrict ourselves solution vector, a rational approximation with denominator to square (m = n), nonsingular matrices with integer entries a power of 2. From this the correct rational entry can be of length d bits or fewer. This is the core problem. Also, in reconstructed. Our iteration is a numeric-symbolic hybrid this paper we are concerned with dense matrices, which is in that it uses an approximate numeric solver at each step to say, matrices that do not have so many zero entries that together with a symbolic (exact arithmetic) residual com- more specialized sparse matrix techniques should be applied. putation and symbolic rational reconstruction. The rational We do anticipate that the refined numeric-symbolic iterative solution may be checked symbolically (exactly). However, approach presented here will also apply effectively to sparse there is some possibility of failure of convergence, usually systems. due to numeric ill-conditioning. Alternatively, the algorithm We present a method which is a refinement of the numeric- may be used to obtain an extended precision floating point symbolic method of Wan[27, 26].