Chapter 3. Basic Properties of Matrices

Total Page:16

File Type:pdf, Size:1020Kb

Chapter 3. Basic Properties of Matrices 3.1. Basic Definitions and Notations 1 Chapter 3. Basic Properties of Matrices Note. This long chapter (over 100 pages) contains the bulk of the material for this course. As in Chapter 2, unless stated otherwise, all matrix entries are real numbers. Chapters 8 and 9 include material on applications of this material to regression and statistics. Section 3.1. Basic Definitions and Notation Note. In this section, we reintroduce many of the definitions you should be famil- iar with from sophomore Linear Algebra (diagonal, trace, minor matrix, cofactor, determinant). We also introduce some notation and matrix/vector manipulations which might be new to you. Definition. Let A be an n×m matrix. Treating the columns of A as n-vectors, the vector space generated by these column vectors (that is, the span of the columns of A) is the column space of matrix A. This is sometimes called the range or manifold of A and is denoted span(A). Treating the rows of A as m-vectors, the vectors generated by these row vectors (that is, the span of the rows of A) is the row space of matrix A. 3.1. Basic Definitions and Notations 2 Definition. For matrix A = [aij], define scalar multiplication entry-wise as cA = c[aij] = [caij]. For A = [aij] and B = [bij] m × n matrices, define matrix addition entry-wise as A + B = [aij] + [bij] = [aij + bij]. The diagonal elements of A = [aij] are the elements aii for i = 1, 2,..., min{m, n} (these elements make up the principal diagonal), elements aij where i < j are above the diagonal, and elements aij where i > j are below the diagonal. The elements ai,i+k (for fixed k) form the kth codiagonal or kth minor diagonal. For n × m matrix A, the elements ai,m+1−i for i = 1, 2,..., min{m, n} are the skew diagonal elements of A. Definition. Square matrix A of size n × n is symmetric if aij = aji for all 1 ≤ i, j ≤ n. Square matrix A is skew symmetric if aij = −aji for 1 ≤ i, j ≤ n. Square matrix A with complex entries is Hermetian (or self-adjoint) if aij = aji (where aji represents the complex conjugate of aji). Definition. If all entries of a square matrix are 0 except for the principal diagonal elements, then matrix A is a diagonal matrix. If all the principal diagonal elements are 0, the matrix is a hollow matrix. If all entries except the principal skew diagonal elements of a matrix are 0, the matrix is a skew diagonal matrix. Note. A skew symmetric matrix must satisfy aii = −aii for 1 ≤ i ≤ n and so is necessarily hollow. 3.1. Basic Definitions and Notations 3 Pm Definition. An n × m matrix A for which |aii| > j=1,j6=i |aij| for each i = 1, 2, . , n is row diagonally dominant (that is, the absolute value of the diagonal entry is larger than the sum of the absolute values of the other entries in the row Pn containing that diagonal entry [for each row]). Similarly, if |ajj| > i=1,i6=j |aij| for each j = 1, 2, . , m then A is column diagonally dominant. Note. If A is symmetric (and hence square) then row and column diagonal domi- nance are equivalent in which case A is simply called diagonally dominant. Definition. A square matrix such that all elements below the diagonal are 0 is an upper triangular matrix; if all entries above the diagonal are 0 the matrix is a lower triangular matrix. If a nonsquare matrix has all entries below the diagonal {aii} equal to 0 (or all entries above the diagonal equal to 0) then the matrix is trapezoidal. Definition. A square matrix A for which all elements are 0 except for ai,i+ck for some (“small”) ck where ck ∈ {−w`, −w` + 1,..., −1, 0, 1, . , wu−1, wu} is a band matrix (or banded matrix). Value w` is the lower band width and wu is the upper band width. A band matrix with lower and upper band widths of 1 is tridiagonal. A matrix is in upper Hessenburg form if it is upper triangular except for the first (lower) subdiagonal. Note. Gentle states that band matrices arise in time series, stochastic processes, and differential equations (see page 43). 3.1. Basic Definitions and Notations 4 Definition. For m × n matrix A = [aij], the transpose matrix is the n × m matrix T A = [aji]. If the entries of A are complex numbers, then the adjoint of A is the H conjugate transpose A = [aji], also called the Hermetian of A. Note. As you’ve seen in sophomore Linear Algebra, we often put vectors into ma- trices and extract vectors from matrices. We now define such operations formally. Definition. Define the “constructor function” diag : Rn → Rn×n as d1 0 0 ··· 0 0 d2 0 ··· 0 diag([d1, d2, . , dn]) = 0 0 d ··· 0 3 . . ... 0 0 0 ··· dn which maps an n-vector into an n × n diagonal matrix with diagonal entries equal to the entries in the n-vector. Define vecdiag : Rn×m → Rk where k = min{m, n} as vecdiag(A) = [a11, a22, . , akk]. Define vec : Rn×m → Rnm as T T T vec(A) = [a1 , a2 , . , am] T T T where a1, a2, . , am are the columns of matrix A (so “[a1 , a2 , . , am]” denotes an nm row vector). Define vech : Rm×n → Rm(m+1)/2 as vech(A) = [a11, a21, . , am1, a22, a32, . , am2, a33, a43, . , amm] (Gentle restricts vech to the symmetric matrices in Rm×n so that vech(A) produces a vector with entries as those entries on and above the diagonal of A). 3.1. Basic Definitions and Notations 5 Note. Gentle doesn’t give a formal definition of a partitioned matrix. You are familiar with the idea of a partitioned matrix since the system of equations A~x = ~b (in the notation of sophomore Linear Algebra) is dealt with in terms of the partitioned matrix (or “augmented matrix”) [A |~b]. “Definition.” An n×m matrix A can be partitioned into four matrices A11,A12,A21, and A as 22 A11 A12 A = A21 A22 where A11 and A12 have the same number of rows (say r1); A21 and A22 have the same number of rows (say r2); A11 and A21 have the same number of columns (say c1); and A12 and A22 have the same number of columns (say c2). The ith row of A11 “combined with” the ith row of A12 gives the ith row of A for 1 ≤ i ≤ r1; the ith row of A21 “combined with” the ith row of A22 gives the (r1 + i)th row of A for 1 ≤ i ≤ r2; the jth column of A11 “combined with” the jth column of A21 gives the jth column of A for 1 ≤ j ≤ c1; and the jth column of A12 “combined with” the jth column of A22 gives the (c1 + j)th column of A for 1 ≤ j ≤ c2. Definition. Matrix B is a submatrix of matrix A if B is obtained from A by deleting some of the rows and some of the columns of A. A square submatrix whose principal diagonal elements are elements of the principal diagonal of matrix A is a principal submatrix of A. A principal submatrix B obtained from n × m matrix A such that if column j is eliminated from A then all columns j + 1, j + 2, . m are eliminated in producing B, and if row i is eliminated from A then all rows i + 1, i + 2, . , n are eliminated, is a leading principal submatrix. 3.1. Basic Definitions and Notations 6 Theorem 3.1.1. Suppose matrix A is diagonally dominant (that is, A is symmetric and row and column diagonally dominant). If B is a principal submatrix of A then B is also diagonally dominant. Definition. A partitioned matrix is block diagonal if it is of the form A1 0 ··· 0 0 A2 ··· 0 . .. . 0 0 ··· Ak for i = 1, 2, . , k and where “0” represents a submatrix of all 0’s of appropriate dimensions to yield a partitioned matrix. Similar to the function diag : Rn → Rn×n, we denote the matrix above as diag(A1,A2,...,Ak). Definition. Let A = [aij] and B = [bij] be n × m matrices. The sum A + B is the n × m matrix C = [cij] where cij = aij + bij. For r ∈ R a scalar, rA is the matrix D = [dij] where dij = raij. Theorem 3.1.2. Let A and B be n × m matrices and r, s ∈ R. Then (a) A + B = B + A (Commutivity Law of Matrix Addition). (b) (A + B) + C = A + (B + C) (Associative Law of Matrix Addition). (c) A + 0 = 0 + A where 0 is the n × m zero matrix (Matrix Additive Identity). 3.1. Basic Definitions and Notations 7 (d) r(A + B) = rA + rB (Distribution of Scalar Multiplication over Matrix Addi- tion). (e) (r+s)A = rA+sA (Distribution of Scalar Multiplication over Scalar Addition). (f) (rs)A = r(sA) (Associativity Law of Scalar Multiplication). Note. Since we have defined matrix addition and scalar multiplication, we can show that Rn×m is in fact a vector space (of dimension nm). A subspace of Rn×m is the space of all symmetric n × n matrices; Rn×n is of dimension n(n + 1)/2 (see Exercise 3.1). Notation. For a ∈ R and A = [aij] we define A + a = [aij + a]. If all elements of real matrix A are positive, we write A > 0; if all elements of A are nonnegative we write A ≥ 0.
Recommended publications
  • Periodic Staircase Matrices and Generalized Cluster Structures
    PERIODIC STAIRCASE MATRICES AND GENERALIZED CLUSTER STRUCTURES MISHA GEKHTMAN, MICHAEL SHAPIRO, AND ALEK VAINSHTEIN Abstract. As is well-known, cluster transformations in cluster structures of geometric type are often modeled on determinant identities, such as short Pl¨ucker relations, Desnanot–Jacobi identities and their generalizations. We present a construction that plays a similar role in a description of general- ized cluster transformations and discuss its applications to generalized clus- ter structures in GLn compatible with a certain subclass of Belavin–Drinfeld Poisson–Lie brackets, in the Drinfeld double of GLn, and in spaces of periodic difference operators. 1. Introduction Since the discovery of cluster algebras in [4], many important algebraic varieties were shown to support a cluster structure in a sense that the coordinate rings of such variety is isomorphic to a cluster algebra or an upper cluster algebra. Lie theory and representation theory turned out to be a particularly rich source of varieties of this sort including but in no way limited to such examples as Grassmannians [5, 18], double Bruhat cells [1] and strata in flag varieties [15]. In all these examples, cluster transformations that connect distinguished coordinate charts within a ring of regular functions are modeled on three-term relations such as short Pl¨ucker relations, Desnanot–Jacobi identities and their Lie-theoretic generalizations of the kind considered in [3]. This remains true even in the case of exotic cluster structures on GLn considered in [8, 10] where cluster transformations can be obtained by applying Desnanot–Jacobi type identities to certain structured matrices of a size far exceeding n.
    [Show full text]
  • Arxiv:Math/0010243V1
    FOUR SHORT STORIES ABOUT TOEPLITZ MATRIX CALCULATIONS THOMAS STROHMER∗ Abstract. The stories told in this paper are dealing with the solution of finite, infinite, and biinfinite Toeplitz-type systems. A crucial role plays the off-diagonal decay behavior of Toeplitz matrices and their inverses. Classical results of Gelfand et al. on commutative Banach algebras yield a general characterization of this decay behavior. We then derive estimates for the approximate solution of (bi)infinite Toeplitz systems by the finite section method, showing that the approximation rate depends only on the decay of the entries of the Toeplitz matrix and its condition number. Furthermore, we give error estimates for the solution of doubly infinite convolution systems by finite circulant systems. Finally, some quantitative results on the construction of preconditioners via circulant embedding are derived, which allow to provide a theoretical explanation for numerical observations made by some researchers in connection with deconvolution problems. Key words. Toeplitz matrix, Laurent operator, decay of inverse matrix, preconditioner, circu- lant matrix, finite section method. AMS subject classifications. 65T10, 42A10, 65D10, 65F10 0. Introduction. Toeplitz-type equations arise in many applications in mathe- matics, signal processing, communications engineering, and statistics. The excellent surveys [4, 17] describe a number of applications and contain a vast list of references. The stories told in this paper are dealing with the (approximate) solution of biinfinite, infinite, and finite hermitian positive definite Toeplitz-type systems. We pay special attention to Toeplitz-type systems with certain decay properties in the sense that the entries of the matrix enjoy a certain decay rate off the diagonal.
    [Show full text]
  • Realizing Euclidean Distance Matrices by Sphere Intersection
    Realizing Euclidean distance matrices by sphere intersection Jorge Alencara,∗, Carlile Lavorb, Leo Libertic aInstituto Federal de Educa¸c~ao,Ci^enciae Tecnologia do Sul de Minas Gerais, Inconfidentes, MG, Brazil bUniversity of Campinas (IMECC-UNICAMP), 13081-970, Campinas-SP, Brazil cCNRS LIX, Ecole´ Polytechnique, 91128 Palaiseau, France Abstract This paper presents the theoretical properties of an algorithm to find a re- alization of a (full) n × n Euclidean distance matrix in the smallest possible embedding dimension. Our algorithm performs linearly in n, and quadratically in the minimum embedding dimension, which is an improvement w.r.t. other algorithms. Keywords: Distance geometry, sphere intersection, euclidean distance matrix, embedding dimension. 2010 MSC: 00-01, 99-00 1. Introduction Euclidean distance matrices and sphere intersection have a strong mathe- matical importance [1, 2, 3, 4, 5, 6], in addition to many applications, such as navigation problems, molecular and nanostructure conformation, network 5 localization, robotics, as well as other problems of distance geometry [7, 8, 9]. Before defining precisely the problem of interest, we need the formal defini- tion for Euclidean distance matrices. Let D be a n × n symmetric hollow (i.e., with zero diagonal) matrix with non-negative elements. We say that D is a (squared) Euclidean Distance Matrix (EDM) if there are points x1; x2; : : : ; xn 2 RK (for a positive integer K) such that 2 D(i; j) = Dij = kxi − xjk ; i; j 2 f1; : : : ; ng; where k · k denotes the Euclidean norm. The smallest K for which such a set of points exists is called the embedding dimension of D, denoted by dim(D).
    [Show full text]
  • THREE STEPS on an OPEN ROAD Gilbert Strang This Note Describes
    Inverse Problems and Imaging doi:10.3934/ipi.2013.7.961 Volume 7, No. 3, 2013, 961{966 THREE STEPS ON AN OPEN ROAD Gilbert Strang Massachusetts Institute of Technology Cambridge, MA 02139, USA Abstract. This note describes three recent factorizations of banded invertible infinite matrices 1. If A has a banded inverse : A=BC with block{diagonal factors B and C. 2. Permutations factor into a shift times N < 2w tridiagonal permutations. 3. A = LP U with lower triangular L, permutation P , upper triangular U. We include examples and references and outlines of proofs. This note describes three small steps in the factorization of banded matrices. It is written to encourage others to go further in this direction (and related directions). At some point the problems will become deeper and more difficult, especially for doubly infinite matrices. Our main point is that matrices need not be Toeplitz or block Toeplitz for progress to be possible. An important theory is already established [2, 9, 10, 13-16] for non-Toeplitz \band-dominated operators". The Fredholm index plays a key role, and the second small step below (taken jointly with Marko Lindner) computes that index in the special case of permutation matrices. Recall that banded Toeplitz matrices lead to Laurent polynomials. If the scalars or matrices a−w; : : : ; a0; : : : ; aw lie along the diagonals, the polynomial is A(z) = P k akz and the bandwidth is w. The required index is in this case a winding number of det A(z). Factorization into A+(z)A−(z) is a classical problem solved by Plemelj [12] and Gohberg [6-7].
    [Show full text]
  • Arxiv:2004.05118V1 [Math.RT]
    GENERALIZED CLUSTER STRUCTURES RELATED TO THE DRINFELD DOUBLE OF GLn MISHA GEKHTMAN, MICHAEL SHAPIRO, AND ALEK VAINSHTEIN Abstract. We prove that the regular generalized cluster structure on the Drinfeld double of GLn constructed in [9] is complete and compatible with the standard Poisson–Lie structure on the double. Moreover, we show that for n = 4 this structure is distinct from a previously known regular generalized cluster structure on the Drinfeld double, even though they have the same compatible Poisson structure and the same collection of frozen variables. Further, we prove that the regular generalized cluster structure on band periodic matrices constructed in [9] possesses similar compatibility and completeness properties. 1. Introduction In [6, 8] we constructed an initial seed Σn for a complete generalized cluster D structure GCn in the ring of regular functions on the Drinfeld double D(GLn) and proved that this structure is compatible with the standard Poisson–Lie structure on D(GLn). In [9, Section 4] we constructed a different seed Σ¯ n for a regular D D generalized cluster structure GCn on D(GLn). In this note we prove that GCn D shares all the properties of GCn : it is complete and compatible with the standard Poisson–Lie structure on D(GLn). Moreover, we prove that the seeds Σ¯ 4(X, Y ) and T T Σ4(Y ,X ) are not mutationally equivalent. In this way we provide an explicit example of two different regular complete generalized cluster structures on the same variety with the same compatible Poisson structure and the same collection of frozen D variables.
    [Show full text]
  • The Spectra of Large Toeplitz Band Matrices with A
    MATHEMATICS OF COMPUTATION Volume 72, Number 243, Pages 1329{1348 S 0025-5718(03)01505-9 Article electronically published on February 3, 2003 THESPECTRAOFLARGETOEPLITZBANDMATRICES WITH A RANDOMLY PERTURBED ENTRY A. BOTTCHER,¨ M. EMBREE, AND V. I. SOKOLOV Abstract. (j;k) This paper is concerned with the union spΩ Tn(a) of all possible spectra that may emerge when perturbing a large n n Toeplitz band matrix × Tn(a)inthe(j; k) site by a number randomly chosen from some set Ω. The main results give descriptive bounds and, in several interesting situations, even (j;k) provide complete identifications of the limit of sp Tn(a)asn .Also Ω !1 discussed are the cases of small and large sets Ω as well as the \discontinuity of (j;k) the infinite volume case", which means that in general spΩ Tn(a)doesnot converge to something close to sp(j;k) T (a)asn ,whereT (a)isthecor- Ω !1 responding infinite Toeplitz matrix. Illustrations are provided for tridiagonal Toeplitz matrices, a notable special case. 1. Introduction and main results For a complex-valued continuous function a on the complex unit circle T,the infinite Toeplitz matrix T (a) and the finite Toeplitz matrices Tn(a) are defined by n T (a)=(aj k)j;k1 =1 and Tn(a)=(aj k)j;k=1; − − where a` is the `th Fourier coefficient of a, 2π 1 iθ i`θ a` = a(e )e− dθ; ` Z: 2π 2 Z0 Here, we restrict our attention to the case where a is a trigonometric polynomial, a , implying that at most a finite number of the Fourier coefficients are nonzero; equivalently,2P T (a) is a banded matrix.
    [Show full text]
  • Arxiv:1701.03378V1 [Math.RA] 12 Jan 2017 Setao Apihe Rusadfe Probability” Free and Groups T Lamplighter by on Supported Technology
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by TUGraz OPEN Library Linearizing the Word Problem in (some) Free Fields Konrad Schrempf∗ January 13, 2017 Abstract We describe a solution of the word problem in free fields (coming from non- commutative polynomials over a commutative field) using elementary linear algebra, provided that the elements are given by minimal linear representa- tions. It relies on the normal form of Cohn and Reutenauer and can be used more generally to (positively) test rational identities. Moreover we provide a construction of minimal linear representations for the inverse of non-zero elements. Keywords: word problem, minimal linear representation, linearization, realiza- tion, admissible linear system, rational series AMS Classification: 16K40, 16S10, 03B25, 15A22 Introduction arXiv:1701.03378v1 [math.RA] 12 Jan 2017 Free (skew) fields arise as universal objects when it comes to embed the ring of non-commutative polynomials, that is, polynomials in (a finite number of) non- commuting variables, into a skew field [Coh85, Chapter 7]. The notion of “free fields” goes back to Amitsur [Ami66]. A brief introduction can be found in [Coh03, Section 9.3], for details we refer to [Coh95, Section 6.4]. In the present paper we restrict the setting to commutative ground fields, as a special case. See also [Rob84]. In [CR94], Cohn and Reutenauer introduced a normal form for elements in free fields in order to extend results from the theory of formal languages. In particular they characterize minimality of linear representations in terms of linear independence of ∗Contact: [email protected], Department of Discrete Mathematics (Noncommutative Structures), Graz University of Technology.
    [Show full text]
  • Conversion of Sparse Matrix to Band Matrix Using an Fpga For
    CONVERSION OF SPARSE MATRIX TO BAND MATRIX USING AN FPGA FOR HIGH-PERFORMANCE COMPUTING by Anjani Chaudhary, M.S. A thesis submitted to the Graduate Council of Texas State University in partial fulfillment of the requirements for the degree of Master of Science with a Major in Engineering December 2020 Committee Members: Semih Aslan, Chair Shuying Sun William A Stapleton COPYRIGHT by Anjani Chaudhary 2020 FAIR USE AND AUTHOR’S PERMISSION STATEMENT Fair Use This work is protected by the Copyright Laws of the United States (Public Law 94-553, section 107). Consistent with fair use as defined in the Copyright Laws, brief quotations from this material are allowed with proper acknowledgement. Use of this material for financial gain without the author’s express written permission is not allowed. Duplication Permission As the copyright holder of this work, I, Anjani Chaudhary, authorize duplication of this work, in whole or in part, for educational or scholarly purposes only. ACKNOWLEDGEMENTS Throughout my research and report writing process, I am grateful to receive support and suggestions from the following people. I would like to thank my supervisor and chair, Dr. Semih Aslan, Associate Professor, Ingram School of Engineering, for his time, support, encouragement, and patience, without which I would not have made it. My committee members Dr. Bill Stapleton, Associate Professor, Ingram School of Engineering, and Dr. Shuying Sun, Associate Professor, Department of Mathematics, for their essential guidance throughout the research and my graduate advisor, Dr. Vishu Viswanathan, for providing the opportunity and valuable suggestions which helped me to come one more step closer to my goal.
    [Show full text]
  • Copyrighted Material
    JWST600-IND JWST600-Penney September 28, 2015 7:40 Printer Name: Trim: 6.125in × 9.25in INDEX Additive property of determinants, 244 Cofactor expansion, 240 Argument of a complex number, 297 Column space, 76 Column vector, 2 Band matrix, 407 Complementary vector, 319 Band width, 407 Complex eigenvalue, 302 Basis Complex eigenvector, 302 chain, 438 Complex linear transformation, 304 coordinate matrix for a basis, 217 Complex matrices, 301 coordinate transformation, 223 Complex number definition, 104 argument, 297 normalization, 315 conjugate, 302 ordered, 216 Complex vector space, 303 orthonormal, 315 Conjugate of a complex number, 302 point transformation, 223 Consumption matrix, 198 standard Coordinate matrix, 217 M(m,n), 119 Coordinates n, 119 coordinate vector, 216 Rn, 118 coordinate matrix for a basis, 217 COPYRIGHTEDcoordinate MATERIAL vector, 216 Cn, 302 point matrix, 217 Chain basis, 438 point transformation, 223 Characteristic polynomial, 275 Cramer’s rule, 265 Coefficient matrix, 75 Current, 41 Linear Algebra: Ideas and Applications, Fourth Edition. Richard C. Penney. © 2016 John Wiley & Sons, Inc. Published 2016 by John Wiley & Sons, Inc. Companion website: www.wiley.com/go/penney/linearalgebra 487 JWST600-IND JWST600-Penney September 28, 2015 7:40 Printer Name: Trim: 6.125in × 9.25in 488 INDEX Data compression, 349 Haar function, 346 Determinant Hermitian additive property, 244 adjoint, 412 cofactor expansion, 240 symmetric, 414 Cramer’s rule, 265 Hermitian, orthogonal, 413 inverse matrix, 267 Homogeneous system, 80 Laplace
    [Show full text]
  • Left Eigenvector of a Stochastic Matrix
    Advances in Pure Mathematics, 2011, 1, 105-117 doi:10.4236/apm.2011.14023 Published Online July 2011 (http://www.SciRP.org/journal/apm) Left Eigenvector of a Stochastic Matrix Sylvain Lavalle´e Departement de mathematiques, Universite du Quebec a Montreal, Montreal, Canada E-mail: [email protected] Received January 7, 2011; revised June 7, 2011; accepted June 15, 2011 Abstract We determine the left eigenvector of a stochastic matrix M associated to the eigenvalue 1 in the commu- tative and the noncommutative cases. In the commutative case, we see that the eigenvector associated to the eigenvalue 0 is (,,NN1 n ), where Ni is the ith principal minor of NMI= n , where In is the 11 identity matrix of dimension n . In the noncommutative case, this eigenvector is (,P1 ,Pn ), where Pi is the sum in aij of the corresponding labels of nonempty paths starting from i and not passing through i in the complete directed graph associated to M . Keywords: Generic Stochastic Noncommutative Matrix, Commutative Matrix, Left Eigenvector Associated To The Eigenvalue 1, Skew Field, Automata 1. Introduction stochastic free field and that the vector 11 (,,PP1 n ) is fixed by our matrix; moreover, the sum 1 It is well known that 1 is one of the eigenvalue of a of the Pi is equal to 1, hence they form a kind of stochastic matrix (i.e. the sum of the elements of each noncommutative limiting probability. row is equal to 1) and its associated right eigenvector is These results have been proved in [1] but the proof the vector (1,1, ,1)T .
    [Show full text]
  • NAG Library Chapter Contents F16 – NAG Interface to BLAS F16 Chapter Introduction
    F16 – NAG Interface to BLAS Contents – F16 NAG Library Chapter Contents f16 – NAG Interface to BLAS f16 Chapter Introduction Function Mark of Name Introduction Purpose f16dbc 7 nag_iload Broadcast scalar into integer vector f16dlc 9 nag_isum Sum elements of integer vector f16dnc 9 nag_imax_val Maximum value and location, integer vector f16dpc 9 nag_imin_val Minimum value and location, integer vector f16dqc 9 nag_iamax_val Maximum absolute value and location, integer vector f16drc 9 nag_iamin_val Minimum absolute value and location, integer vector f16eac 24 nag_ddot Dot product of two vectors, allows scaling and accumulation. f16ecc 7 nag_daxpby Real weighted vector addition f16ehc 9 nag_dwaxpby Real weighted vector addition preserving input f16elc 9 nag_dsum Sum elements of real vector f16fbc 7 nag_dload Broadcast scalar into real vector f16gcc 24 nag_zaxpby Complex weighted vector addition f16ghc 9 nag_zwaxpby Complex weighted vector addition preserving input f16glc 9 nag_zsum Sum elements of complex vector f16hbc 7 nag_zload Broadcast scalar into complex vector f16jnc 9 nag_dmax_val Maximum value and location, real vector f16jpc 9 nag_dmin_val Minimum value and location, real vector f16jqc 9 nag_damax_val Maximum absolute value and location, real vector f16jrc 9 nag_damin_val Minimum absolute value and location, real vector f16jsc 9 nag_zamax_val Maximum absolute value and location, complex vector f16jtc 9 nag_zamin_val Minimum absolute value and location, complex vector f16pac 8 nag_dgemv Matrix-vector product, real rectangular matrix
    [Show full text]
  • THE EFFECTIVE POTENTIAL of an M-MATRIX Marcel Filoche, Svitlana Mayboroda, Terence Tao
    THE EFFECTIVE POTENTIAL OF AN M-MATRIX Marcel Filoche, Svitlana Mayboroda, Terence Tao To cite this version: Marcel Filoche, Svitlana Mayboroda, Terence Tao. THE EFFECTIVE POTENTIAL OF AN M- MATRIX. 2021. hal-03189307 HAL Id: hal-03189307 https://hal.archives-ouvertes.fr/hal-03189307 Preprint submitted on 3 Apr 2021 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. THE EFFECTIVE POTENTIAL OF AN M-MATRIX MARCEL FILOCHE, SVITLANA MAYBORODA, AND TERENCE TAO Abstract. In the presence of a confining potential V, the eigenfunctions of a con- tinuous Schrodinger¨ operator −∆ + V decay exponentially with the rate governed by the part of V which is above the corresponding eigenvalue; this can be quan- tified by a method of Agmon. Analogous localization properties can also be es- tablished for the eigenvectors of a discrete Schrodinger¨ matrix. This note shows, perhaps surprisingly, that one can replace a discrete Schrodinger¨ matrix by any real symmetric Z-matrix and still obtain eigenvector localization estimates. In the case of a real symmetric non-singular M-matrix A (which is a situation that arises in several contexts, including random matrix theory and statistical physics), the landscape function u = A−11 plays the role of an effective potential of localiza- tion.
    [Show full text]