Frobenius Normal Forms of Doubly Stochastic Matrices Received August 15, 2019; Accepted November 22, 2019

Total Page:16

File Type:pdf, Size:1020Kb

Frobenius Normal Forms of Doubly Stochastic Matrices Received August 15, 2019; Accepted November 22, 2019 Spec. Matrices 2019; 7:213–217 Note Open Access Special Issue Dedicated to Charles R. Johnson Pietro Paparella* Frobenius normal forms of doubly stochastic matrices https://doi.org/10.1515/spma-2019-0015 Received August 15, 2019; accepted November 22, 2019 Abstract: An elementary proof of a fundamental result on doubly stochastic matrices in Frobenius normal form is given. This result is used to establish several well-known results concerning permutations, including a theorem due to Runi. Keywords: Frobenius normal form, doubly stochastic matrix, permutation, permutation matrix MSC: 15A21; 05A05 1 Introduction In 1965, Perfect and Mirsky [5, Lemma 3] stated, without proof, that a doubly stochastic matrix is permuta- tionally similar to a direct sum of irreducible, doubly stochastic matrices. Furthermore, they state that the result “is almost certainly well-known” and omit a proof because it “follows very easily from the denitions” [5, p. 38]. This result is fundamental in the Perfect–Mirsky conjecture, but, to the best of our knowledge, it ap- pears sparingly in the literature: Liu and Lai [4, Theorem 2.7.4] prove the weaker result that a doubly stochastic matrix is either irreducible or permutationally similar to a direct sum of two doubly stochastic matrices; and Hartel and Spellman [2, Lemma 1(b)] give a proof via strong induction. In this work, an elementary proof of this fundamental result is provided that relies on weak induction and the Frobenius normal form of a matrix. To demonstrate its import and utility, we apply it to permutation matrices and and characterize the Frobenius normal forms of a permutation matrix. This, in turn, is used to derive two well-known results concerning permutations, including the disjoint cyclic form and the result due to Runi that the order of a permutation in disjoint cyclic form is the least common multiple of the lengths of its disjoint cycles. 2 Background n For F = C or F = R, we let Mn(F) denote the set of n-by-n matrices with entries over F, and F denote the collection of all column vectors of length n over F. We let In denote the n-by-n identity matrix and e denotes the all-ones vector (the size of which is determined by the context in which it appears). Finally, for n 2 N, we let hni := f1, ... , ng. A directed graph (or simply digraph) Γ = (V, E) consists of a nite, nonempty set V of vertices, together with a set of arcs E ⊆ V × V. A digraph Γ is strongly connected if, for any two vertices u and v of V, there is a (directed) walk in Γ from u to v Every vertex of V is considered strongly connected to itself so that strong *Corresponding Author: Pietro Paparella: University of Washington Bothell, E-mail: [email protected] Open Access. © 2020 Pietro Paparella, published by De Gruyter. This work is licensed under the Creative Commons Attribution alone 4.0 License. 214 Ë Pietro Paparella connectivity denes an equivalence relation, and hence a partition, of the vertices into strongly connected components. Notice that Γ is strongly connected if and only if Γ possesses one strongly connected component. If A 2 Mn(C), then the digraph of A, denoted by Γ(A), has vertices V = hni and arcs E = f(i, j) 2 V × V | aij ≠ 0g. For n ≥ 2, a matrix A 2 Mn(C), is reducible if there is a permutation matrix P such that " # > A A P AP = 11 12 , 0 A22 in which A11 and A22 are nonempty square matrices and 0 is a zero block. If A is not reducible, then A is called irreducible. It is well-known that A is irreducible if and only if Γ (A) is strongly connected (see, e.g., Brualdi and Ryser [1, Theorem 3.2.1]). If A 2 Mn(C), then there is a permutation matrix P such that 2 3 A11 ··· A1k > 6 7 P AP = 6 .. 7 , 4 . 5 Akk > in which the matrices A11, ... , Akk are irreducible [1, Theorem 3.2.4]. The matrix P AP is called a Frobe- nius normal form (FNF) of A and is not unique. The irreducible matrices A11, ... , Akk, called the irreducible components of A, are unique up to permutation similarity. 3 Main Result > Recall that a nonnegative matrix A 2 Mn(R) is called doubly stochastic if Ae = e = A e. With this denition, we are now ready to present the main result. Theorem 3.1. Let A 2 Mn(R) and suppose that 2 3 A11 ··· A1k > 6 7 P AP = 6 .. 7 4 . 5 Akk is a Frobenius normal form of A. If A is doubly stochastic, then Aij = 0, i < j, i.e., 2 3 k A11 > M 6 7 P AP = Aii = 6 .. 7 4 . 5 i=1 Akk and the irreducible components A11 2 Mn1 (R), ... , Akk 2 Mnk (R) are doubly stochastic. Proof. Proceed by induction on k, the number of irreducible blocks in any FNF of A. When k = 1, then A is irreducible and the result is clear. For the induction-step, assume that k ≥ 2 and that the result holds for any doubly stochastic matrix having k − 1 irreducible blocks in any FNF. > > > Since Pe = P e = e, it follows that P AP is doubly stochastic. Consequently, A11e = e and A11e ≤ e (inequality here is considered entrywise). We claim that A11e = e. Otherwise, > > > > n1 = e e = e A11 e = e (A11e) < e e = n1, Spectrally Perron Polynomials Ë 215 a contradiction. Thus, A11 is doubly stochastic and 2 3 A11 6 A A 7 > 6 22 ··· 2k7 P AP = 6 7 . 6 .. 7 4 . 5 Akk The submatrix 2 3 A22 ··· A2k 6 7 6 .. 7 4 . 5 Akk > of P AP is doubly stochastic and has k − 1 irreducible blocks; as such, the induction-hypothesis applies and the result is established. n > n Remark 3.2. For α > 0 let CSα := fA 2 Mn(R) | A ≥ 0, Ae = αe = A eg. Notice that A 2 CSα if and only if n n ri = cj = α, 8i, j 2 hni. Furthermore, A 2 CSα if and only if B := A/α 2 CS1, therefore Theorem 3.1 applies to n matrices in CSα, α > 0. 4 Permutations and Permutation Matrices. Denote by Sn the symmetric group of hni. For σ 2 Sn, the permutation matrix corresponding to σ, denoted by Pσ, is the the n-by-n matrix such that (i, j)-entry is δi,σ−1(i), where δij denotes the Kronecker delta. When the context is clear, Pσ is abbreviated to P. As is well-known, Pσ Pγ = Pσγ and so the set of all n-by-n permutation matrices, denoted by Pn, forms a group under matrix multiplication and the map Φ : Sn −! Pn, dened by Φ(σ) = Pσ is an isomorphism. Observation 4.1. If P 2 Mn(R) is a permutation matrix, then P is irreducible if and only if Γ (P) is an n-cycle. Proof. Let v 2 hni. Since Γ (P) is strongly connected, Γ (P) has a k-cycle of distinct vertices v := v1 −! ··· −! vk −! v1 with 1 ≤ k ≤ n. We claim that k = n; otherwise, if k < n, then Γ (P) clearly contains more than one connected component, which is a contradiction. Thus, k = n, i.e., Γ (P) is an n-cycle. Conversely, the matrix P is clearly irreducible when Γ (P) is an n-cycle. " # 0 1 Observation 4.2. For a positive integer n ≥ 2, the matrix Cn := is called a basic circulant. No- In−1 0 tice that Cn is the permutation matrix corresponding to the cyclic permutation 1 ··· n . Thus, if P is an > irreducible permutation matrix, then there is a permutation matrix Q such that Q PQ = Cn. Corollary 4.3 (FNF of a permuation matrix). If P is a permutation matrix, then there is a permutation matrix Q Q>PQ Lk C k n such that = i=1 ni , 1 ≤ ≤ . Lemma 4.4. If α = α1 ··· αk 2 Sn and β = β1 ··· β` 2 Sn are disjoint, cyclic permutations (i.e., −1 −1 αi ≠ βj) and γ 2 Sn, then γαγ and γβγ are disjoint, cyclic permutations. −1 Proof. If yi := γ(αi), 1 ≤ i ≤ k, then γ(α(γ (yi))) = γ(α(αi)) = γ(αi+1) = yi+1, where, for convenience, −1 −1 −1 αk := α and yk := y . If y 2 hninfy , ... , ykg and x := γ (y), then (γαγ )(y) = γ(α(γ (y))) = γ(α(x)) = +1 1 +1 1 1 −1 −1 γ(x) = y. Thus, γαγ = y1 ··· yk . A similar argument demonstrates that γβγ = z1 ··· z` with zi := γ(βi), 1 ≤ i ≤ `. 216 Ë Pietro Paparella For contradiction, if γαγ−1 and γβγ−1 are not disjoint, then there are positive integers i and j such that γ(αi) = yi = zj = γ(βj), i.e., αi = βj, a contradiction. Theorem 4.5 (disjoint cyclic form). If σ 2 Sn, then there are disjoint, cyclic permutations σ1, ... , σk 2 Sn, k n σ Qk σ 1 ≤ ≤ , such that = i=1 i. Proof. If P is the permutation matrix corresponding to σ, then, following Corollary 4.3, there is a permutation matrix Q corresponding to γ 2 Sn such that k ! k M > M > P = Q Cni Q = QCni Q , 1 ≤ k ≤ n. i=1 i=1 > If σˆi is the cyclic permutation corresponding to Cni and σi is the permutation corresponding to QCni Q , i.e., σ γσ γ−1 σ ... σ σ Qk σ i = ˆi , then 1, , k are pairwise disjoint (Lemma 4.4) and = i=1 i, as desired.
Recommended publications
  • Math 511 Advanced Linear Algebra Spring 2006
    MATH 511 ADVANCED LINEAR ALGEBRA SPRING 2006 Sherod Eubanks HOMEWORK 2 x2:1 : 2; 5; 9; 12 x2:3 : 3; 6 x2:4 : 2; 4; 5; 9; 11 Section 2:1: Unitary Matrices Problem 2 If ¸ 2 σ(U) and U 2 Mn is unitary, show that j¸j = 1. Solution. If ¸ 2 σ(U), U 2 Mn is unitary, and Ux = ¸x for x 6= 0, then by Theorem 2:1:4(g), we have kxkCn = kUxkCn = k¸xkCn = j¸jkxkCn , hence j¸j = 1, as desired. Problem 5 Show that the permutation matrices in Mn are orthogonal and that the permutation matrices form a sub- group of the group of real orthogonal matrices. How many different permutation matrices are there in Mn? Solution. By definition, a matrix P 2 Mn is called a permutation matrix if exactly one entry in each row n and column is equal to 1, and all other entries are 0. That is, letting ei 2 C denote the standard basis n th element of C that has a 1 in the i row and zeros elsewhere, and Sn be the set of all permutations on n th elements, then P = [eσ(1) j ¢ ¢ ¢ j eσ(n)] = Pσ for some permutation σ 2 Sn such that σ(k) denotes the k member of σ. Observe that for any σ 2 Sn, and as ½ 1 if i = j eT e = σ(i) σ(j) 0 otherwise for 1 · i · j · n by the definition of ei, we have that 2 3 T T eσ(1)eσ(1) ¢ ¢ ¢ eσ(1)eσ(n) T 6 .
    [Show full text]
  • Chapter Four Determinants
    Chapter Four Determinants In the first chapter of this book we considered linear systems and we picked out the special case of systems with the same number of equations as unknowns, those of the form T~x = ~b where T is a square matrix. We noted a distinction between two classes of T ’s. While such systems may have a unique solution or no solutions or infinitely many solutions, if a particular T is associated with a unique solution in any system, such as the homogeneous system ~b = ~0, then T is associated with a unique solution for every ~b. We call such a matrix of coefficients ‘nonsingular’. The other kind of T , where every linear system for which it is the matrix of coefficients has either no solution or infinitely many solutions, we call ‘singular’. Through the second and third chapters the value of this distinction has been a theme. For instance, we now know that nonsingularity of an n£n matrix T is equivalent to each of these: ² a system T~x = ~b has a solution, and that solution is unique; ² Gauss-Jordan reduction of T yields an identity matrix; ² the rows of T form a linearly independent set; ² the columns of T form a basis for Rn; ² any map that T represents is an isomorphism; ² an inverse matrix T ¡1 exists. So when we look at a particular square matrix, the question of whether it is nonsingular is one of the first things that we ask. This chapter develops a formula to determine this. (Since we will restrict the discussion to square matrices, in this chapter we will usually simply say ‘matrix’ in place of ‘square matrix’.) More precisely, we will develop infinitely many formulas, one for 1£1 ma- trices, one for 2£2 matrices, etc.
    [Show full text]
  • On Finding Multiplicities of Characteristic Polynomial Factors Of
    On finding multiplicities of characteristic polynomial factors of black-box matrices∗. Jean-Guillaume Dumas† Cl´ement Pernet‡ B. David Saunders§ November 2, 2018 Abstract We present algorithms and heuristics to compute the characteristic polynomial of a matrix given its minimal polynomial. The matrix is rep- resented as a black-box, i.e., by a function to compute its matrix-vector product. The methods apply to matrices either over the integers or over a large enough finite field. Experiments show that these methods perform efficiently in practice. Combined in an adaptive strategy, these algorithms reach significant speedups in practice for some integer matrices arising in an application from graph theory. Keywords: Characteristic polynomial ; black-box matrix ; finite field. 1 Introduction Computing the characteristic polynomial of an integer matrix is a classical math- ematical problem. It is closely related to the computation of the Frobenius nor- mal form which can be used to test two matrices for similarity, or computing invariant subspaces under the action of the matrix. Although the Frobenius normal form contains more information on the matrix than the characteristic arXiv:0901.4747v2 [cs.SC] 18 May 2009 polynomial, most algorithms to compute it are based on computations of char- acteristic polynomials (see for example [25, 9.7]). Several matrix representations are used in§ computational linear algebra. In the dense representation, a m n matrix is considered as the array of all the m n coefficients. The sparse representation× only considers non-zero coefficients using× different possible data structures. In the black-box representation, the matrix ∗Saunders supported by National Science Foundation Grants CCF-0515197, CCF-0830130.
    [Show full text]
  • Rotation Matrix - Wikipedia, the Free Encyclopedia Page 1 of 22
    Rotation matrix - Wikipedia, the free encyclopedia Page 1 of 22 Rotation matrix From Wikipedia, the free encyclopedia In linear algebra, a rotation matrix is a matrix that is used to perform a rotation in Euclidean space. For example the matrix rotates points in the xy -Cartesian plane counterclockwise through an angle θ about the origin of the Cartesian coordinate system. To perform the rotation, the position of each point must be represented by a column vector v, containing the coordinates of the point. A rotated vector is obtained by using the matrix multiplication Rv (see below for details). In two and three dimensions, rotation matrices are among the simplest algebraic descriptions of rotations, and are used extensively for computations in geometry, physics, and computer graphics. Though most applications involve rotations in two or three dimensions, rotation matrices can be defined for n-dimensional space. Rotation matrices are always square, with real entries. Algebraically, a rotation matrix in n-dimensions is a n × n special orthogonal matrix, i.e. an orthogonal matrix whose determinant is 1: . The set of all rotation matrices forms a group, known as the rotation group or the special orthogonal group. It is a subset of the orthogonal group, which includes reflections and consists of all orthogonal matrices with determinant 1 or -1, and of the special linear group, which includes all volume-preserving transformations and consists of matrices with determinant 1. Contents 1 Rotations in two dimensions 1.1 Non-standard orientation
    [Show full text]
  • Alternating Sign Matrices, Extensions and Related Cones
    See discussions, stats, and author profiles for this publication at: https://www.researchgate.net/publication/311671190 Alternating sign matrices, extensions and related cones Article in Advances in Applied Mathematics · May 2017 DOI: 10.1016/j.aam.2016.12.001 CITATIONS READS 0 29 2 authors: Richard A. Brualdi Geir Dahl University of Wisconsin–Madison University of Oslo 252 PUBLICATIONS 3,815 CITATIONS 102 PUBLICATIONS 1,032 CITATIONS SEE PROFILE SEE PROFILE Some of the authors of this publication are also working on these related projects: Combinatorial matrix theory; alternating sign matrices View project All content following this page was uploaded by Geir Dahl on 16 December 2016. The user has requested enhancement of the downloaded file. All in-text references underlined in blue are added to the original document and are linked to publications on ResearchGate, letting you access and read them immediately. Alternating sign matrices, extensions and related cones Richard A. Brualdi∗ Geir Dahly December 1, 2016 Abstract An alternating sign matrix, or ASM, is a (0; ±1)-matrix where the nonzero entries in each row and column alternate in sign, and where each row and column sum is 1. We study the convex cone generated by ASMs of order n, called the ASM cone, as well as several related cones and polytopes. Some decomposition results are shown, and we find a minimal Hilbert basis of the ASM cone. The notion of (±1)-doubly stochastic matrices and a generalization of ASMs are introduced and various properties are shown. For instance, we give a new short proof of the linear characterization of the ASM polytope, in fact for a more general polytope.
    [Show full text]
  • More Gaussian Elimination and Matrix Inversion
    Week 7 More Gaussian Elimination and Matrix Inversion 7.1 Opening Remarks 7.1.1 Introduction * View at edX 289 Week 7. More Gaussian Elimination and Matrix Inversion 290 7.1.2 Outline 7.1. Opening Remarks..................................... 289 7.1.1. Introduction..................................... 289 7.1.2. Outline....................................... 290 7.1.3. What You Will Learn................................ 291 7.2. When Gaussian Elimination Breaks Down....................... 292 7.2.1. When Gaussian Elimination Works........................ 292 7.2.2. The Problem.................................... 297 7.2.3. Permutations.................................... 299 7.2.4. Gaussian Elimination with Row Swapping (LU Factorization with Partial Pivoting) 304 7.2.5. When Gaussian Elimination Fails Altogether................... 311 7.3. The Inverse Matrix.................................... 312 7.3.1. Inverse Functions in 1D.............................. 312 7.3.2. Back to Linear Transformations.......................... 312 7.3.3. Simple Examples.................................. 314 7.3.4. More Advanced (but Still Simple) Examples................... 320 7.3.5. Properties...................................... 324 7.4. Enrichment......................................... 325 7.4.1. Library Routines for LU with Partial Pivoting................... 325 7.5. Wrap Up.......................................... 327 7.5.1. Homework..................................... 327 7.5.2. Summary...................................... 327 7.1. Opening
    [Show full text]
  • Pattern Avoidance in Alternating Sign Matrices
    PATTERN AVOIDANCE IN ALTERNATING SIGN MATRICES ROBERT JOHANSSON AND SVANTE LINUSSON Abstract. We generalize the definition of a pattern from permu- tations to alternating sign matrices. The number of alternating sign matrices avoiding 132 is proved to be counted by the large Schr¨oder numbers, 1, 2, 6, 22, 90, 394 . .. We give a bijection be- tween 132-avoiding alternating sign matrices and Schr¨oder-paths, which gives a refined enumeration. We also show that the 132, 123- avoiding alternating sign matrices are counted by every second Fi- bonacci number. 1. Introduction For some time, we have emphasized the use of permutation matrices when studying pattern avoidance. This and the fact that alternating sign matrices are generalizations of permutation matrices, led to the idea of studying pattern avoidance in alternating sign matrices. The main results, concerning 132-avoidance, are presented in Section 3. In Section 5, we enumerate sets of alternating sign matrices avoiding two patterns of length three at once. In Section 6 we state some open problems. Let us start with the basic definitions. Given a permutation π of length n we will represent it with the n×n permutation matrix with 1 in positions (i, π(i)), 1 ≤ i ≤ n and 0 in all other positions. We will also suppress the 0s. 1 1 3 1 2 1 Figure 1. The permutation matrix of 132. In this paper, we will think of permutations in terms of permutation matrices. Definition 1.1. (Patterns in permutations) A permutation π is said to contain a permutation τ as a pattern if the permutation matrix of τ is a submatrix of π.
    [Show full text]
  • Matrices That Commute with a Permutation Matrix
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Elsevier - Publisher Connector Matrices That Commute With a Permutation Matrix Jeffrey L. Stuart Department of Mathematics University of Southern Mississippi Hattiesburg, Mississippi 39406-5045 and James R. Weaver Department of Mathematics and Statistics University of West Florida Pensacola. Florida 32514 Submitted by Donald W. Robinson ABSTRACT Let P be an n X n permutation matrix, and let p be the corresponding permuta- tion. Let A be a matrix such that AP = PA. It is well known that when p is an n-cycle, A is permutation similar to a circulant matrix. We present results for the band patterns in A and for the eigenstructure of A when p consists of several disjoint cycles. These results depend on the greatest common divisors of pairs of cycle lengths. 1. INTRODUCTION A central theme in matrix theory concerns what can be said about a matrix if it commutes with a given matrix of some special type. In this paper, we investigate the combinatorial structure and the eigenstructure of a matrix that commutes with a permutation matrix. In doing so, we follow a long tradition of work on classes of matrices that commute with a permutation matrix of a specified type. In particular, our work is related both to work on the circulant matrices, the results for which can be found in Davis’s compre- hensive book [5], and to work on the centrosymmetric matrices, which can be found in the book by Aitken [l] and in the papers by Andrew [2], Bamett, LZNEAR ALGEBRA AND ITS APPLICATIONS 150:255-265 (1991) 255 Q Elsevier Science Publishing Co., Inc., 1991 655 Avenue of the Americas, New York, NY 10010 0024-3795/91/$3.50 256 JEFFREY L.
    [Show full text]
  • The Waldspurger Transform of Permutations and Alternating Sign Matrices
    Séminaire Lotharingien de Combinatoire 78B (2017) Proceedings of the 29th Conference on Formal Power Article #41, 12 pp. Series and Algebraic Combinatorics (London) The Waldspurger Transform of Permutations and Alternating Sign Matrices Drew Armstrong∗ and James McKeown Department of Mathematics, University of Miami, Coral Gables, FL Abstract. In 2005 J. L. Waldspurger proved the following theorem: Given a finite real reflection group G, the closed positive root cone is tiled by the images of the open weight cone under the action of the linear transformations 1 g. Shortly after this − E. Meinrenken extended the result to affine Weyl groups and then P. V. Bibikov and V. S. Zhgoon gave a uniform proof for a discrete reflection group acting on a simply- connected space of constant curvature. In this paper we show that the Waldspurger and Meinrenken theorems of type A give an interesting new perspective on the combinatorics of the symmetric group. In particular, for each permutation matrix g S we define a non-negative integer 2 n matrix WT(g), called the Waldspurger transform of g. The definition of the matrix WT(g) is purely combinatorial but it turns out that its columns are the images of the fundamental weights under 1 g, expressed in simple root coordinates. The possible − columns of WT(g) (which we call UM vectors) biject to many interesting structures including: unimodal Motzkin paths, abelian ideals in the Lie algebra sln(C), Young diagrams with maximum hook length n, and integer points inside a certain polytope. We show that the sum of the entries of WT(g) is half the entropy of the corresponding permutation g, which is known to equal the rank of g in the MacNeille completion of the Bruhat order.
    [Show full text]
  • Observavility Brunovsky Normal Form: Multi-Output Linear Dynamical Systems Driss Boutat, Frédéric Kratz, Jean-Pierre Barbot
    Observavility Brunovsky Normal Form: Multi-Output Linear Dynamical Systems Driss Boutat, Frédéric Kratz, Jean-Pierre Barbot To cite this version: Driss Boutat, Frédéric Kratz, Jean-Pierre Barbot. Observavility Brunovsky Normal Form: Multi- Output Linear Dynamical Systems. American Control Conference, ACC09, IEEE, Jun 2009, St. Louis, MO, United States. hal-00772281 HAL Id: hal-00772281 https://hal.inria.fr/hal-00772281 Submitted on 10 Jan 2013 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Observavility Brunovsky Normal Form: Multi-Output Linear Dynamical Systems Driss Boutat, Fred´ eric´ Kratz and Jean-Pierre Barbot Abstract— This paper gives the sufficient and necessary II. NOTATION AND PROBLEM STATEMENT conditions to guarantee the existence of a linear change of co- ordinates to transform a multi-output linear dynamical system (modulo a nonlinear term depending on inputs and outputs) in the observability Brunovsky canonical form. Consider the following multi-output dynamical system: I. INTRODUCTION x˙ = Ax + γ(y, u) (1) y = Cx (2) For a single output dynamical linear system the observability rank condition is a necessary and sufficient condition to transform it into the Brunovsky observability where: normal form. In this last form, it is possible to use classical observer such that [8] observer and, [5] observer.
    [Show full text]
  • A Rough Guide to Linear Algebra
    A rough guide to linear algbera Dongryul Kim Version of April 18, 2020 ii Contents Preface v 1 Sets 1 1.1 Sets and maps . .1 1.2 Products, coproducts, and sets of maps . .4 1.3 Fun with diagrams . .7 1.4 Equivalence classes . .9 2 Vector spaces 13 2.1 Fields . 13 2.2 Vector spaces . 15 2.3 Matrices . 18 2.4 Products and direct sums . 21 2.5 Subspaces and quotients . 27 2.6 Vector spaces from linear maps . 29 2.7 Bases and dimension . 34 2.8 Dual spaces . 44 2.9 Linear algebra in combinatorics . 50 3 Multilinear algebra 55 3.1 Bilinear maps and tensor products . 55 3.2 Symmetric and exterior algebras . 62 3.3 The determinant . 68 3.4 Computing the inverse matrix . 72 4 Linear algebra without division 79 4.1 Commutative rings . 79 4.2 Modules . 82 4.3 Classification of finitely generated modules over a PID . 86 4.4 Frobenius and Jordan normal form . 97 4.5 Eigenvalues and eigenvectors . 103 iii iv CONTENTS 5 Linear algebra over R and C 109 5.1 A bit of analysis . 109 5.2 Inner products . 112 5.3 Operators on Hilbert spaces . 118 5.4 The spectral theorem . 121 5.5 Positivity of operators . 125 5.6 Duality in linear programming . 130 Epilogue 131 Preface At a certain point, one adopts a mode of learning mathematics. Then at a later point, after progressing through various styles of teaching and writing, one develops one's own point of view for how mathematics is to be taught.
    [Show full text]
  • On the Generalized Nullspace of M-Matrices and Z-Matrices*
    On the Generalized Nullspace of M-Matrices and Z-Matrices* Daniel Hershkowitz Department of Mathematics Technion - lsrael Institute of Technology Haifa, 32000 Israel and Hans Schneider Mathematics Department University of Wisconsin -Madison Madison, Wisconsin 53706 Submitted by Richard A. Bxualdi ABSTRACT A proof is given for the preferred basis theorem for the generalized nullspace of a given M-matrix. The theorem is then generalized and extended to the case of a Z-matrix. 1. INTRODUCTION In this paper we give a proof for a known result, namely, the preferred basis theorem for the generalized nullspace of a given M-matrix. We then generalize and extend the theorem to the case of a Z-matrix. Let A be a Zmatrix. A preferred set is an ordered set of semipositive vectors such that the image of each vector under - A is a nonnegative linear combination of the subsequent vectors. Furthermore, the positivity of the *The research of both authors was supported by their joint grant No. 8500153 from the United States-Israel Binational Science Foundation (BSF), J erusalem, Israel. The second author was supported also by NSF grant DMS8521521 and by ONR grant NOOO14-85K-1613. LINEAR ALGEBRA AND ITS APPLICATlONS 10&S-23 (1988) 5 0 Elsevier Science Publishing Co., Inc., 1988 52 Vanderbilt Ave., New York, NY 10017 0024-3795/88/$3.50 6 DANIEL HERSHKOWITZ AND HANS SCHNEIDER entries of the vectors and the coefficients in the linear combinations depends entirely on the graph structure of A in a specified manner. The preferred basis theorem asserts that the generalized nullspace E(A) of an M-matrix A has a basis which is a preferred set.
    [Show full text]