Majorization and the Schur-Horn Theorem
Total Page:16
File Type:pdf, Size:1020Kb
MAJORIZATION AND THE SCHUR-HORN THEOREM A Thesis Submitted to the Faculty of Graduate Studies and Research In Partial Fulfillment of the Requirements for the Degree of Master of Science In Mathematics University of Regina By Maram Albayyadhi Regina, Saskatchewan January 2013 c Copyright 2013: Maram Albayyadhi UNIVERSITY OF REGINA FACULTY OF GRADUATE STUDIES AND RESEARCH SUPERVISORY AND EXAMINING COMMITTEE Maram Albayyadhi, candidate for the degree of Master of Science in Mathematics, has presented a thesis titled, Majorization and the Schur-Horn Theorem, in an oral examination held on December 18, 2012. The following committee members have found the thesis acceptable in form and content, and that the candidate demonstrated satisfactory knowledge of the subject material. External Examiner: Dr. Daryl Hepting, Department of Computer Science Supervisor: Dr. Martin Argerami, Department of Mathematics and Statistics Committee Member: Dr. Douglas Farenick, Department of Mathematics and Statistics Chair of Defense: Dr. Sandra Zilles, Department of Computer Science Abstract We study majorization in Rn and some of its properties. The concept of ma- jorization plays an important role in matrix analysis by producing several useful re- lationships. We find out that there is a strong relationship between majorization and doubly stochastic matrices; this relation has been perfectly described in Birkhoff's Theorem. On the other hand, majorization characterizes the connection between the eigenvalues and the diagonal elements of self adjoint matrices. This relation is summarized in the Schur-Horn Theorem. Using this theorem, we prove versions of Kadison's Carpenter's Theorem. We discuss A. Neumann's extension of the concept of majorization to infinite dimension to that provides a Schur-Horn Theorem in this context. Finally, we detail the work of W. Arveson and R.V. Kadison in proving a strict Schur-Horn Theorem for positive trace-class operators. i Acknowledgments Throughout my studying, I could have not done my project without the support of my professors whom I insist on thanking even though my words cannot adequately express my gratitude. Dr. Martin Argerami, I would like to thank you from the bottom of my heart for all the support and the guidance that you provided for me. Dr. Shaun Fallat and Dr. Remus Floricel, if it were not for your classes, I would not have learned as much about my field. I'm also honored to thank all my amazing colleagues in the math department, especially my friend Angshuman Bhattacharya. To my father Ibrahim Albayyadhi and my Mother Suad Bakkari, words cannot express my love for you. Your prayers, belief in me and encouragement are the main reasons for my success. If I kept on thanking you all of my life, I could not pay you back. To my husband, Dr. Hadi Mufti, you always make it easier for me whenever I face obstacles; you have always been the wind beneath my wings. Finally, I would like to thank the one who kept wiping my tears on the hard days and saying, \Mom ::: don't give up" { my son Yazan. ii Contents Abstract i Acknowledgments ii Table of Contents iii 1 Preliminaries 1 1.1 Majorization . 1 1.2 Doubly Stochastic Matrices . 4 1.3 Doubly Stochastic Matrices and Majorization . 6 2 The Schur-Horn Theorem in the Finite Dimensional Case 14 2.1 The Pythagorean Theorem in Finite Dimension . 14 2.2 The Schur-Horn Theorem in the Finite Dimensional Case . 20 2.3 A Pythagorean Theorem for Finite Doubly Stochastic Matrices . 24 3 The Carpenter Theorem in the Infinite Dimensional Case 27 3.1 The Subspaces K and K? both have Infinite Dimension . 28 3.2 One of the Subspaces K; K? has Finite Dimension . 45 3.3 A Pythagorean Theorem for Infinite Doubly Stochastic Matrices . 47 iii 4 A Schur-Horn Theorem in Infinite Dimensional Case 52 4.1 Majorization in Infinite Dimension . 52 4.2 Neumann's Schur-Horn Theorem . 55 4.3 A Strict Schur-Horn Theorem for Positive Trace-Class Operators . 57 5 Conclusion and Future Work 66 iv Chapter 1 Preliminaries In this chapter we provide some basic information about majorization and some of it properties, which we will use later. The material in this chapter is basic and can be found in many matrix analysis books [4, 7]. 1.1 Majorization n " " " # # # Let x = (x1; ··· ; xn) 2 R . Let x = (x1; ··· ; xn) and x = (x1; ··· ; xn) denote the vector x with it's coordinates rearranged in increasing and decreasing orders respectively. Then " # xi = xn−i+1; 1 ≤ i ≤ n: Definition 1.1. Given x; y in Rn, we say x is majorized by y, denoted by x ≺ y, 1 if k k X # X # xi ≤ yi ; 1 ≤ k ≤ n (1.1) i=1 i=1 and n n X # X # xi = yi : (1.2) i=1 i=1 Pn Example 1.2. If xi 2 [0; 1], and i=1 xi = 1, then we have 1 1 ( ; ··· : ) ≺ (x ; ··· ; x ) ≺ (1; 0; ··· ; 0): n n 1 n Proposition 1.3. If x ≺ y and y ≺ x, then there exists a permutation matrix P such that y = Px. Proof. Assume first that x; y are rearranged in decreasing order,i.e. x1 ≥ x2 ≥ · · · ≥ xn, y1 ≥ y2 · · · ≥ yn. We proceed by induction on k. For k = 1 we get x1 ≤ y1, and y1 ≤ x1 from the majorization equation (1.1), which implies x1 = y1: By induction hypothesis we assume that xi = yi, for i = 1; ··· ; k and we will show that it works up to k + 1. From ( 1.1) x1 + x2 + ··· + xk + xk+1 ≤ y1 + y2 + ··· + yk + yk+1, and by our assumption x1 + x2 + ··· + xk = y1 + y2 + ··· + yk, so if we cancel them we get xk+1 ≤ yk+1, as the roles of x and y can be reversed yk+1 ≤ xk+1, and this implies xk+1 = yk+1: In general, if we write x# and y# for the non-increasing rearrangements, there exist # # permutations P1 and P2 such that x = P1x, y = P2y. By the first part of the proof, # # −1 x = y , i.e. x = P1 P2y. Although majorization is defined through non-increasing rearrangements, it can 2 also be done by non-decreasing ones: Proposition 1.4. If x; y 2 Rn, then x ≺ y if and only if k k X " X " xi ≥ yi ; 1 ≤ k ≤ n (1.3) i=1 i=1 and n n X " X " xi = yi (1.4) i=1 i=1 Pn # Pn # Pn " Proof. For equation(1.4), we have i=1 xi = i=1 yi , since x ≺ y; but i=1 xi = Pn # Pn " Pn # Pn # Pn " Pn " Pn " i=1 xi ; so i=1 xi = i=1 xi = i=1 yi = i=1 yi ; and i=1 xi = i=1 yi : And for equation (1.3), we know that " # xi = xn−i+1; 1 ≤ i ≤ n: So k k X " X # xi = xn−i+1: i=1 i=1 Then k n n n−k X " X # X X # xi = xl = xl − xl i=1 l=n−k+1 l=1 l=1 n−k X # = tr (x) − xl l=1 n−k X # ≥ tr (y) − yl l=1 n n−k n k X X # X # X " = yl − yl = yl = yi : l=1 l=1 l=n−k+1 i=1 3 1.2 Doubly Stochastic Matrices There is a deep relation between majorization and doubly stochastic matrices; we will discuss that relation in this section. Definition 1.5. Let B = (bij) be a square matrix. We say B is doubly stochastic if : • bij ≥ 0 8i; j, Pn • i=1 bij = 1 8j, Pn • j=1 bij = 1 8i. Proposition 1.6. The set of square doubly stochastic matrices is a convex set and it is closed under multiplication and the adjoint operation. But it is not a group. Pr Proof. If t1; ··· ; tr 2 [0; 1] with j=1 tj = 1 and P1; ··· ;Pr are permutations, let Pr A = j=1 tjPj. Clearly every entry of A is non-negative. Also, n n r X X X Akl = tj(Pj)kl k=1 k=1 j=1 r n X X = tj (Pj)kl j=1 k=1 r X = tj = 1; 8l: j=1 Pn A similar computation shows that l=1 Akl = 1 for all k. If A; B are doubly stochastic matrices then AB is also doubly stochastic. Indeed, 4 the sum over the rows is n n n X X X (AB)ij = AikBkj j=1 j=1 k=1 n n X X = Aik Bkj k=1 j=1 n X = Aik = 1: k=1 And the same thing can be done for the columns, which shows that AB is also doubly stochastic. Also it is clear that if A doubly stochastic, then so is its adjoint A∗: The class of doubly stochastic matrices is not a group since not every doubly stochastic matrix is invertible; for example if we take the 2 × 2 matrix, with all its 1 entries equal to 2 , this matrix is doubly stochastic and its determinant is zero. Proposition 1.7. Every permutation matrix is doubly stochastic and is an extreme point of the convex set of all doubly stochastic matrices. Proof. A permutation matrix has exactly one entry +1 in each row and in each column and all other entries are zero. So it is doubly stochastic. Now let A = α1B + α2C, with A a permutation matrix such that α1; α2 2 (0; 1), α1 + α2 = 1, and B; C are doubly stochastic matrices. Then every entry of B; C that corresponds to the zero element aij = 0 of A must be zero; indeed, 0 = aij = α1bij + α2cij.