Chapter 6 Jordan Canonical Form

Total Page:16

File Type:pdf, Size:1020Kb

Chapter 6 Jordan Canonical Form Chapter 6 Jordan Canonical Form 175 6.1 Primary Decomposition Theorem For a general matrix, we shall now try to obtain the analogues of the ideas we have in the case of diagonalizable matrices. Diagonalizable Case General Case Minimal Polynomial: Minimal Polynomial: k k Y rj Y rj mA(λ) = (λ − λj) mA(λ) = (λ − λj) j=1 j=1 where rj = 1 for all j where 1 ≤ rj ≤ aj for all j Eigenspaces: Eigenspaces: Wj = Null Space of (A − λjI) Wj = Null Space of (A − λjI) dim(Wj) = gj and gj = aj for all j dim(Wj) = gj and gj ≤ aj for all j Generalized Eigenspaces: rj Vj = Null Space of (A − λjI) dim(Vj) = aj Decomposition: Decomposition: n n F = W1 ⊕ W2 ⊕ Wj ⊕ · · · ⊕ Wk W = W1⊕W2⊕Wj⊕· · ·⊕Wk is a subspace of F Primary Decomposition Theorem: n F = V1 ⊕ V2 ⊕ Vj ⊕ Vk n n x 2 F =) 9 unique x 2 F =) 9 unique xj 2 Wj, 1 ≤ j ≤ k such that vj 2 Vj, 1 ≤ j ≤ k such that x = x1 + x2 + ··· + xj + ··· + xk x = v1 + v2 + ··· + vj + ··· + vk 176 6.2 Analogues of Lagrange Polynomial Diagonalizable Case General Case Define Define k k Y Y ri fj(λ) = (λ − λi) Fj(λ) = (λ − λi) i=1 i=1 i6=j i6=j This is a set of coprime polynomials This is a set of coprime polynomials and and hence there exist polynomials hence there exist polynomials Gj(λ), 1 ≤ gj(λ), 1 ≤ j ≤ k such that j ≤ k such that k k X X fj(λ)gj(λ) = 1 Fj(λ)Gj(λ) = 1 j=1 j=1 The polynomials The polynomials `j(λ) = fj(λ)gj(λ); 1 ≤ j ≤ k Lj(λ) = Fj(λ)Gj(λ); 1 ≤ j ≤ k are the Lagrange Interpolation are the counterparts of the polynomials Lagrange Interpolation polynomials `1(λ) + `2(λ) + ··· + `k(λ) = 1 L1(λ) + L2(λ) + ··· + Lk(λ) = 1 λ1`1(λ)+λ2`2(λ)+···+λk`k(λ) = λ λ1L1(λ) + λ2L2(λ) + ··· + λkLk(λ) = d(λ) (a polynomial but it need not be = λ) For i 6= j the product `i(λ)`j(λ) has For i 6= j the product Li(λ)Lj(λ) has a a factor mA (λ) factor mA (λ) 6.3 Matrix Decomposition In the diagonalizable case, using the Lagrange polynomials we obtained a decomposition of the matrix A. We shall next look at the analogous situation for a general matrix. 177 Diagonalizable Case General Case Define the matrices Analogously define the matrices Aj = `j(A) for 1 ≤ j ≤ k Aj = Lj(A) for 1 ≤ j ≤ k We have Analogously we have A1 + A2 + ··· + Ak = I A1 + A2 + ··· + Ak = I since since `1(λ) + `2(λ) + ··· + `k(λ) = 1 L1(λ) + L2(λ) + ··· + Lk(λ) = 1 AiAj = 0(n×n) if i 6= j AiAj = 0(n×n) since, then `i(λ)`j(λ) has the mini- since, then Li(λ)Lj(λ) has the minimal mal polynomial as a factor polynomial as a factor 2 2 Aj = Aj for 1 ≤ j ≤ k Aj = Aj for 1 ≤ j ≤ k Range of Aj is Wj and Ajx = x for Range of Aj is Vj and Ajx = x for all all x 2 Wj x 2 Vj We have We will not have λ1A1 + λ2A2 + ··· + λkAk = A λ1A1 + λ2A2 + ··· + λkAk = A since since λ1`1(λ)+λ2`2(λ)+···+λk`k(λ) = λ λ1L1(λ) + λ2L2(λ) + ··· + Lk(λ) ; in general, may not be equal to λ. Let D = λ1A1 + λ2A2 + ··· + λkAk Then D is a diagonalizable matrix (by The- orem 5.7.1). 178 We then look at by how much A differs form D. For this purpose we define N = A − D so that A = D + N (In the case where A is a diagonalizable matrix we have N = 0(n×n) and A = D). Thus, in the general case, in addition to a diagonalizable matrix, we also have to analyse the part N = A − D. We shall do this in the next section. 6.4 Nilpotent Matrices We want to analyse the difference N = A − D (6.4.1) where D is as defined in the previous section. Using the fact that I = A1 + A2 + ··· + Ak we get A = AIn×n = AA1 + AA2 + ··· + AAk Combining this and the fact that D = λ1A1 + λ2A2 + ··· + λkAk we get A − D = (AA1 + AA2 + ··· + AAk) − (λ1A1 + λ2A2 + ··· + λkAk) = (A − λ1In×n)A1 + (A − λ2In×n)A2 + ··· + (A − λkIn×n)Ak All the Aj, being polynomials in A, commute with A, and furher, AiAj = 0(n×n). Using these facts we get r r r r (A − D) = (A − λ1In×n) A1 + (A − λ2In×n) A2 + ··· + (A − λkIn×n) Ak (6.4.2) r If r ≥ Max: fr1; r2; ··· ; rjg then (A − λjIn×n) Aj has a factor mA (A), and hence r (A − λjIn×n) Aj = 0n×n for all j if r ≥ Max:fr1; r2; ··· ; rjg (6.4.3) 179 Hence we get from (6.4.2) r r N = (A − D) = 0n×n for r ≥ Max:fr1; r2; ··· ; rjg (6.4.4) Thus the difference between A and the diagonalizable matrix D is such that a power of it vanishes. This leads us to the notion of a nilpotent matrix. We have n×n Definition 6.4.1 A matrix N 2 C is said to be a \NILPOTENT" r matrix if there exists a positive integer r such that N = 0n×n. For r a nilpotent matrix the smallest power r for which N = 0n×n is called the \order of nilpotency" of N and is denoted by γ . N Thus we have that the matrix N = A − D is a nilpotent matrix the order of nilpotency being equal to Max: fr1; r2; ··· ; rkg. We can summarise our analysis above as follows: n×n Theorem 6.4.1 Every matrix A 2 C can be expressed as the sum, A = D + N; (6.4.5) of a diagonalizable matrix D and a nilpotent matrix N We have already analysed diagonalizable matrices in the last chapter. We shall now analyse a nilpotent matrix. 6.5 Nilpotent Matrices Let N 2 n×n be a nilpotent matrix with order of nilpotency γ . This means F N that γ N N = 0n×n and (6.5.1) N r 6= 0 for any r < γ (6.5.2) n×n N From (6.5.1) we also get, N r = 0 for all r ≥ γ (6.5.3) n×n N γ From this it follows that the polynomial p(λ) = λ N is an annihilating poly- γ nomial for N. Hence the minimal polynomial must divide λ N . Thus the 180 minimal polynomial must be of the form m (λ) = λr where r ≤ γ . How- N N ever, from (6.5.2) we have that N r 6= 0 for any r < γ . Hence we get n×n N that the minimal polynomial of N is given by γ N mN (λ) = (λ) Thus we see that λ1 = 0 is the only eigenvalue of any nilpotent matrix N. Since the characteristic polynomial of N is a monic polynomial of degree n and its roots are the eigenvalues we get that the characteristic polynomial of N is given by n cN (λ) = λ Hence we have, n×n Theorem 6.5.1 If N 2 F is a nilpotent matrix with order of nilpotency as γ then its characteristic and minimal polynomials N are given by n cN (λ) = λ , and (6.5.4) γ N mN (λ) = λ (6.5.5) From the above simple properties we now obtain a useful result. Suppose A 2 n×n is both nilpotent and diagonalizable. Since it is nilpotent matrix F γ A its minimal polynomial is of the form mA = λ , where γA is its order of nilpotency. Combining this with the fact that A is diagonalizable we get that its minimal polynomial must be mA (λ) = λ. Hence mA (A) = 0n×n gives us that A = 0n×n. Thus it follows that n×n Theorem 6.5.2 The only matrix in F which is both diagonaliz- able and nilpotent is the zero matrix We shall now see an example to show that the sum and product of two nilpotent matrices need not be nilpotent. n×n Example 6.5.1 . Consider the matrices N1;N2 2 C given below: 0 1 ! N = (6.5.6) 1 0 0 0 0 ! N = (6.5.7) 2 1 0 181 It is easy to see that 2 N1 = 02×2 2 N2 = 02×2 Thus both are nilpotent matrices both with order of nilpotency as 2. How- ever, we have, 0 1 ! N + N = 1 2 1 0 This matrix has characteristic polynomial given by λ2 − 1 = 0 Since there are two eigenvalues λ1 = 1; λ2 = −1, both eigenvalues having algebraic and geometric multiplicity both as 1, the matrix is diagonalizable. Since this is not the zero matrix, by the Theorem 6.5.2, this matrix cannot be nilpotent. Thus we see that the sum N1 +N2 of the two nilpotent matrices N1 and N2 is not nilpotent. Similarly we have, 1 0 ! N N = 1 2 0 0 and this is a diagonal matrix. Again by Theorem 6.5.2 it follows that N1N2 is not nilpotent. Thus the product N1N2 of the two nilpotent matrices N1 and N2 is not nilpotent. However, we can show that the following is true: when two nilpotent matrices commute with each other, then their sum and product will also be nilpotent We shall now use this result to investigate the uniqueness of the decompo- sition of a matrix as the sum D + N of a diagonalizable matrix D and a nilpotent matrix N obtained in Section 6.3.
Recommended publications
  • MATH 2370, Practice Problems
    MATH 2370, Practice Problems Kiumars Kaveh Problem: Prove that an n × n complex matrix A is diagonalizable if and only if there is a basis consisting of eigenvectors of A. Problem: Let A : V ! W be a one-to-one linear map between two finite dimensional vector spaces V and W . Show that the dual map A0 : W 0 ! V 0 is surjective. Problem: Determine if the curve 2 2 2 f(x; y) 2 R j x + y + xy = 10g is an ellipse or hyperbola or union of two lines. Problem: Show that if a nilpotent matrix is diagonalizable then it is the zero matrix. Problem: Let P be a permutation matrix. Show that P is diagonalizable. Show that if λ is an eigenvalue of P then for some integer m > 0 we have λm = 1 (i.e. λ is an m-th root of unity). Hint: Note that P m = I for some integer m > 0. Problem: Show that if λ is an eigenvector of an orthogonal matrix A then jλj = 1. n Problem: Take a vector v 2 R and let H be the hyperplane orthogonal n n to v. Let R : R ! R be the reflection with respect to a hyperplane H. Prove that R is a diagonalizable linear map. Problem: Prove that if λ1; λ2 are distinct eigenvalues of a complex matrix A then the intersection of the generalized eigenspaces Eλ1 and Eλ2 is zero (this is part of the Spectral Theorem). 1 Problem: Let H = (hij) be a 2 × 2 Hermitian matrix. Use the Min- imax Principle to show that if λ1 ≤ λ2 are the eigenvalues of H then λ1 ≤ h11 ≤ λ2.
    [Show full text]
  • Arxiv:1508.00183V2 [Math.RA] 6 Aug 2015 Atclrb Ed Yasemigroup a by field
    A THEOREM OF KAPLANSKY REVISITED HEYDAR RADJAVI AND BAMDAD R. YAHAGHI Abstract. We present a new and simple proof of a theorem due to Kaplansky which unifies theorems of Kolchin and Levitzki on triangularizability of semigroups of matrices. We also give two different extensions of the theorem. As a consequence, we prove the counterpart of Kolchin’s Theorem for finite groups of unipotent matrices over division rings. We also show that the counterpart of Kolchin’s Theorem over division rings of characteristic zero implies that of Kaplansky’s Theorem over such division rings. 1. Introduction The purpose of this short note is three-fold. We give a simple proof of Kaplansky’s Theorem on the (simultaneous) triangularizability of semi- groups whose members all have singleton spectra. Our proof, although not independent of Kaplansky’s, avoids the deeper group theoretical aspects present in it and in other existing proofs we are aware of. Also, this proof can be adjusted to give an affirmative answer to Kolchin’s Problem for finite groups of unipotent matrices over division rings. In particular, it follows from this proof that the counterpart of Kolchin’s Theorem over division rings of characteristic zero implies that of Ka- plansky’s Theorem over such division rings. We also present extensions of Kaplansky’s Theorem in two different directions. M arXiv:1508.00183v2 [math.RA] 6 Aug 2015 Let us fix some notation. Let ∆ be a division ring and n(∆) the algebra of all n × n matrices over ∆. The division ring ∆ could in particular be a field. By a semigroup S ⊆ Mn(∆), we mean a set of matrices closed under multiplication.
    [Show full text]
  • An Effective Lie–Kolchin Theorem for Quasi-Unipotent Matrices
    Linear Algebra and its Applications 581 (2019) 304–323 Contents lists available at ScienceDirect Linear Algebra and its Applications www.elsevier.com/locate/laa An effective Lie–Kolchin Theorem for quasi-unipotent matrices Thomas Koberda a, Feng Luo b,∗, Hongbin Sun b a Department of Mathematics, University of Virginia, Charlottesville, VA 22904-4137, USA b Department of Mathematics, Rutgers University, Hill Center – Busch Campus, 110 Frelinghuysen Road, Piscataway, NJ 08854, USA a r t i c l e i n f oa b s t r a c t Article history: We establish an effective version of the classical Lie–Kolchin Received 27 April 2019 Theorem. Namely, let A, B ∈ GLm(C)be quasi-unipotent Accepted 17 July 2019 matrices such that the Jordan Canonical Form of B consists Available online 23 July 2019 of a single block, and suppose that for all k 0the Submitted by V.V. Sergeichuk matrix ABk is also quasi-unipotent. Then A and B have a A, B < C MSC: common eigenvector. In particular, GLm( )is a primary 20H20, 20F38 solvable subgroup. We give applications of this result to the secondary 20F16, 15A15 representation theory of mapping class groups of orientable surfaces. Keywords: © 2019 Elsevier Inc. All rights reserved. Lie–Kolchin theorem Unipotent matrices Solvable groups Mapping class groups * Corresponding author. E-mail addresses: [email protected] (T. Koberda), fl[email protected] (F. Luo), [email protected] (H. Sun). URLs: http://faculty.virginia.edu/Koberda/ (T. Koberda), http://sites.math.rutgers.edu/~fluo/ (F. Luo), http://sites.math.rutgers.edu/~hs735/ (H.
    [Show full text]
  • Dimensions of Nilpotent Algebras
    DIMENSIONS OF NILPOTENT ALGEBRAS AMAJOR QUALIFYING PROJECT REPORT SUBMITTED TO THE FACULTY OF WORCESTER POLYTECHNIC INSTITUTE IN PARTIAL FULFILLMENT OF THE REQUIREMENTS FOR THE DEGREE OF BACHELOR OF SCIENCE WRITTEN BY KWABENA A. ADWETEWA-BADU APPROVED BY: MARCH 14, 2021 THIS REPORT REPRESENTS THE WORK OF WPI UNDERGRADUATE STUDENTS SUBMITTED TO THE FACULTY AS EVIDENCE OF COMPLETION OF A DEGREE REQUIREMENT. 1 Abstract In this thesis, we study algebras of nilpotent matrices. In the first chapter, we give complete proofs of some well-known results on the structure of a linear operator. In particular, we prove that every nilpotent linear transformation can be represented by a strictly upper triangular matrix. In Chapter 2, we generalise this result to algebras of nilpotent matrices. We prove the famous Lie-Kolchin theorem, which states that an algebra is nilpotent if and only if it is conjugate to an algebra of strictly upper triangular matrices. In Chapter 3 we introduce a family of algebras constructed from graphs and directed graphs. We characterise the directed graphs which give nilpotent algebras. To be precise, a digraph generates an algebra of nilpotency class d if and only if it is acyclic with no paths of length ≥ d. Finally, in Chapter 4, we give Jacobson’s proof of a Theorem of Schur on the maximal dimension of a subalgebra of Mn(k) of nilpotency class 2. We relate this result to problems in external graph theory, and give lower bounds on the dimension of subalgebras of nilpotency class d in Mn(k) for every pair of integers d and n.
    [Show full text]
  • Jordan Canonical Form of a Nilpotent Matrix Math 422 Schur's Triangularization Theorem Tells Us That Every Matrix a Is Unitari
    Jordan Canonical Form of a Nilpotent Matrix Math 422 Schur’s Triangularization Theorem tells us that every matrix A is unitarily similar to an upper triangular matrix T . However, the only thing certain at this point is that the the diagonal entries of T are the eigenvalues of A. The off-diagonal entries of T seem unpredictable and out of control. Recall that the Core-Nilpotent Decomposition of a singular matrix A of index k produces a block diagonal matrix C 0 0 L ∙ ¸ similar to A in which C is non-singular, rank (C)=rank Ak , and L is nilpotent of index k.Isitpossible to simplify C and L via similarity transformations and obtain triangular matrices whose off-diagonal entries are predictable? The goal of this lecture is to do exactly this¡ ¢ for nilpotent matrices. k 1 k Let L be an n n nilpotent matrix of index k. Then L − =0and L =0. Let’s compute the eigenvalues × k k 16 k 1 k 1 k 2 of L. Suppose x =0satisfies Lx = λx; then 0=L x = L − (Lx)=L − (λx)=λL − x = λL − (Lx)= 2 k 2 6 k λ L − x = = λ x; thus λ =0is the only eigenvalue of L. ··· 1 Now if L is diagonalizable, there is an invertible matrix P and a diagonal matrix D such that P − LP = D. Since the diagonal entries of D are the eigenvalues of L, and λ =0is the only eigenvalue of L,wehave 1 D =0. Solving P − LP =0for L gives L =0. Thus a diagonalizable nilpotent matrix is the zero matrix, or equivalently, a non-zero nilpotent matrix L is not diagonalizable.
    [Show full text]
  • A Polynomial in a of the Diagonalizable and Nilpotent Parts of A
    Western Oregon University Digital Commons@WOU Honors Senior Theses/Projects Student Scholarship 6-1-2017 A Polynomial in A of the Diagonalizable and Nilpotent Parts of A Kathryn Wilson Western Oregon University Follow this and additional works at: https://digitalcommons.wou.edu/honors_theses Recommended Citation Wilson, Kathryn, "A Polynomial in A of the Diagonalizable and Nilpotent Parts of A" (2017). Honors Senior Theses/Projects. 145. https://digitalcommons.wou.edu/honors_theses/145 This Undergraduate Honors Thesis/Project is brought to you for free and open access by the Student Scholarship at Digital Commons@WOU. It has been accepted for inclusion in Honors Senior Theses/Projects by an authorized administrator of Digital Commons@WOU. For more information, please contact [email protected], [email protected], [email protected]. A Polynomial in A of the Diagonalizable and Nilpotent Parts of A By Kathryn Wilson An Honors Thesis Submitted in Partial Fulfillment of the Requirements for Graduation from the Western Oregon University Honors Program Dr. Scott Beaver, Thesis Advisor Dr. Gavin Keulks, Honor Program Director June 2017 Acknowledgements I would first like to thank my advisor, Dr. Scott Beaver. He has been an amazing resources throughout this entire process. Your support was what made this project possible. I would also like to thank Dr. Gavin Keulks for his part in running the Honors Program and allowing me the opportunity to be part of it. Further I would like to thank the WOU Mathematics department for making my education here at Western immensely fulfilling, and instilling in me the quality of thinking that will create my success outside of school.
    [Show full text]
  • 6 Sep 2016 a Note on Products of Nilpotent Matrices
    A note on Products of Nilpotent Matrices Christiaan Hattingh∗† September 7, 2016 Abstract Let F be a field. A matrix A ∈ Mn(F ) is a product of two nilpotent matrices if and only if it is singular, except if A is a nonzero nilpotent matrix of order 2×2. This result was proved independently by Sourour [6] and Laffey [4]. While these results remain true and the general strategies and principles of the proofs correct, there are certain problematic details in the original proofs which are resolved in this article. A detailed and rigorous proof of the result based on Laffey’s original proof [4] is provided, and a problematic detail in the Proposition which is the main device underpinning the proof by Sourour is resolved. 1 Introduction First I will fix some notation. A standard basis vector, which is a column vector with a one in position i and zeros elsewhere, is indicated as ei. A matrix with a one in entry (i, j) and zeros elsewhere is indicated as E(i,j). Scripted letters such as F indicate a field, Mn(F ) indicate the set of all matrices of order n×n over the field F . A block diagonal matrix with diagonal blocks A and B (in that order from left top to bottom right) is indicated as Dg[A, B]. The notation [a,b,...,k] indicates a matrix with the column vectors a,b,...,k as columns. A simple Jordan block matrix of order k, and with eigenvalue λ is indicated as Jk(λ). In this text the convention arXiv:1608.04666v3 [math.RA] 6 Sep 2016 of ones on the subdiagonal is assumed for the Jordan canonical form, unless otherwise indicated.
    [Show full text]
  • Spring 2015 Complex and Linear Algebra
    Department of Mathematics OSU Qualifying Examination Spring 2015 PART II: COMPLEX ANALYSIS and LINEAR ALGEBRA • Do any of the two problems in Part CA, use the correspondingly marked blue book and indicate on the selection sheet with your identification number those problems that you wish graded. Similarly for Part LA. • Your solutions should contain all mathematical details. Please write them up as clearly as possible. • Explicitly state any standard theorems, including hypotheses, that are necessary to justify your reasoning. • On problems with multiple parts, individual parts may be weighted differently in grading. • You have three hours to complete Part II. • When you are done with the examination, place examination blue books and selec- tion sheets back into the envelope in which the test materials came. You will hand in all materials. If you use extra examination books, be sure to place your code number on them. 1 PART CA : COMPLEX ANALYSIS QUALIFYING EXAM 1. Let f : C → C be an entire function such that f(R) ⊆ R and f(iR) ⊆ iR, where iR = {it | t ∈ R}. Show that f(−z) = −f(z) for all z ∈ C. 2. Let D = {z ∈ C | |z| < 1}, and let f : D → D be a holomorphic map. Show that for any distinct z, w ∈ D, f(z) − f(w) z − w ≤ . f(w)f(z) − 1 wz − 1 3. Find the number of roots of z7 − 4z3 − 11 = 0 contained in the annulus 1 < |z| < 2. Exam continues on next page ... 2 PART LA: LINEAR ALGEBRA QUALIFYING EXAM ∗ 1. Let A = (aij) be a nonsingular n × n matrix with entries in C, and let A = (aji) 2 2 1/2 be its complex conjugate transpose (or adjoint).
    [Show full text]
  • High-Order Automatic Differentiation of Unmodified Linear Algebra
    High-Order Automatic Differentiation of Unmodified Linear Algebra Routines via Nilpotent Matrices by Benjamin Z Dunham B.A., Carroll College, 2009 M.S., University of Colorado, 2011 A thesis submitted to the Faculty of the Graduate School of the University of Colorado in partial fulfillment of the requirements for the degree of Doctor of Philosophy Department of Aerospace Engineering Sciences 2017 This thesis entitled: High-Order Automatic Differentiation of Unmodified Linear Algebra Routines via Nilpotent Matrices written by Benjamin Z Dunham has been approved for the Department of Aerospace Engineering Sciences Prof. Kurt K. Maute Prof. Alireza Doostan Date The final copy of this thesis has been examined by the signatories, and we find that both the content and the form meet acceptable presentation standards of scholarly work in the above mentioned discipline. iii Dunham, Benjamin Z (Ph.D., Aerospace Engineering Sciences) High-Order Automatic Differentiation of Unmodified Linear Algebra Routines via Nilpotent Matrices Thesis directed by Prof. Kurt K. Maute This work presents a new automatic differentiation method, Nilpotent Matrix Differentiation (NMD), capable of propagating any order of mixed or univariate derivative through common linear algebra functions – most notably third-party sparse solvers and decomposition routines, in addition to basic matrix arithmetic operations and power series – without changing data-type or modifying code line by line; this allows differentiation across sequences of ar- bitrarily many such functions with minimal implementation effort. NMD works by enlarging the matrices and vectors passed to the routines, replacing each original scalar with a matrix block augmented by derivative data; these blocks are constructed with special sparsity structures, termed “stencils,” each designed to be isomorphic to a particular mul- tidimensional hypercomplex algebra.
    [Show full text]
  • Construction of a System of Linear Differential Equations from a Scalar
    ISSN 0081-5438, Proceedings of the Steklov Institute of Mathematics, 2010, Vol. 271, pp. 322–338. c Pleiades Publishing, Ltd., 2010. Original Russian Text c I.V. V’yugin, R.R. Gontsov, 2010, published in Trudy Matematicheskogo Instituta imeni V.A. Steklova, 2010, Vol. 271, pp. 335–351. Construction of a System of Linear Differential Equations from a Scalar Equation I. V. V’yugin a andR.R.Gontsova Received January 2010 Abstract—As is well known, given a Fuchsian differential equation, one can construct a Fuch- sian system with the same singular points and monodromy. In the present paper, this fact is extended to the case of linear differential equations with irregular singularities. DOI: 10.1134/S008154381004022X 1. INTRODUCTION The classical Riemann–Hilbert problem, i.e., the question of existence of a system dy = B(z)y, y(z) ∈ Cp, (1) dz of p linear differential equations with given Fuchsian singular points a1,...,an ∈ C and monodromy χ: π1(C \{a1,...,an},z0) → GL(p, C), (2) has a negative solution in the general case. Recall that a singularity ai of system (1) is said to be Fuchsian if the matrix differential 1-form B(z) dz has simple poles at this point. The monodromy of a system is a representation of the fundamental group of a punctured Riemann sphere in the space of nonsingular complex matrices of dimension p. Under this representation, a loop γ is mapped to a matrix Gγ such that Y (z)=Y (z)Gγ ,whereY (z) is a fundamental matrix of the system in the neighborhood of the point z0 and Y (z) is its analytic continuation along γ.
    [Show full text]
  • Math 265, Midterm 2
    Math 265, Midterm 2 Name: This exam consists of 7 pages including this front page. Ground Rules 1. No calculator is allowed. 2. Show your work for every problem unless otherwise stated. Score 1 16 2 16 3 16 4 18 5 16 6 18 Total 100 1 Notations: R denotes the set of real number; C denotes the set of complex numbers. Rm×n denotes the set of m × n-matrices with entries in R; Rn = Rn×1 denotes the set of n-column vectors; Similar notations for matrices of complex numbers C; Pn[t] denotes the set of polynomials with coefficients in R and the most degree n, that is, n n−1 Pn[t] = ff(t) = ant + an−1t + ··· + a1t + a0; ai 2 R; 8ig: 1. The following are true/false questions. You don't have to justify your an- swers. Just write down either T or F in the table below. A; B; C; X; b are always matrices here. (2 points each) (a) If a matrix A has r linearly dependent rows then it has also r linearly independent columns. (b) Let A be 4×5-matrix with rank 4 then the linear system AX = b always has infinity many solutions for any b 2 R4. (c) If A is similar to B then A and B share the same eigenvectors. (d) Let v1 and v2 be eigenvectors with eigenvalues 0 and 1 respectively then v1 and v2 must be linearly independent. (e) Let A 2 Cn×n be a complex matrix. Then v 2 Cn is an eigenvector of A if and only if the complex conjugationv ¯ is.
    [Show full text]
  • MATH 8430 Fundamental Theory of Ordinary Differential Equations Lecture Notes Julien Arino Department of Mathematics University
    MATH 8430 Fundamental Theory of Ordinary Differential Equations Lecture Notes Julien Arino Department of Mathematics University of Manitoba Fall 2006 Contents 1 General theory of ODEs 3 1.1 ODEs, IVPs, solutions . 3 1.1.1 Ordinary differential equation, initial value problem . 3 1.1.2 Solutions to an ODE . 4 1.1.3 Geometric interpretation . 8 1.2 Existence and uniqueness theorems . 9 1.2.1 Successive approximations . 9 1.2.2 Local existence and uniqueness – Proof by fixed point . 10 1.2.3 Local existence and uniqueness – Proof by successive approximations 13 1.2.4 Local existence (non Lipschitz case) . 16 1.2.5 Some examples of existence and uniqueness . 21 1.3 Continuation of solutions . 24 1.3.1 Maximal interval of existence . 27 1.3.2 Maximal and global solutions . 28 1.4 Continuous dependence on initial data, on parameters . 29 1.5 Generality of first order systems . 32 1.6 Generality of autonomous systems . 34 1.7 Suggested reading, Further problems . 34 2 Linear systems 35 2.1 Existence and uniqueness of solutions . 35 2.2 Linear systems . 36 2.2.1 The vector space of solutions . 37 2.2.2 Fundamental matrix solution . 38 2.2.3 Resolvent matrix . 41 2.2.4 Wronskian . 43 2.2.5 Autonomous linear systems . 43 2.3 Affine systems . 46 2.3.1 The space of solutions . 46 2.3.2 Construction of solutions . 46 2.3.3 Affine systems with constant coefficients . 47 2.4 Systems with periodic coefficients . 48 2.4.1 Linear systems: Floquet theory .
    [Show full text]