Linear Algebra (XXVIII)

Total Page:16

File Type:pdf, Size:1020Kb

Linear Algebra (XXVIII) Linear Algebra (XXVIII) Yijia Chen 1. Quadratic Forms Definition 1.1. Let A be an n × n symmetric matrix over K, i.e., A = aij i,j2[n] where aij = aji 0 1 x1 B . C for all i, j 2 [n]. Then for the variable vector x = @ . A xn 0 1 0 1 a11 ··· a1n x1 0 x ··· x B . .. C B . C x Ax = 1 n @ . A @ . A an1 ··· ann xn = aij · xi · xj i,Xj2[n] 2 = aii · xi + 2 aij · xi · xj. 1 i<j n iX2[n] 6X6 is a quadratic form. Definition 1.2. A quadratic form x0Ax is diagonal if the matrix A is diagonal. Definition 1.3. Let A and B be two n × n-matrices. If there exists an invertible C with B = C0AC, then B is congruent to A. Lemma 1.4. The matrix congruence is an equivalence relation. More precisely, for all n × n-matrices A, B, and C, (i) A is congruent to A, (ii) if A is congruent to B, then B is congruent to A as well, and (iii) if A is congruent to B, B is congruent to C, then A is congruent to C. Lemma 1.5. Let A and B be two n × n-matrices. If B is obtained from A by the following elementary operations, then B is congruent to A. (i) Switch the i-th and j-th rows, then switch the i-th and j-th columns, where 1 6 i < j 6 j 6 n. (ii) Multiply the i-th row by k, then multiply the i-th column by k, where i 2 [n] and k 2 K n f0g. (iii) Multiply the i-th row by k and add the result to the j-th row, then multiply the i-th column by k and add the result to the j-th column, where i, j 2 [n] with i 6= j and k 2 K. Proof: (i) Recall that switching the i-th and j-th rows of the matrix A is equivalent to multiplying A on its left by the elementary matrix Pij. Similarly the column switching corresponds to multiplying the same Pij on A’s right. Thus 0 B = PijAPij = PijAPij. 1 1 Furthermore, Pij is invertible . Thus B is congruent to A. (ii) and (iii) can be shown in exactly the same fashion. 2 Theorem 1.6. Every symmetric matrix is congruent to a diagonal matrix. We first prove a technical lemma. Lemma 1.7. Let A be an n × n symmetric nonzero matrix. Then there exists an invertible matrix C such that 0 (C AC)11 6= 0. 0 0 Recall that (C AC)11 denotes the element on the first row and the first column of the matrix C AC. Proof: If (A)11 6= 0, then we are done by taking C = In. Otherwise, assume that (A)ii 6= 0 for some 2 6 i 6 n, then we switch the first and the i-th rows, and then switch the first and the i-th columns. Thereby, we move (A)ii to the first row and the first column. The result then follows from Lemma 1.5 (i). Now we are left with the case that all diagonal elements of A are 0. Since A is symmetric and nonzero, there exist some 1 6 i < j 6 n with Aij 6= 0. We add the i-th row to the j-th row, and then add the i-th column to the j-th column. Let B the resulting matrix. Thereby (B)jj = (A)jj + (A)ij + (A)ii + (A)ji = 2 · Aij 6= 0. B is congruent to A by Lemma 1.5 (ii). So we can apply the above second case and in addition use Lemma 1.5 (iii). 2 Proof of Theorem 1.6: Assume that an n × n-matrix A is symmetric, we need to show that A is congruent to a diagonal matrix. The case for n = 1 is trivial. So let n > 2. By Lemma 1.7 and Lemma 1.5 we can assume without loss of generality 0 1 a11 a12 ··· a1n Ba21 a22 ··· a2n C A = B C with a 6= 0. B . .. C 11 @ . A an1 an2 ··· ann We multiply the first row by -a21=a11 and add it to the second row, then multiply the first column by -a21=a11 = -a12=a11 and add it to the second column. This yields a matrix of the form 0 1 a11 0 ··· a1n B 0 - ··· - C B C . B . .. C @ . A an1 - ··· ann Then we do the same to the third row and the third column, and so on. Eventually, we multiply the first row by -an1=a11 and add it to the n-th row, then multiply the first column by -an1=a11 = -a1n=a11 and add it to the n-th column. The resulting matrix has the form a 0 11 , 0 B where B is (n - 1) × (n - 1) and symmetric. 1 In fact, PijPij = In. 2 By induction hypothesis, there exists an invertible D such that D0BD is diagonal. Then 1 0 a 0 1 0 a 0 11 = 11 0 D0 0 B 0 D 0 D0BD is diagonal. Since D is invertible, 1 0 1 0 = I . 0 D 0 D-1 n 1 0 a 0 Thus is invertible as well. So by Lemma 1.4, A is congruent to the diagonal 11 . 0 D 0 D0BD 2 1.1. Sylvester’s Law of Inertia. Let A be an n × n-matrix over R. By Theorem 1.6 there exists an invertible n × n-matrix C over R such that C0AC is diagonal. Without loss of generality, we can assume that for some d1,..., dr 6= 0 0 1 d1 ··· 0 S 0 C0AC = with S = B . .. C . 0 0 @ . A 0 ··· dr We can further assume that for some p 2 [r] d1,..., dp > 0 and dp+1,..., dr < 0. In terms of quadratic form, 0 0 2 2 y (C AC)y = d1x1 + ··· drxr. By further taking p p z1 := d1y1, ··· , zp := dpyp, p p zp+1 := -dp+1y1, ··· , zr := -dryr, and zr+1 := yr+1, ··· , zr := yn, we obtain a very simple quadratic from 2 2 2 2 z1 + ··· + zp - zp+1 - ··· - zr. The question arises whether the numbers r and p are uniquely determined by the original matrix A. Theorem 1.8 (Sylvester’s Law of Inertia). Assume that an n × n symmetric matrix A over R is congruent to two diagonal matrices 0 1 0 1 Ip1 0 0 Ip2 0 0 A1 = @ 0 -Ir1-p1 0A and A1 = @ 0 -Ir2-p2 0A . 0 0 0 0 0 0 Then r1 = r2 = rank(A) and p1 = p2. We give a proof which is very different from the one in the textbook, which might offer more insights. To that end, some preparations are in order. Lemma 1.9. Let A be an m × n-matrix. Then for every invertible m × m-matrix B rank(A) = rank(BA). Similarly, for every invertible n × n-matrix C, we have rank(A) = rank(AC). 3 Proof: By the Rank-Nullity Theorem n = Rank(A) + dim N(A), and n = Rank(A) + dim N(BA). n Thus it suffices to show N(A) = N(BA), i.e., for every x 2 K , Ax = 0 () BAx = 0. The direction from left to right is trivial. For the converse, assume that Ax = b 6= 0 and Bb = 0. But this means that the column vectors of B are linearly dependent, contradicting that B is invertible, or equivalently, column rank(B) = rank(B) = n. The second property can be proved similarly by observing the row vectors of C. 2 Corollary 1.10. Any two congruent matrices have the same rank. Proof: Assume A is congruent to B, i.e., there exists an invertible C with B = C0AC. Then Lemma 1.9 implies that rank(B) = rank(C0AC) = rank(AC) = rank(A). 2 The proofs of the next two lemmas are rather routine from the definitions. Lemma 1.11. Let A be an n × n-matrix of K. Then the mapping x 7! Ax n n is an isomorphism from K to K if and only if A is invertible. Lemma 1.12. Let ' : V ! V be an isomorphism (i.e., an automorphism of V). Furthermore, let W be a subspace of V. (i) '(W) is a subspace of V. (ii) If W is finite-dimensional, then dim(W) = dim '(W). Proof of Lemma 1.8: Corollary 1.10 implies r1 = rank(A1) = rank(A) = rank(A2) = r2. Next, we show that p1 = p2. Observe that by Lemma 1.4, the two matrices A1 and A2 are congruent. Hence, there is an invertible n × n-matrix C such that 0 A1 = C A2C. (1) For i 2 [2] let 0 W := x ··· x 0 ··· 0 x ,..., x 2 ⊆ n. i 1 pi 1 pi R R In particular, for any x 2 Wi n f0g, 0 x Aix > 0, and furthermore dim(Wi) = pi. Therefore, we aim to show dim(W1) = dim(W2). 4 Consider two further subspaces 0 N := 0 ··· 0 x ··· x x ,..., x 2 ⊆ n. i pi+1 n pi+1 n R R Then for every x 2 N1 we have 0 x A1x 6 0, (2) n n and additionally dim(N1) = n - p1. Let ' : R ! R be defined by '(x) := C-1x. n By Lemma 1.11 ' is an isomorphism, and then Lemma 1.12 implies '(W2) is a subspace of R with dim '(W2) = dim(W2) = p2.
Recommended publications
  • Chapter 0 Preliminaries
    Chapter 0 Preliminaries Objective of the course Introduce basic matrix results and techniques that are useful in theory and applications. It is important to know many results, but there is no way I can tell you all of them. I would like to introduce techniques so that you can obtain the results you want. This is the standard teaching philosophy of \It is better to teach students how to fish instead of just giving them fishes.” Assumption You are familiar with • Linear equations, solution sets, elementary row operations. • Matrices, column space, row space, null space, ranks, • Determinant, eigenvalues, eigenvectors, diagonal form. • Vector spaces, basis, change of bases. • Linear transformations, range space, kernel. • Inner product, Gram Schmidt process for real matrices. • Basic operations of complex numbers. • Block matrices. Notation • Mn(F), Mm;n(F) are the set of n × n and m × n matrices over F = R or C (or a more general field). n • F is the set of column vectors of length n with entries in F. • Sometimes we write Mn;Mm;n if F = C. A1 0 • If A = 2 Mm+n(F) with A1 2 Mm(F) and A2 2 Mn(F), we write A = A1 ⊕ A2. 0 A2 • xt;At denote the transpose of a vector x and a matrix A. • For a complex matrix A, A denotes the matrix obtained from A by replacing each entry by its complex conjugate. Furthermore, A∗ = (A)t. More results and examples on block matrix multiplications. 1 1 Similarity, Jordan form, and applications 1.1 Similarity Definition 1.1.1 Two matrices A; B 2 Mn are similar if there is an invertible S 2 Mn such that S−1AS = B, equivalently, A = T −1BT with T = S−1.
    [Show full text]
  • Publications of Roger Horn
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Elsevier - Publisher Connector Linear Algebra and its Applications 424 (2007) 3–7 www.elsevier.com/locate/laa Publications of Roger Horn 1. A heuristic asymptotic formula concerning the distribution of prime numbers (with P.T. Bateman), Math. Comp. 16 (1962), 363–367 (MR26#6139). 2. Primes represented by irreducible polynomials in one variable, Theory of Numbers (with P.T. Bateman), In: Proceedings of Symposia in Pure Mathematics, vol. VIII, American Mathematical Society, Providence, Rhode Island, 1965, pp. 119–132 (MR31#1234). 3. On boundary values of a Schlicht mapping, Proc. Amer. Math. Soc. 18 (1967) 782–787 (MR36#2792). 4. On infinitely divisible matrices, kernels, and functions, Z. Wahrscheinlichkeitstheorie verw. Gebiete 8 (1967) 219–230 (MR36#925). 5. The theory of infinitely divisible matrices and kernels, Trans. Amer. Math. Soc. 136 (1969) 269–286 (MR41#9327). 6. Infinitely divisible positive definite sequences, Trans. Amer. Math. Soc. 136 (1969) 287–303 (MR58#17681). 7. On a functional equation arising in probability (with R.D. Meredith), Amer. Math. Monthly 76 (1969) 802–804 (MR40#4622). 8. On the Wronskian test for independence, Amer. Math. Monthly 77 (1970) 65–66. 9. On certain power series and sequences, J. London Math. Soc. 2 (2) (1970) 160–162 (MR41#5629). 10. On moment sequences and renewal sequences, J. Math. Anal. Appl. 31 (1970) 130–135 (MR42#6541). 11. On Fenchel’s theorem, Amer. Math. Monthly 78 (1971) 380–381 (MR44#2142). 12. Schlicht mappings and infinitely divisible kernels, Pacific J.
    [Show full text]
  • SYMPLECTIC GROUPS MATHEMATICAL SURVEYS' Number 16
    SYMPLECTIC GROUPS MATHEMATICAL SURVEYS' Number 16 SYMPLECTIC GROUPS BY O. T. O'MEARA 1978 AMERICAN MATHEMATICAL SOCIETY PROVIDENCE, RHODE ISLAND TO JEAN - my wild Irish rose Library of Congress Cataloging in Publication Data O'Meara, O. Timothy, 1928- Symplectic groups. (Mathematical surveys; no. 16) Based on lectures given at the University of Notre Dame, 1974-1975. Bibliography: p. Includes index. 1. Symplectic groups. 2. Isomorphisms (Mathematics) I. Title. 11. Series: American Mathematical Society. Mathematical surveys; no. 16. QA171.046 512'.2 78-19101 ISBN 0-8218-1516-4 AMS (MOS) subject classifications (1970). Primary 15-02, 15A33, 15A36, 20-02, 20B25, 20045, 20F55, 20H05, 20H20, 20H25; Secondary 10C30, 13-02, 20F35, 20G15, 50025,50030. Copyright © 1978 by the American Mathematical Society Printed in the United States of America AIl rights reserved except those granted to the United States Government. Otherwise, this book, or parts thereof, may not be reproduced in any form without permission of the publishers. CONTENTS Preface....................................................................................................................................... ix Prerequisites and Notation...................................................................................................... xi Chapter 1. Introduction 1.1. Alternating Spaces............................................................................................. 1 1.2. Projective Transformations................................................................................
    [Show full text]
  • Canonical Forms for Congruence of Matrices and T -Palindromic Matrix Pencils: a Tribute to H
    Noname manuscript No. (will be inserted by the editor) Canonical forms for congruence of matrices and T -palindromic matrix pencils: a tribute to H. W. Turnbull and A. C. Aitken Fernando De Ter´an the date of receipt and acceptance should be inserted later Abstract A canonical form for congruence of matrices was introduced by Turnbull and Aitken in 1932. More than 70 years later, in 2006, another canonical form for congruence has been introduced by Horn and Sergeichuk. The main purpose of this paper is to draw the attention on the pioneer work of Turnbull and Aitken and to compare both canonical forms. As an application, we also show how to recover the spectral structure of T -palindromic matrix pencils from the canonical form of matrices by Horn and Sergeichuk. Keywords matrices, canonical forms, congruence, T -palindromic matrix pencils, equivalence Mathematics Subject Classification (2000) 15A21, 15A22 1 Introduction 1.1 Congruence and similarity We consider the following two actions of the general linear group of n × n invertible matrices with complex n×n entries (denoted by GL(n; C)) on the set of n × n matrices with complex entries (denoted by C ): GL(n; ) × n×n −! n×n GL(n; ) × n×n −! n×n C C C ; and C C C (P; A) 7−! P AP −1 (P; A) 7−! P AP T : The first one is known as the action of similarity and the second one as the action of congruence. Accordingly, n×n n×n −1 two matrices A; B 2 C are said to be similar if there is a nonsingular P 2 C such that P AP = B and they are said to be congruent if there is a nonsingular P such that P AP T = B.
    [Show full text]
  • Matrix Canonical Forms
    Matrix Canonical Forms Roger Horn University of Utah ICTP School: Linear Algebra: Friday, June 26, 2009 ICTP School: Linear Algebra: Friday, June 26, 2009 1 Roger Horn (University of Utah) Matrix Canonical Forms / 22 Canonical forms for congruence and *congruence Congruence A SAST (change variables in quadratic form xT Ax) ! *Congruence A SAS (change variables in Hermitian form x Ax) ! Congruence and *congruence are simpler than similarity: no inverses; identical row and column operations for congruence (complex conjugates for *congruence). The singular and nonsingular canonical structures are fundamentally di¤erent. ICTP School: Linear Algebra: Friday, June 26, 2009 2 Roger Horn (University of Utah) Matrix Canonical Forms / 22 Example: *Congruence for Hermitian matrices Sylvester’sInertia Theorem (1852): Two Hermitian matrices are *congruent if and only if they have the same number of positive eigenvalues and the same number of negative eigenvalues (and hence also the same number of zero eigenvalues). Reformulate: Two Hermitian matrices are *congruent if and only if they have the same number of eigenvalues on each of the two open rays tei0 : t > 0 and teiπ : t > 0 in the complex plane. f g i0 f iπ g Canonical form: e In+ e In 0n0 ICTP School: Linear Algebra: Friday, June 26, 2009 3 Roger Horn (University of Utah) Matrix Canonical Forms / 22 Example: *Congruence for normal matrices Unitary *congruence: Two normal matrices are unitarily *congruent (unitarily similar!) if and only if they have the same eigenvalues. Ikramov (2001): Two normal matrices are *congruent if and only if they have the same number of eigenvalues on each open ray teiθ : t > 0 , θ [0, 2π) in the complex plane.
    [Show full text]
  • Congruence of Symmetric Matrices Over Local Rings
    CORE Metadata, citation and similar papers at core.ac.uk Provided by Elsevier - Publisher Connector Linear Algebra and its Applications 431 (2009) 1687–1690 Contents lists available at ScienceDirect Linear Algebra and its Applications journal homepage: www.elsevier.com/locate/laa Congruence of symmetric matrices over local rings ∗ Yonglin Cao a,1, Fernando Szechtman b, a Institute of Applied Mathematics, School of Sciences, Shandong University of Technology, Zibo, Shandong 255091, PR China b Department of Mathematics and Statistics, University of Regina, Saskatchewan, Canada ARTICLE INFO ABSTRACT Article history: Let R be a commutative, local, and principal ideal ring with maximal Received 25 January 2009 ideal m and residue class field F. Suppose that every element of Accepted 4 June 2009 1 + m is square. Then the problem of classifying arbitrary symmet- Available online 15 July 2009 ric matrices over R by congruence naturally reduces, and is actually Submitted by V. Sergeichuk equivalent to, the problem of classifying invertible symmetric matrices over F by congruence. AMS classification: © 2009 Elsevier Inc. All rights reserved. 15A21 15A33 Keywords: Matrix congruence Bilinear form Local ring Principal ideal ring 1. Introduction We consider the problem of classifying symmetric bilinear forms defined on a free R-module of finite rank, where R is a commutative, local and principal ideal ring with maximal ideal m = Rπ, ∗ residue field F = R/m and unit group R . This is equivalent to the problem of classifying symmetric matrices over R by congruence. For the ring of integers of a local field of characteristic not 2 the problem was studied by O’Meara [6], who in 1953 gave a complete solution provided 2 is a unit or does not ramify.
    [Show full text]
  • Matrices with Orthogonal Groups Admitting Only Determinant One
    Available online at www.sciencedirect.com Linear Algebra and its Applications 428 (2008) 796–813 www.elsevier.com/locate/laa Matrices with orthogonal groups admitting only determinant one Edward S. Coakley a,∗, Froilán M. Dopico b, Charles R. Johnson c a Department of Mathematics, Harvard University, Cambridge, MA 02138, USA b Departamento de Matemáticas, Universidad Carlos III de Madrid, Avenida de la Universidad 30, 28911 Leganés, Spain c Department of Mathematics, The College of William and Mary, Williamsburg, VA 23187-8795, USA Received 6 July 2006; accepted 13 August 2007 Available online 1 November 2007 Submitted by V. Mehrmann Abstract The K-orthogonal group of an n-by-n matrix K is defined as the set of nonsingular n-by-n matrices A satisfying ATKA = K, where the superscript T denotes transposition. These form a group under matrix multiplication. It is well-known that if K is skew-symmetric and nonsingular the determinant of every element of the K-Orthogonal group is +1, i.e., the determinant of any symplectic matrix is +1. We present necessary and sufficient conditions on a real or complex matrix K so that all elements of the K-Orthogonal group have determinant +1. These necessary and sufficient conditions can be simply stated in terms of the symmetric and skew-symmetric parts of K, denoted by Ks and Kw respectively, as follows: the determinant of every element in the K-Orthogonal group is +1 if and only if the matrix pencil Kw − λKs is regular and the matrix −1 (Kw − λ0Ks) Kw has no Jordan blocks associated to the zero eigenvalue with odd dimension, where λ0 is any number such that det(Kw − λ0Ks)/= 0.
    [Show full text]
  • CONSISTENCY and EFFICIENT SOLUTION of the SYLVESTER EQUATION for ⋆-CONGRUENCE: AX + X⋆B = C 1. Introduction. the Sylvester E
    CONSISTENCY AND EFFICIENT SOLUTION OF THE SYLVESTER EQUATION FOR ?-CONGRUENCE: AX + X?B = C ∗ y z FERNANDO DE TERÁN AND FROILÁN M. DOPICO Abstract. We consider the matrix equation AX + X?B = C , where the matrices A and B have sizes m × n and n × m, respectively, the size of the unknown X is n × m, and the operator (·)? denotes either the transpose or the conjugate transpose of a matrix. In the first part of the paper, we review necessary and sufficient conditions for the existence and uniqueness of solutions. These conditions were obtained previously by Wimmer and by Byers, Kressner, Schröder and Watkins. In this review, we generalize to fields of characteristic different from two the existence condition that Wimmer originally proved for the complex field. In the second part, we develop, in the real or complex square case m = n, an algorithm to solve the equation in O(n3) flops when the solution is unique. This algorithm is based on the generalized Schur decomposition of the matrix pencil A − λB?. The equation AX + X?B = C is connected with palindromic eigenvalue problems and, as a consequence, the square complex case has attracted recently the attention of several authors. For instance, Byers, Kressner, Schröder and Watkins have considered this equation in the study of the conditioning and in the development of structured algorithms for palindromic eigenvalue problems, while De Terán and Dopico have solved the homogeneous equations AX + X?A = 0 and studied their relationship with orbits of matrices and palindromic pencils under the action of ?-congruence. Key words.
    [Show full text]
  • More on Regular Subgroups of the Affine Group 3
    MORE ON REGULAR SUBGROUPS OF THE AFFINE GROUP M.A. PELLEGRINI AND M.C. TAMBURINI BELLANI Abstract. This paper is a new contribution to the study of regular subgroups of the affine group AGLn(F), for any field F. In particular we associate to any partition λ =6 (1n+1) of n + 1 abelian regular subgroups in such a way that different partitions define non-conjugate subgroups. Moreover, we classify the regular subgroups of certain natural types for n ≤ 4. Our classification is equivalent to the classification of split local algebras of dimension n + 1 over F. Our methods, based on classical results of linear algebra, are computer free. 1. Introduction Let F be any field. We identify the affine group AGLn(F) with the subgroup of T GLn+1(F) consisting of the matrices having (1, 0,..., 0) as first column. With this n notation, AGLn(F) acts on the right on the set = (1, v): v F of affine points. A { ∈ } Clearly, there exists an epimorphism π : AGLn(F) GLn(F) induced by the action n → of AGLn(F) on F . A subgroup (or a subset) R of AGLn(F) is called regular if it acts regularly on , namely if, for every v Fn, there exists a unique element in R A ∈ having (1, v) as first row. Thus R is regular precisely when AGLn(F)= GLn(F) R, with GLn(F) R = In , where GLn(F) denotes the stabilizer of (1, 0,..., 0). ∩ { +1} A subgroup H of AGLn(F) is indecomposable if there exists no decompositionc of Fn asc a direct sum of non-trivial π(cH)-invariant subspaces.
    [Show full text]
  • An Informal Introduction to Perturbations of Matrices Determined up to Similarity Or Congruence
    S~aoPaulo Journal of Mathematical Sciences 8, 1 (2014), 1{22 An informal introduction to perturbations of matrices determined up to similarity or congruence Lena Klimenko National Technical University of Ukraine \Kyiv Polytechnic Institute", Prospect Peremogy 37, Kiev, Ukraine. E-mail address: [email protected] Vladimir V. Sergeichuk* Institute of Mathematics, Tereshchenkivska 3, Kiev, Ukraine. E-mail address: [email protected] Abstract. The reductions of a square complex matrix A to its canon- ical forms under transformations of similarity, congruence, or *con- gruence are unstable operations: these canonical forms and reduction transformations depend discontinuously on the entries of A. We sur- vey results about their behavior under perturbations of A and about normal forms of all matrices A + E in a neighborhood of A with re- spect to similarity, congruence, or *congruence. These normal forms are called miniversal deformations of A; they are not uniquely deter- mined by A + E, but they are simple and depend continuously on the entries of E. 1. Introduction The purpose of this survey is to give an informal introduction into the theory of perturbations of a square complex matrix A determined up to transformations of similarity S−1AS, or congruence ST AS, or *congruence S∗AS, in which S is nonsingular and S∗ ∶= S¯T . 2010 Mathematics Subject Classification. 15A21, 15A63, 47A07, 47A55. Key words: similarity, congruence, *congruence, perturbations, miniversal deforma- tions, closure graphs. * Supported by the FAPESP grant 2012/18139-2. The work was done while this author was visiting the University of S~aoPaulo, whose hospitality is gratefully acknowledged.
    [Show full text]
  • Mathematics in Physical Chemistry
    Mathematics in Physical Chemistry Ali Nassimi [email protected] Chemistry Department Sharif University of Technology May 8, 2021 1/1 Aim Your most valuable asset is your learning ability. This course is a practice in learning and specially improves your deduction skills. This course provides you with tools applicable in and necessary for modeling many natural phenomena. \The fundamental laws necessary for the mathematical treatment of a large part of physics and the whole of chemistry are thus completely known, and the difficulty lies only in the fact that application of these laws leads to equations that are too complex to be solved." The first part of the course reviews Linear algebra and calculus while introducing some very useful notations. In the second part of the course, we study ordinary differential equations. End of semester objective: Become familiar with mathematical tools needed for understanding physical chemistry. 2/1 Course Evaluation Final exam 11 Bahman 3 PM 40% Midterm exam 24 Azar 9 AM 40% Quizz 10% Class presentation 10% Office hours: Due to special situation of corona pandemic office hour is not set, email for an appointment instead. 3/1 To be covered in the course Complex numbers, Vector analysis and Linear algebra Vector rotation, vector multiplication and vector derivatives. Series expansion of analytic functions Integration and some theorems from calculus Dirac delta notation and Fourier transformation Curvilinear coordinates. Matrices 4/1 To be covered in the course When we know the relation between change in dependent variable with changes in independent variable we are facing a differential equation.
    [Show full text]
  • Products of Positive Definite Symplectic Matrices
    Products of positive definite symplectic matrices by Daryl Q. Granario A dissertation submitted to the Graduate Faculty of Auburn University in partial fulfillment of the requirements for the Degree of Doctor of Philosophy Auburn, Alabama August 8, 2020 Keywords: symplectic matrix, positive definite matrix, matrix decomposition Copyright 2020 by Daryl Q. Granario Approved by Tin-Yau Tam, Committee Chair and Professor Emeritus of Mathematics Ming Liao, Professor of Mathematics Huajun Huang, Associate Professor of Mathematics Ulrich Albrecht, Department Chair and Professor of Mathematics Abstract We show that every symplectic matrix is a product of five positive definite symplectic matrices and five is the best in the sense that there are symplectic matrices which are not product of less. In Chapter 1, we provide a historical background and motivation behind the study. We highlight the important works in the subject that lead to the formulation of the problem. In Chapter 2, we present the necessary mathematical prerequisites and construct a symplectic ∗congruence canonical form for Sp(2n; C). In Chapter 3, we give the proof of the main the- orem. In Chapter 4, we discuss future research. The main results in this dissertation can be found in [18]. ii Acknowledgments I would like to express my sincere gratitude to everyone who have supported me through- out this process. To my advisor Professor Tin-Yau Tam for the continuous support of my Ph.D study and related research, for their patience, motivation, and immense knowledge. Thank you for their insightful comments and constant encouragement, and for the challenging questions that pushed me to grow as a mathematician.
    [Show full text]