High-Order Automatic Differentiation of Unmodified Linear Algebra

Total Page:16

File Type:pdf, Size:1020Kb

High-Order Automatic Differentiation of Unmodified Linear Algebra High-Order Automatic Differentiation of Unmodified Linear Algebra Routines via Nilpotent Matrices by Benjamin Z Dunham B.A., Carroll College, 2009 M.S., University of Colorado, 2011 A thesis submitted to the Faculty of the Graduate School of the University of Colorado in partial fulfillment of the requirements for the degree of Doctor of Philosophy Department of Aerospace Engineering Sciences 2017 This thesis entitled: High-Order Automatic Differentiation of Unmodified Linear Algebra Routines via Nilpotent Matrices written by Benjamin Z Dunham has been approved for the Department of Aerospace Engineering Sciences Prof. Kurt K. Maute Prof. Alireza Doostan Date The final copy of this thesis has been examined by the signatories, and we find that both the content and the form meet acceptable presentation standards of scholarly work in the above mentioned discipline. iii Dunham, Benjamin Z (Ph.D., Aerospace Engineering Sciences) High-Order Automatic Differentiation of Unmodified Linear Algebra Routines via Nilpotent Matrices Thesis directed by Prof. Kurt K. Maute This work presents a new automatic differentiation method, Nilpotent Matrix Differentiation (NMD), capable of propagating any order of mixed or univariate derivative through common linear algebra functions – most notably third-party sparse solvers and decomposition routines, in addition to basic matrix arithmetic operations and power series – without changing data-type or modifying code line by line; this allows differentiation across sequences of ar- bitrarily many such functions with minimal implementation effort. NMD works by enlarging the matrices and vectors passed to the routines, replacing each original scalar with a matrix block augmented by derivative data; these blocks are constructed with special sparsity structures, termed “stencils,” each designed to be isomorphic to a particular mul- tidimensional hypercomplex algebra. The algebras are in turn designed such that Taylor expansions of hypercomplex function evaluations are finite in length and thus exactly track derivatives without approximation error. Although this use of the method in the “forward mode” is unique in its own right, it is also possible to apply it to existing implementations of the (first-order) discrete adjoint method to find high-order derivatives with lowered cost complexity; for example, for a problem with N inputs and an adjoint solver whose cost is independent of N – i.e., O.1/ – the N × N Hessian can be found in O.N/ time, which is comparable to existing second-order adjoint methods that require far more problem-specific implementation effort. Higher derivatives are likewise less expensive 2 – e.g., a N × N × N rank-three tensor can be found in O N . Alternatively, a Hessian-vector product can be found in O.1/ time, which may open up many matrix-based simulations to a range of existing optimization or surrogate modeling approaches. As a final corollary in parallel to the NMD-adjoint hybrid method, the existing complex-step differentiation (CD) technique is also shown to be capable of finding the Hessian-vector product. All variants are implemented on a stochastic diffusion problem and compared in-depth with various cost and accuracy metrics. Contents Abstract iii Table of Contents iv List of Figures x I Motivation and Background 1 1 Overview 2 1.1 Contributions . 2 1.2 A Motivating Example . 3 1.3 Differentiation Approaches . 5 1.4 An Example of Nilpotent Matrix Differentiation . 7 1.5 Dissertation Goals and Organization . 11 2 Motivation: Numerical Uses of High-Order Derivatives 14 2.1 Optimization . 14 2.1.1 Newton’s Method . 15 2.1.2 Linear and Nonlinear Conjugate Gradient Methods . 16 2.1.3 Truncated-Newton and Hessian-Free Optimization . 18 2.2 Uncertainty Quantification and Sensitivity Analysis . 20 2.3 Surrogate Modeling . 21 v 2.4 A Selection of Higher-Order Applications . 23 3 Standard Differentiation Approaches 25 3.1 Leading Non-Automatic Tools . 26 3.1.1 Finite Differences: Easy, Approximate, Forward-Mode Differentiation . 26 3.1.2 The Discrete Adjoint Method: Intrusive, Exact, Reverse-Mode Differentiation . 29 3.1.3 Second-Order Adjoints . 34 3.1.4 Directional Derivatives and Implicit Products . 39 3.2 Standard Automatic Differentiation . 40 3.2.1 Forward-Mode AD . 42 3.2.2 Reverse-Mode AD . 44 3.2.3 Automatic Source Modification vs Operator Overloading . 45 3.3 Multivariate and High-Order AD . 47 3.3.1 Vectorization . 47 3.3.2 Naive Induction . 49 3.4 Matrix Operations and AD . 50 4 Special Algebras for Differentiation 54 4.1 Complex-Step Differentiation (CD) . 54 4.1.1 History and Derivation . 54 4.1.2 Practical Usage . 58 4.2 Complex Variants . 60 4.2.1 Multicomplex Variables . 61 4.2.2 Quasi-Complex Gradients . 63 4.2.3 Shortcomings of Imaginary Elements . 65 4.3 Dual Numbers . 68 4.4 Hyper-Dual Numbers . 71 4.5 Truncated Taylor Polynomials . 74 vi II Hypercomplex Algebras and Nilpotent Matrix Differentiation 78 5 Generalized Differentiation via Nilpotent Elements 79 5.1 Hypercomplex Numbers . 79 5.2 Tensor Products and Indexing Notation . 82 5.3 Generalized Hypercomplex Algebras for AD . 89 6 Differentiation via Matrix Representations of Algebras 92 6.1 Introduction: 2 × 2 Matrix Representations . 93 6.1.1 Complex and Dual Numbers Expressed as Matrices . 93 6.1.2 A Single-Variable Example – Derivative of a Matrix Inverse . 94 6.2 Addressing Limitations . 98 6.3 Representations of Other Existing Algebras . 99 6.3.1 Multicomplex Numbers . 99 6.3.2 Multidual Numbers . 101 6.3.3 High-Order Univariate Nilpotents . 103 6.4 Representing Aribtrary Hypercomplex Algebras with Stencils . 107 7 Hybrid Methods and Directional Derivatives 114 7.1 High-Order Forward Directional Derivatives . 114 7.2 A Nilpotent-Complex Hybrid . 117 7.3 Nilpotent and Complex Adjoint Methods . 121 III Demonstrating Nilpotent Matrix Differentiation 126 8 Differentiating a Stochastic Diffusion Model 127 8.1 Detailed Problem Description . 128 8.1.1 A Deterministic Known Solution: The Heat Equation . 130 8.1.2 Galerkin Projection and Karhunen-Loève Expansion . 131 8.1.3 Sampling the Diffusivity Field . 136 vii 8.1.4 PDE Discretization, Solution, and Quantity of Interest . 137 8.1.5 Undifferentiated Program Results . 144 8.1.6 Adjoint Formulation . 147 8.2 Experiments in Nilpotent and Complex Differentiation . 152 8.3 Results: Differentiation Accuracy . 158 8.4 Results: Differentiation Costs . 162 9 Conclusions 165 Bibliography 167 Appendix A Time and Memory Benchmarks 174 A.1 Costs vs Number of Eigenfunctions . 175 A.1.1 Original Forward Problem . 175 A.1.2 Forward Complex-Step Differentiation . 176 A.1.3 Forward Dual-Number Stencil Nilpotent Matrix Differentiation . 176 A.1.4 Forward Two-Variable, Second-Order Stencil Nilpotent Matrix Differentiation . 177 A.1.5 Original Adjoint Problem . 178 A.1.6 Adjoint Complex-Step Differentiation . 179 A.1.7 Adjoint Dual-Number Stencil Nilpotent Matrix Differentiation . 179 A.1.8 Adjoint Two-Variable, First-Order Stencil Nilpotent Matrix Differentiation . 180 A.2 Costs vs Size of System (Number of Mesh Nodes) . 181 A.2.1 Original Forward Problem . 181 A.2.2 Forward Complex-Step Differentiation . 182 A.2.3 Forward Dual-Number Stencil Nilpotent Matrix Differentiation . 183 A.2.4 Forward Two-Variable, Second-Order Stencil Nilpotent Matrix Differentiation . 184 A.2.5 Original Adjoint Problem . 185 viii A.2.6 Adjoint Complex-Step Differentiation . 186 A.2.7 Adjoint Dual-Number Stencil Nilpotent Matrix Differentiation . 187 A.2.8 Adjoint Two-Variable, First-Order Stencil Nilpotent Matrix Differentiation . 188 Figures Figure 1.1 Simplified Schematic of an Unmodified User Program . 4 1.2 Simplified Schematic of an AD-Augmented Program . 12 3.1 Finite Differencing Error vs Step-Size . 28 3.2 Vectorization Savings vs Fixed Overhead . ..
Recommended publications
  • MATH 2370, Practice Problems
    MATH 2370, Practice Problems Kiumars Kaveh Problem: Prove that an n × n complex matrix A is diagonalizable if and only if there is a basis consisting of eigenvectors of A. Problem: Let A : V ! W be a one-to-one linear map between two finite dimensional vector spaces V and W . Show that the dual map A0 : W 0 ! V 0 is surjective. Problem: Determine if the curve 2 2 2 f(x; y) 2 R j x + y + xy = 10g is an ellipse or hyperbola or union of two lines. Problem: Show that if a nilpotent matrix is diagonalizable then it is the zero matrix. Problem: Let P be a permutation matrix. Show that P is diagonalizable. Show that if λ is an eigenvalue of P then for some integer m > 0 we have λm = 1 (i.e. λ is an m-th root of unity). Hint: Note that P m = I for some integer m > 0. Problem: Show that if λ is an eigenvector of an orthogonal matrix A then jλj = 1. n Problem: Take a vector v 2 R and let H be the hyperplane orthogonal n n to v. Let R : R ! R be the reflection with respect to a hyperplane H. Prove that R is a diagonalizable linear map. Problem: Prove that if λ1; λ2 are distinct eigenvalues of a complex matrix A then the intersection of the generalized eigenspaces Eλ1 and Eλ2 is zero (this is part of the Spectral Theorem). 1 Problem: Let H = (hij) be a 2 × 2 Hermitian matrix. Use the Min- imax Principle to show that if λ1 ≤ λ2 are the eigenvalues of H then λ1 ≤ h11 ≤ λ2.
    [Show full text]
  • THREE STEPS on an OPEN ROAD Gilbert Strang This Note Describes
    Inverse Problems and Imaging doi:10.3934/ipi.2013.7.961 Volume 7, No. 3, 2013, 961{966 THREE STEPS ON AN OPEN ROAD Gilbert Strang Massachusetts Institute of Technology Cambridge, MA 02139, USA Abstract. This note describes three recent factorizations of banded invertible infinite matrices 1. If A has a banded inverse : A=BC with block{diagonal factors B and C. 2. Permutations factor into a shift times N < 2w tridiagonal permutations. 3. A = LP U with lower triangular L, permutation P , upper triangular U. We include examples and references and outlines of proofs. This note describes three small steps in the factorization of banded matrices. It is written to encourage others to go further in this direction (and related directions). At some point the problems will become deeper and more difficult, especially for doubly infinite matrices. Our main point is that matrices need not be Toeplitz or block Toeplitz for progress to be possible. An important theory is already established [2, 9, 10, 13-16] for non-Toeplitz \band-dominated operators". The Fredholm index plays a key role, and the second small step below (taken jointly with Marko Lindner) computes that index in the special case of permutation matrices. Recall that banded Toeplitz matrices lead to Laurent polynomials. If the scalars or matrices a−w; : : : ; a0; : : : ; aw lie along the diagonals, the polynomial is A(z) = P k akz and the bandwidth is w. The required index is in this case a winding number of det A(z). Factorization into A+(z)A−(z) is a classical problem solved by Plemelj [12] and Gohberg [6-7].
    [Show full text]
  • Polynomials and Hankel Matrices
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Elsevier - Publisher Connector Polynomials and Hankel Matrices Miroslav Fiedler Czechoslovak Academy of Sciences Institute of Mathematics iitnci 25 115 67 Praha 1, Czechoslovakia Submitted by V. Ptak ABSTRACT Compatibility of a Hankel n X n matrix W and a polynomial f of degree m, m < n, is defined. If m = n, compatibility means that HC’ = CfH where Cf is the companion matrix of f With a suitable generalization of Cr, this theorem is gener- alized to the case that m < n. INTRODUCTION By a Hankel matrix [S] we shall mean a square complex matrix which has, if of order n, the form ( ai +k), i, k = 0,. , n - 1. If H = (~y~+~) is a singular n X n Hankel matrix, the H-polynomial (Pi of H was defined 131 as the greatest common divisor of the determinants of all (r + 1) x (r + 1) submatrices~of the matrix where r is the rank of H. In other words, (Pi is that polynomial for which the nX(n+l)matrix I, 0 0 0 %fb) 0 i 0 0 0 1 LINEAR ALGEBRA AND ITS APPLICATIONS 66:235-248(1985) 235 ‘F’Elsevier Science Publishing Co., Inc., 1985 52 Vanderbilt Ave., New York, NY 10017 0024.3795/85/$3.30 236 MIROSLAV FIEDLER is the Smith normal form [6] of H,. It has also been shown [3] that qN is a (nonzero) polynomial of degree at most T. It is known [4] that to a nonsingular n X n Hankel matrix H = ((Y~+~)a linear pencil of polynomials of degree at most n can be assigned as follows: f(x) = fo + f,x + .
    [Show full text]
  • Nilpotent Elements Control the Structure of a Module
    Nilpotent elements control the structure of a module David Ssevviiri Department of Mathematics Makerere University, P.O BOX 7062, Kampala Uganda E-mail: [email protected], [email protected] Abstract A relationship between nilpotency and primeness in a module is investigated. Reduced modules are expressed as sums of prime modules. It is shown that presence of nilpotent module elements inhibits a module from possessing good structural properties. A general form is given of an example used in literature to distinguish: 1) completely prime modules from prime modules, 2) classical prime modules from classical completely prime modules, and 3) a module which satisfies the complete radical formula from one which is neither 2-primal nor satisfies the radical formula. Keywords: Semisimple module; Reduced module; Nil module; K¨othe conjecture; Com- pletely prime module; Prime module; and Reduced ring. MSC 2010 Mathematics Subject Classification: 16D70, 16D60, 16S90 1 Introduction Primeness and nilpotency are closely related and well studied notions for rings. We give instances that highlight this relationship. In a commutative ring, the set of all nilpotent elements coincides with the intersection of all its prime ideals - henceforth called the prime radical. A popular class of rings, called 2-primal rings (first defined in [8] and later studied in [23, 26, 27, 28] among others), is defined by requiring that in a not necessarily commutative ring, the set of all nilpotent elements coincides with the prime radical. In an arbitrary ring, Levitzki showed that the set of all strongly nilpotent elements coincides arXiv:1812.04320v1 [math.RA] 11 Dec 2018 with the intersection of all prime ideals, [29, Theorem 2.6].
    [Show full text]
  • Classifying the Representation Type of Infinitesimal Blocks of Category
    Classifying the Representation Type of Infinitesimal Blocks of Category OS by Kenyon J. Platt (Under the Direction of Brian D. Boe) Abstract Let g be a simple Lie algebra over the field C of complex numbers, with root system Φ relative to a fixed maximal toral subalgebra h. Let S be a subset of the simple roots ∆ of Φ, which determines a standard parabolic subalgebra of g. Fix an integral weight ∗ µ µ ∈ h , with singular set J ⊆ ∆. We determine when an infinitesimal block O(g,S,J) := OS of parabolic category OS is nonzero using the theory of nilpotent orbits. We extend work of Futorny-Nakano-Pollack, Br¨ustle-K¨onig-Mazorchuk, and Boe-Nakano toward classifying the representation type of the nonzero infinitesimal blocks of category OS by considering arbitrary sets S and J, and observe a strong connection between the theory of nilpotent orbits and the representation type of the infinitesimal blocks. We classify certain infinitesimal blocks of category OS including all the semisimple infinitesimal blocks in type An, and all of the infinitesimal blocks for F4 and G2. Index words: Category O; Representation type; Verma modules Classifying the Representation Type of Infinitesimal Blocks of Category OS by Kenyon J. Platt B.A., Utah State University, 1999 M.S., Utah State University, 2001 M.A, The University of Georgia, 2006 A Thesis Submitted to the Graduate Faculty of The University of Georgia in Partial Fulfillment of the Requirements for the Degree Doctor of Philosophy Athens, Georgia 2008 c 2008 Kenyon J. Platt All Rights Reserved Classifying the Representation Type of Infinitesimal Blocks of Category OS by Kenyon J.
    [Show full text]
  • (Hessenberg) Eigenvalue-Eigenmatrix Relations∗
    (HESSENBERG) EIGENVALUE-EIGENMATRIX RELATIONS∗ JENS-PETER M. ZEMKE† Abstract. Explicit relations between eigenvalues, eigenmatrix entries and matrix elements are derived. First, a general, theoretical result based on the Taylor expansion of the adjugate of zI − A on the one hand and explicit knowledge of the Jordan decomposition on the other hand is proven. This result forms the basis for several, more practical and enlightening results tailored to non-derogatory, diagonalizable and normal matrices, respectively. Finally, inherent properties of (upper) Hessenberg, resp. tridiagonal matrix structure are utilized to construct computable relations between eigenvalues, eigenvector components, eigenvalues of principal submatrices and products of lower diagonal elements. Key words. Algebraic eigenvalue problem, eigenvalue-eigenmatrix relations, Jordan normal form, adjugate, principal submatrices, Hessenberg matrices, eigenvector components AMS subject classifications. 15A18 (primary), 15A24, 15A15, 15A57 1. Introduction. Eigenvalues and eigenvectors are defined using the relations Av = vλ and V −1AV = J. (1.1) We speak of a partial eigenvalue problem, when for a given matrix A ∈ Cn×n we seek scalar λ ∈ C and a corresponding nonzero vector v ∈ Cn. The scalar λ is called the eigenvalue and the corresponding vector v is called the eigenvector. We speak of the full or algebraic eigenvalue problem, when for a given matrix A ∈ Cn×n we seek its Jordan normal form J ∈ Cn×n and a corresponding (not necessarily unique) eigenmatrix V ∈ Cn×n. Apart from these constitutional relations, for some classes of structured matrices several more intriguing relations between components of eigenvectors, matrix entries and eigenvalues are known. For example, consider the so-called Jacobi matrices.
    [Show full text]
  • Hyperbolicity of Hermitian Forms Over Biquaternion Algebras
    HYPERBOLICITY OF HERMITIAN FORMS OVER BIQUATERNION ALGEBRAS NIKITA A. KARPENKO Abstract. We show that a non-hyperbolic hermitian form over a biquaternion algebra over a field of characteristic 6= 2 remains non-hyperbolic over a generic splitting field of the algebra. Contents 1. Introduction 1 2. Notation 2 3. Krull-Schmidt principle 3 4. Splitting off a motivic summand 5 5. Rost correspondences 7 6. Motivic decompositions of some isotropic varieties 12 7. Proof of Main Theorem 14 References 16 1. Introduction Throughout this note (besides of x3 and x4) F is a field of characteristic 6= 2. The basic reference for the staff related to involutions on central simple algebras is [12].p The degree deg A of a (finite dimensional) central simple F -algebra A is the integer dimF A; the index ind A of A is the degree of a central division algebra Brauer-equivalent to A. Conjecture 1.1. Let A be a central simple F -algebra endowed with an orthogonal invo- lution σ. If σ becomes hyperbolic over the function field of the Severi-Brauer variety of A, then σ is hyperbolic (over F ). In a stronger version of Conjecture 1.1, each of two words \hyperbolic" is replaced by \isotropic", cf. [10, Conjecture 5.2]. Here is the complete list of indices ind A and coindices coind A = deg A= ind A of A for which Conjecture 1.1 is known (over an arbitrary field of characteristic 6= 2), given in the chronological order: • ind A = 1 | trivial; Date: January 2008. Key words and phrases.
    [Show full text]
  • Arxiv:1508.00183V2 [Math.RA] 6 Aug 2015 Atclrb Ed Yasemigroup a by field
    A THEOREM OF KAPLANSKY REVISITED HEYDAR RADJAVI AND BAMDAD R. YAHAGHI Abstract. We present a new and simple proof of a theorem due to Kaplansky which unifies theorems of Kolchin and Levitzki on triangularizability of semigroups of matrices. We also give two different extensions of the theorem. As a consequence, we prove the counterpart of Kolchin’s Theorem for finite groups of unipotent matrices over division rings. We also show that the counterpart of Kolchin’s Theorem over division rings of characteristic zero implies that of Kaplansky’s Theorem over such division rings. 1. Introduction The purpose of this short note is three-fold. We give a simple proof of Kaplansky’s Theorem on the (simultaneous) triangularizability of semi- groups whose members all have singleton spectra. Our proof, although not independent of Kaplansky’s, avoids the deeper group theoretical aspects present in it and in other existing proofs we are aware of. Also, this proof can be adjusted to give an affirmative answer to Kolchin’s Problem for finite groups of unipotent matrices over division rings. In particular, it follows from this proof that the counterpart of Kolchin’s Theorem over division rings of characteristic zero implies that of Ka- plansky’s Theorem over such division rings. We also present extensions of Kaplansky’s Theorem in two different directions. M arXiv:1508.00183v2 [math.RA] 6 Aug 2015 Let us fix some notation. Let ∆ be a division ring and n(∆) the algebra of all n × n matrices over ∆. The division ring ∆ could in particular be a field. By a semigroup S ⊆ Mn(∆), we mean a set of matrices closed under multiplication.
    [Show full text]
  • Chapter 6 Jordan Canonical Form
    Chapter 6 Jordan Canonical Form 175 6.1 Primary Decomposition Theorem For a general matrix, we shall now try to obtain the analogues of the ideas we have in the case of diagonalizable matrices. Diagonalizable Case General Case Minimal Polynomial: Minimal Polynomial: k k Y rj Y rj mA(λ) = (λ − λj) mA(λ) = (λ − λj) j=1 j=1 where rj = 1 for all j where 1 ≤ rj ≤ aj for all j Eigenspaces: Eigenspaces: Wj = Null Space of (A − λjI) Wj = Null Space of (A − λjI) dim(Wj) = gj and gj = aj for all j dim(Wj) = gj and gj ≤ aj for all j Generalized Eigenspaces: rj Vj = Null Space of (A − λjI) dim(Vj) = aj Decomposition: Decomposition: n n F = W1 ⊕ W2 ⊕ Wj ⊕ · · · ⊕ Wk W = W1⊕W2⊕Wj⊕· · ·⊕Wk is a subspace of F Primary Decomposition Theorem: n F = V1 ⊕ V2 ⊕ Vj ⊕ Vk n n x 2 F =) 9 unique x 2 F =) 9 unique xj 2 Wj, 1 ≤ j ≤ k such that vj 2 Vj, 1 ≤ j ≤ k such that x = x1 + x2 + ··· + xj + ··· + xk x = v1 + v2 + ··· + vj + ··· + vk 176 6.2 Analogues of Lagrange Polynomial Diagonalizable Case General Case Define Define k k Y Y ri fj(λ) = (λ − λi) Fj(λ) = (λ − λi) i=1 i=1 i6=j i6=j This is a set of coprime polynomials This is a set of coprime polynomials and and hence there exist polynomials hence there exist polynomials Gj(λ), 1 ≤ gj(λ), 1 ≤ j ≤ k such that j ≤ k such that k k X X fj(λ)gj(λ) = 1 Fj(λ)Gj(λ) = 1 j=1 j=1 The polynomials The polynomials `j(λ) = fj(λ)gj(λ); 1 ≤ j ≤ k Lj(λ) = Fj(λ)Gj(λ); 1 ≤ j ≤ k are the Lagrange Interpolation are the counterparts of the polynomials Lagrange Interpolation polynomials `1(λ) + `2(λ) + ··· + `k(λ) = 1 L1(λ) + L2(λ) + ··· + Lk(λ) = 1 λ1`1(λ)+λ2`2(λ)+···+λk`k(λ) = λ λ1L1(λ) + λ2L2(λ) + ··· + λkLk(λ) = d(λ) (a polynomial but it need not be = λ) For i 6= j the product `i(λ)`j(λ) has For i 6= j the product Li(λ)Lj(λ) has a a factor mA (λ) factor mA (λ) 6.3 Matrix Decomposition In the diagonalizable case, using the Lagrange polynomials we obtained a decomposition of the matrix A.
    [Show full text]
  • Mathematische Annalen Digital Object Identifier (DOI) 10.1007/S002080100153
    Math. Ann. (2001) Mathematische Annalen Digital Object Identifier (DOI) 10.1007/s002080100153 Pfaff τ-functions M. Adler · T. Shiota · P. van Moerbeke Received: 20 September 1999 / Published online: ♣ – © Springer-Verlag 2001 Consider the two-dimensional Toda lattice, with certain skew-symmetric initial condition, which is preserved along the locus s =−t of the space of time variables. Restricting the solution to s =−t, we obtain another hierarchy called Pfaff lattice, which has its own tau function, being equal to the square root of the restriction of 2D-Toda tau function. We study its bilinear and Fay identities, W and Virasoro symmetries, relation to symmetric and symplectic matrix integrals and quasiperiodic solutions. 0. Introduction Consider the set of equations ∂m∞ n ∂m∞ n = Λ m∞ , =−m∞(Λ ) ,n= 1, 2,..., (0.1) ∂tn ∂sn on infinite matrices m∞ = m∞(t, s) = (µi,j (t, s))0≤i,j<∞ , M. Adler∗ Department of Mathematics, Brandeis University, Waltham, MA 02454–9110, USA (e-mail: [email protected]) T. Shiota∗∗ Department of Mathematics, Kyoto University, Kyoto 606–8502, Japan (e-mail: [email protected]) P. van Moerbeke∗∗∗ Department of Mathematics, Universit´e de Louvain, 1348 Louvain-la-Neuve, Belgium Brandeis University, Waltham, MA 02454–9110, USA (e-mail: [email protected] and @math.brandeis.edu) ∗ The support of a National Science Foundation grant # DMS-98-4-50790 is gratefully ac- knowledged. ∗∗ The support of Japanese ministry of education’s grant-in-aid for scientific research, and the hospitality of the University of Louvain and Brandeis University are gratefully acknowledged.
    [Show full text]
  • Elementary Invariants for Centralizers of Nilpotent Matrices
    J. Aust. Math. Soc. 86 (2009), 1–15 doi:10.1017/S1446788708000608 ELEMENTARY INVARIANTS FOR CENTRALIZERS OF NILPOTENT MATRICES JONATHAN BROWN and JONATHAN BRUNDAN ˛ (Received 7 March 2007; accepted 29 March 2007) Communicated by J. Du Abstract We construct an explicit set of algebraically independent generators for the center of the universal enveloping algebra of the centralizer of a nilpotent matrix in the general linear Lie algebra over a field of characteristic zero. In particular, this gives a new proof of the freeness of the center, a result first proved by Panyushev, Premet and Yakimova. 2000 Mathematics subject classification: primary 17B35. Keywords and phrases: nilpotent matrices, centralizers, symmetric invariants. 1. Introduction Let λ D (λ1; : : : ; λn/ be a composition of N such that either λ1 ≥ · · · ≥ λn or λ1 ≤ · · · ≤ λn. Let g be the Lie algebra glN .F/, where F is an algebraically closed field of characteristic zero. Let e 2 g be the nilpotent matrix consisting of Jordan blocks of sizes λ1; : : : ; λn in order down the diagonal, and write ge for the centralizer of e in g. g Panyushev et al. [PPY] have recently proved that S.ge/ e , the algebra of invariants for the adjoint action of ge on the symmetric algebra S.ge/, is a free polynomial algebra on N generators. Moreover, viewing S.ge/ as a graded algebra as usual so that ge is concentrated in degree one, they show that if x1;:::; xN are homogeneous generators g for S.ge/ e of degrees d1 ≤ · · · ≤ dN , then the sequence .d1;:::; dN / of invariant degrees is equal to λ1 1’s λ2 2’s λn n’s z }| { z }| { z }| { .1;:::; 1; 2;:::; 2;:::; n;:::; n/ if λ1 ≥ · · · ≥ λn, .1;:::; 1; 2;:::; 2;:::; n;:::; n/ if λ1 ≤ · · · ≤ λn.
    [Show full text]
  • Commutative Algebra
    Commutative Algebra Andrew Kobin Spring 2016 / 2019 Contents Contents Contents 1 Preliminaries 1 1.1 Radicals . .1 1.2 Nakayama's Lemma and Consequences . .4 1.3 Localization . .5 1.4 Transcendence Degree . 10 2 Integral Dependence 14 2.1 Integral Extensions of Rings . 14 2.2 Integrality and Field Extensions . 18 2.3 Integrality, Ideals and Localization . 21 2.4 Normalization . 28 2.5 Valuation Rings . 32 2.6 Dimension and Transcendence Degree . 33 3 Noetherian and Artinian Rings 37 3.1 Ascending and Descending Chains . 37 3.2 Composition Series . 40 3.3 Noetherian Rings . 42 3.4 Primary Decomposition . 46 3.5 Artinian Rings . 53 3.6 Associated Primes . 56 4 Discrete Valuations and Dedekind Domains 60 4.1 Discrete Valuation Rings . 60 4.2 Dedekind Domains . 64 4.3 Fractional and Invertible Ideals . 65 4.4 The Class Group . 70 4.5 Dedekind Domains in Extensions . 72 5 Completion and Filtration 76 5.1 Topological Abelian Groups and Completion . 76 5.2 Inverse Limits . 78 5.3 Topological Rings and Module Filtrations . 82 5.4 Graded Rings and Modules . 84 6 Dimension Theory 89 6.1 Hilbert Functions . 89 6.2 Local Noetherian Rings . 94 6.3 Complete Local Rings . 98 7 Singularities 106 7.1 Derived Functors . 106 7.2 Regular Sequences and the Koszul Complex . 109 7.3 Projective Dimension . 114 i Contents Contents 7.4 Depth and Cohen-Macauley Rings . 118 7.5 Gorenstein Rings . 127 8 Algebraic Geometry 133 8.1 Affine Algebraic Varieties . 133 8.2 Morphisms of Affine Varieties . 142 8.3 Sheaves of Functions .
    [Show full text]