A Numerical Method for the Inverse Stochastic

Total Page:16

File Type:pdf, Size:1020Kb

A Numerical Method for the Inverse Stochastic A NUMERICAL METHOD FOR THE INVERSE STOCHASTIC SPECTRUM PROBLEM y MOODY T CHU AND QUANLIN GUO Abstract Inverse sto chastic sp ectrum problem involves the construction of a sto chastic matrix with a prescrib ed sp ectrum A dierential equation aimed to bring forth the steep est descent ow in reducing the distance b etween isosp ectral matrices and nonnegative matrices represented in terms of some general co ordinates is describ ed The ow is further characterized by an analytic singular value decomp osition to maintain the numerical stability and to monitor the proximity to singularity This ow approachcan be used to design Markovchains with sp ecied structure Applications are demonstrated byn umerical examples Key words Sto chastic Matrix Least Squares Steep est Descent Isosp ectral Flow Structured Markovchain Analytic Singular Value Flow AMSMOS sub ject classications F H Intro duction Inverse eigenvalue problems concern the reconstruction of ma trices from prescrib ed sp ectral data The sp ectral data mayinvolve complete or partial information of eigenvalues or eigenvectors Generally a problem without any restric tions on the matrix is of little interest In order that the inverse eigenvalue problem b e meaningful it is often necessary to restrict the construction to sp ecial classes of matri ces such as symmetric To eplitz matrices or matrices with other sp ecial structures In this pap er we limit our attention to the so called sto chastic matrices ie matrices with nonnegative elements where all its row sums are equal to one We prop ose a numer ical pro cedure for the construction of a sto chastic matrix so that its sp ectrum agrees with a prescrib ed set of complex values If the set of prescrib ed values turns out to be infeasible the metho d pro duces a best approximation in the sense of least squares To our knowledge this inverse eigenvalue problem for sto chastic matrices has not b een studied extensively probably due to its diculty as we shall discuss below Neverthe less for a variety of physical problems that can be describ ed in the context of Markov chains an understanding of the inverse eigenvalue problem for sto chastic matrices and a capacity to solve the problem would make it p ossible to construct a system from its natural frequencies The metho d prop osed in the pap er app ears to be the rst attemp at tackling this problem numerically with some success Our technique can also be applied as a numerical way to solve the long standing inverse eigenvalue problems for nonnegative matrices Asso ciated with every inverse eigenvalue problem are two fundamental questions issue on solvability and the practical issue on computability The the theoretic ma jor eort in solvability has been to determine a necessary or a sucient condition under which an inverse eigenvalue problem has a solution whereas the main concern Department of Mathematics North Carolina State University Raleigh NC This researchwas supp orted in part by the National Science Foundation under grant DMS y Department of Mathematics North Carolina State University Raleigh NC 1 0.8 0.6 0.4 0.2 0 −0.2 −0.4 −0.6 −0.8 −1 −1 −0.5 0 0.5 1 Fig by the KarpelevicTheorem in computability has been to develop an algorithm by which knowing a priori that the given sp ectral data are feasible a matrix can be constructed numerically Both questions are dicult and challenging Searching through the literature wehavefound only a handful of inverse eigenvalue problems that have b een completely understo o d or solved The fo cus of this pap er is on the computability for sto chastic matrices For sto chastic matrices the inverse eigenvalue problem is particularly dicult as result on existence by Karp elevic can be seen from the involvement in the best known Karp elevic completely characterized the set of points in the complex n plane that are eigenvalues of sto chastic n n matrices In particular the region is n symmetric ab out the real axis It is contained within the unit circle and its intersections ab with the unit circle are p oints z e where a and b run over all integers satisfying a b n The b oundary of consists of these intersection p oints and of n curvilinear arcs connecting them in circular order These arcs are characterized by sp ecic parametric equations whose formulas can be found in For example a complex number is an eigenvalue for a sto chastic matrix if and only if it b elongs to a region such as the one shown in Figure Complicated though it may seem the Karp elevic theorem characterizes only one complex value a time and do es alues of the not provide further insights into when two or more points in are eigenv n same sto chastic matrix Minc distinctively called the problem we are considering where the entire sp ectrum is given the inverse spectrum problem It is known that the inverse eigenvalue problem for nonnegative matrices is virtually equivalent to that for sto chastic matrices For example a complex nonzero number is an eigenvalue of a nonnegative matrix with a p ositive maximal eigenvalue r if and only if r is an eigenvalue of a sto chastic matrix Our problem is much more complicated b ecause it involves the entire sp ectrum Fortunately based on the following theorem we can pro ceed our computation once a nonnegative matrix is found Theorem If A is a nonnegative matrix with positive maximal eigenvalue r and a positive maximal eigenvector x then D r AD is a stochastic matrix where D diagfx x g n Wethus should turn our attention to the inverse eigenvalue or sp ectrum problems for nonnegative matrices a sub ject that has received considerable interest in the litera ture Some necessary and a few sucient conditions on whether a given set of complex numbers could be the sp ectrum of a nonnegative matrix can be found for example in and the references contained therein Yet numerical metho ds for constructing such a matrix even if the sp ectrum is feasible still need to b e develop ed Some discussion can b e found in Regardless of all the eorts the inverse eigenvalue problem for nonnegative matrices has not been completely resolved to this date In an earlier pap er the rst author has develop ed an algorithm that can con struct symmetric nonnegative matrices with prescrib ed sp ectra by means of dierential equations Symmetry was needed there b ecause the techniques by then were for ows in the group of orthogonal matrices only Up on realizing the existence of an analytic singular value decomp osition ASVD for a real analytic path of matrices we are able to advance the techniques in to general matrices in this pap er This pap er is organized as follows We reformulate the inverse sto chastic sp ectrum problem as that of nding the shortest distance between isosp ectral matrices and non negative matrices In x we intro duce a general co ordinate system to describ e these twotyp es of matrices and discuss how this setting naturally leads to a steep est descent ow This approach generalizes what has b een done b efore but requires the inversion of matrices that is potentially dangerous In x we argue that the steep est descent o w is in fact analytic and hence an analytic singular value decomp osition exists We therefore are able to describ e the ow by a more stable vector eld We illustrate the application of this dierential equation to the inverse sp ectrum problem by numerical examples in x Basic Formulation The given sp ectrum f g may b e complexvalued n It is not dicult to create a simple say tridiagonal realvalued matrix carrying the same sp ectrum For multiple eigenvalues one should also consider the p ossible real valued Jordan canonical form dep ending on the geometric multiplicity Matrices in the set nn M fP P jP R is nonsingularg obviously are isosp ectral to Let nn n fB B jB R g R denote the cone of all nonnegative matrices where means the Hadamard pro duct of n matrices Our basic idea is to nd the intersection of M and R Such an intersection if exists results in a nonnegative matrix isosp ectral to Furthermore if the condition in Theorem holds ie if the eigenvector corresp onding to the p ositive maximal eigenvalue is p ositive then we will have solved the inverse sp ectrum problem for sto chastic matrices by a diagonal similarity transformation The diculty as we have p ointed earlier is the lack of means to determine if the given sp ectrum is feasible An arbitrarily given set of values even if for all i may not be the n i n sp ectrum of any nonnegative matrix In this case it is reasonable to ask for only the b est p ossible approximation To handle b oth problems at the same time we reformulate the inverse sp ectrum problem as that of nding the shortest distance between M n and R minimize F P R kP P R R k where kk represents the Frob enius matrix norm Obviously if is feasible then F P R for some suitable P and R Note that the variable P in resides in the nn op en set of nonsingular matrices whereas R is simple a general matrix in R The optimization in sub jects to no other signicant constraint Since the optimization is over unb ounded op en domain it is p ossible that the minimum do es not exist We shall comment more on this point later The Frechet derivativeofF at P R acting on H K is calculated as follows F P RH K hP P R R H P P P HP K R R K i T T T T T T R R P Hi hP P R R P P P P P hP P R R R K i where h i denotes the Frob enius inner pro duct of two matrices Dene for abbrevia tion M P P P P R M P R R The norm of P R represents
Recommended publications
  • Isospectralization, Or How to Hear Shape, Style, and Correspondence
    Isospectralization, or how to hear shape, style, and correspondence Luca Cosmo Mikhail Panine Arianna Rampini University of Venice Ecole´ Polytechnique Sapienza University of Rome [email protected] [email protected] [email protected] Maks Ovsjanikov Michael M. Bronstein Emanuele Rodola` Ecole´ Polytechnique Imperial College London / USI Sapienza University of Rome [email protected] [email protected] [email protected] Initialization Reconstruction Abstract opt. target The question whether one can recover the shape of a init. geometric object from its Laplacian spectrum (‘hear the shape of the drum’) is a classical problem in spectral ge- ometry with a broad range of implications and applica- tions. While theoretically the answer to this question is neg- ative (there exist examples of iso-spectral but non-isometric manifolds), little is known about the practical possibility of using the spectrum for shape reconstruction and opti- mization. In this paper, we introduce a numerical proce- dure called isospectralization, consisting of deforming one shape to make its Laplacian spectrum match that of an- other. We implement the isospectralization procedure using modern differentiable programming techniques and exem- plify its applications in some of the classical and notori- Without isospectralization With isospectralization ously hard problems in geometry processing, computer vi- sion, and graphics such as shape reconstruction, pose and Figure 1. Top row: Mickey-from-spectrum: we recover the shape of Mickey Mouse from its first 20 Laplacian eigenvalues (shown style transfer, and dense deformable correspondence. in red in the leftmost plot) by deforming an initial ellipsoid shape; the ground-truth target embedding is shown as a red outline on top of our reconstruction.
    [Show full text]
  • Improving the Accuracy of Computed Eigenvalues and Eigenvectors*
    SIAM J. NUMER. ANAL. 1983 Society for Industrial and Applied Mathematics Vol. 20, No. 1, February 1983 0036-1429/83/2001-0002 $01.25/0 IMPROVING THE ACCURACY OF COMPUTED EIGENVALUES AND EIGENVECTORS* J. J. DONGARRA,t C. B. MOLER AND J. H. WILKINSON Abstract. This paper describes and analyzes several variants of a computational method for improving the numerical accuracy of, and for obtaining numerical bounds on, matrix eigenvalues and eigenvectors. The method, which is essentially a numerically stable implementation of Newton's method, may be used to "fine tune" the results obtained from standard subroutines such as those in EISPACK [Lecture Notes in Computer Science 6, 51, Springer-Verlag, Berlin, 1976, 1977]. Extended precision arithmetic is required in the computation of certain residuals. Introduction. The calculation of an eigenvalue , and the corresponding eigenvec- tor x (here after referred to as an eigenpair) of a matrix A involves the solution of the nonlinear system of equations (A AI)x O. Starting from an approximation h and , a sequence of iterates may be determined using Newton's method or variants of it. The conditions on and guaranteeing convergence have been treated extensively in the literature. For a particularly lucid account the reader is referred to the book by Rail [3]. In a recent paper Wilkinson [7] describes an algorithm for determining error bounds for a computed eigenpair based on these mathematical concepts. Considerations of numerical stability were an essential feature of that paper and indeed were its main raison d'etre. In general this algorithm provides an improved eigenpair and error bounds for it; unless the eigenpair is very ill conditioned the improved eigenpair is usually correct to the precision of the computation used in the main body of the algorithm.
    [Show full text]
  • Numerical Analysis Notes for Math 575A
    Numerical Analysis Notes for Math 575A William G. Faris Program in Applied Mathematics University of Arizona Fall 1992 Contents 1 Nonlinear equations 5 1.1 Introduction . 5 1.2 Bisection . 5 1.3 Iteration . 9 1.3.1 First order convergence . 9 1.3.2 Second order convergence . 11 1.4 Some C notations . 12 1.4.1 Introduction . 12 1.4.2 Types . 13 1.4.3 Declarations . 14 1.4.4 Expressions . 14 1.4.5 Statements . 16 1.4.6 Function definitions . 17 2 Linear Systems 19 2.1 Shears . 19 2.2 Reflections . 24 2.3 Vectors and matrices in C . 27 2.3.1 Pointers in C . 27 2.3.2 Pointer Expressions . 28 3 Eigenvalues 31 3.1 Introduction . 31 3.2 Similarity . 31 3.3 Orthogonal similarity . 34 3.3.1 Symmetric matrices . 34 3.3.2 Singular values . 34 3.3.3 The Schur decomposition . 35 3.4 Vector and matrix norms . 37 3.4.1 Vector norms . 37 1 2 CONTENTS 3.4.2 Associated matrix norms . 37 3.4.3 Singular value norms . 38 3.4.4 Eigenvalues and norms . 39 3.4.5 Condition number . 39 3.5 Stability . 40 3.5.1 Inverses . 40 3.5.2 Iteration . 40 3.5.3 Eigenvalue location . 41 3.6 Power method . 43 3.7 Inverse power method . 44 3.8 Power method for subspaces . 44 3.9 QR method . 46 3.10 Finding eigenvalues . 47 4 Nonlinear systems 49 4.1 Introduction . 49 4.2 Degree . 51 4.2.1 Brouwer fixed point theorem .
    [Show full text]
  • Numerically Stable Generation of Correlation Matrices and Their Factors ∗
    BIT 0006-3835/00/4004-0640 $15.00 2000, Vol. 40, No. 4, pp. 640–651 c Swets & Zeitlinger NUMERICALLY STABLE GENERATION OF CORRELATION MATRICES AND THEIR FACTORS ∗ ‡ PHILIP I. DAVIES† and NICHOLAS J. HIGHAM Department of Mathematics, University of Manchester, Manchester, M13 9PL, England email: [email protected], [email protected] Abstract. Correlation matrices—symmetric positive semidefinite matrices with unit diagonal— are important in statistics and in numerical linear algebra. For simulation and testing it is desirable to be able to generate random correlation matrices with specified eigen- values (which must be nonnegative and sum to the dimension of the matrix). A popular algorithm of Bendel and Mickey takes a matrix having the specified eigenvalues and uses a finite sequence of Givens rotations to introduce 1s on the diagonal. We give im- proved formulae for computing the rotations and prove that the resulting algorithm is numerically stable. We show by example that the formulae originally proposed, which are used in certain existing Fortran implementations, can lead to serious instability. We also show how to modify the algorithm to generate a rectangular matrix with columns of unit 2-norm. Such a matrix represents a correlation matrix in factored form, which can be preferable to representing the matrix itself, for example when the correlation matrix is nearly singular to working precision. Key words: Random correlation matrix, Bendel–Mickey algorithm, eigenvalues, sin- gular value decomposition, test matrices, forward error bounds, relative error bounds, IMSL, NAG Library, Jacobi method. AMS subject classification: 65F15. 1 Introduction. An important class of symmetric positive semidefinite matrices is those with unit diagonal.
    [Show full text]
  • Model Problems in Numerical Stability Theory for Initial Value Problems*
    SIAM REVIEW () 1994 Society for Industrial and Applied Mathematics Vol. 36, No. 2, pp. 226-257, June 1994 004 MODEL PROBLEMS IN NUMERICAL STABILITY THEORY FOR INITIAL VALUE PROBLEMS* A.M. STUART AND A.R. HUMPHRIES Abstract. In the past numerical stability theory for initial value problems in ordinary differential equations has been dominated by the study of problems with simple dynamics; this has been motivated by the need to study error propagation mechanisms in stiff problems, a question modeled effectively by contractive linear or nonlinear problems. While this has resulted in a coherent and self-contained body of knowledge, it has never been entirely clear to what extent this theory is relevant for problems exhibiting more complicated dynamics. Recently there have been a number of studies of numerical stability for wider classes of problems admitting more complicated dynamics. This on-going work is unified and, in particular, striking similarities between this new developing stability theory and the classical linear and nonlinear stability theories are emphasized. The classical theories of A, B and algebraic stability for Runge-Kutta methods are briefly reviewed; the dynamics of solutions within the classes of equations to which these theories apply--linear decay and contractive problemsw are studied. Four other categories of equationsmgradient, dissipative, conservative and Hamiltonian systemsmare considered. Relationships and differences between the possible dynamics in each category, which range from multiple competing equilibria to chaotic solutions, are highlighted. Runge-Kutta schemes that preserve the dynamical structure of the underlying problem are sought, and indications of a strong relationship between the developing stability theory for these new categories and the classical existing stability theory for the older problems are given.
    [Show full text]
  • Geometric Algebra and Covariant Methods in Physics and Cosmology
    GEOMETRIC ALGEBRA AND COVARIANT METHODS IN PHYSICS AND COSMOLOGY Antony M Lewis Queens' College and Astrophysics Group, Cavendish Laboratory A dissertation submitted for the degree of Doctor of Philosophy in the University of Cambridge. September 2000 Updated 2005 with typo corrections Preface This dissertation is the result of work carried out in the Astrophysics Group of the Cavendish Laboratory, Cambridge, between October 1997 and September 2000. Except where explicit reference is made to the work of others, the work contained in this dissertation is my own, and is not the outcome of work done in collaboration. No part of this dissertation has been submitted for a degree, diploma or other quali¯cation at this or any other university. The total length of this dissertation does not exceed sixty thousand words. Antony Lewis September, 2000 iii Acknowledgements It is a pleasure to thank all those people who have helped me out during the last three years. I owe a great debt to Anthony Challinor and Chris Doran who provided a great deal of help and advice on both general and technical aspects of my work. I thank my supervisor Anthony Lasenby who provided much inspiration, guidance and encouragement without which most of this work would never have happened. I thank Sarah Bridle for organizing the useful lunchtime CMB discussion group from which I learnt a great deal, and the other members of the Cavendish Astrophysics Group for interesting discussions, in particular Pia Mukherjee, Carl Dolby, Will Grainger and Mike Hobson. I gratefully acknowledge ¯nancial support from PPARC. v Contents Preface iii Acknowledgements v 1 Introduction 1 2 Geometric Algebra 5 2.1 De¯nitions and basic properties .
    [Show full text]
  • Stability Analysis for Systems of Differential Equations
    Stability Analysis for Systems of Differential Equations David Eberly, Geometric Tools, Redmond WA 98052 https://www.geometrictools.com/ This work is licensed under the Creative Commons Attribution 4.0 International License. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/ or send a letter to Creative Commons, PO Box 1866, Mountain View, CA 94042, USA. Created: February 8, 2003 Last Modified: March 2, 2008 Contents 1 Introduction 2 2 Physical Stability 2 3 Numerical Stability 4 3.1 Stability for Single-Step Methods..................................4 3.2 Stability for Multistep Methods...................................5 3.3 Choosing a Stable Step Size.....................................7 4 The Simple Pendulum 7 4.1 Numerical Solution of the ODE...................................8 4.2 Physical Stability for the Pendulum................................ 13 4.3 Numerical Stability of the ODE Solvers.............................. 13 1 1 Introduction In setting up a physical simulation involving objects, a primary step is to establish the equations of motion for the objects. These equations are formulated as a system of second-order ordinary differential equations that may be converted to a system of first-order equations whose dependent variables are the positions and velocities of the objects. Such a system is of the generic form x_ = f(t; x); t ≥ 0; x(0) = x0 (1) where x0 is a specified initial condition for the system. The components of x are the positions and velocities of the objects. The function f(t; x) includes the external forces and torques of the system. A computer implementation of the physical simulation amounts to selecting a numerical method to approximate the solution to the system of differential equations.
    [Show full text]
  • Algorithmic Structure for Geometric Algebra Operators and Application to Quadric Surfaces Stephane Breuils
    Algorithmic structure for geometric algebra operators and application to quadric surfaces Stephane Breuils To cite this version: Stephane Breuils. Algorithmic structure for geometric algebra operators and application to quadric surfaces. Operator Algebras [math.OA]. Université Paris-Est, 2018. English. NNT : 2018PESC1142. tel-02085820 HAL Id: tel-02085820 https://pastel.archives-ouvertes.fr/tel-02085820 Submitted on 31 Mar 2019 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. ÉCOLE DOCTORALE MSTIC THÈSE DE DOCTORAT EN INFORMATIQUE Structure algorithmique pour les opérateurs d’Algèbres Géométriques et application aux surfaces quadriques Auteur: Encadrants: Stéphane Breuils Dr. Vincent NOZICK Dr. Laurent FUCHS Rapporteurs: Prof. Dominique MICHELUCCI Prof. Pascal SCHRECK Examinateurs: Dr. Leo DORST Prof. Joan LASENBY Prof. Raphaëlle CHAINE Soutenue le 17 Décembre 2018 iii Thèse effectuée au Laboratoire d’Informatique Gaspard-Monge, équipe A3SI, dans les locaux d’ESIEE Paris LIGM, UMR 8049 École doctorale Paris-Est Cité Descartes, Cité Descartes, Bâtiment Copernic-5, bd Descartes 6-8 av. Blaise Pascal Champs-sur-Marne Champs-sur-Marne, 77 454 Marne-la-Vallée Cedex 2 77 455 Marne-la-Vallée Cedex 2 v Abstract Geometric Algebra is considered as a very intuitive tool to deal with geometric problems and it appears to be increasingly efficient and useful to deal with computer graphics problems.
    [Show full text]
  • Isospectral Towers of Riemannian Manifolds
    New York Journal of Mathematics New York J. Math. 18 (2012) 451{461. Isospectral towers of Riemannian manifolds Benjamin Linowitz Abstract. In this paper we construct, for n ≥ 2, arbitrarily large fam- ilies of infinite towers of compact, orientable Riemannian n-manifolds which are isospectral but not isometric at each stage. In dimensions two and three, the towers produced consist of hyperbolic 2-manifolds and hyperbolic 3-manifolds, and in these cases we show that the isospectral towers do not arise from Sunada's method. Contents 1. Introduction 451 2. Genera of quaternion orders 453 3. Arithmetic groups derived from quaternion algebras 454 4. Isospectral towers and chains of quaternion orders 454 5. Proof of Theorem 4.1 456 5.1. Orders in split quaternion algebras over local fields 456 5.2. Proof of Theorem 4.1 457 6. The Sunada construction 458 References 461 1. Introduction Let M be a closed Riemannian n-manifold. The eigenvalues of the La- place{Beltrami operator acting on the space L2(M) form a discrete sequence of nonnegative real numbers in which each value occurs with a finite mul- tiplicity. This collection of eigenvalues is called the spectrum of M, and two Riemannian n-manifolds are said to be isospectral if their spectra coin- cide. Inverse spectral geometry asks the extent to which the geometry and topology of M is determined by its spectrum. Whereas volume and scalar curvature are spectral invariants, the isometry class is not. Although there is a long history of constructing Riemannian manifolds which are isospectral but not isometric, we restrict our discussion to those constructions most Received February 4, 2012.
    [Show full text]
  • The Making of a Geometric Algebra Package in Matlab Computer Science Department University of Waterloo Research Report CS-99-27
    The Making of a Geometric Algebra Package in Matlab Computer Science Department University of Waterloo Research Report CS-99-27 Stephen Mann∗, Leo Dorsty, and Tim Boumay [email protected], [email protected], [email protected] Abstract In this paper, we describe our development of GABLE, a Matlab implementation of the Geometric Algebra based on `p;q (where p + q = 3) and intended for tutorial purposes. Of particular note are the C matrix representation of geometric objects, effective algorithms for this geometry (inversion, meet and join), and issues in efficiency and numerics. 1 Introduction Geometric algebra extends Clifford algebra with geometrically meaningful operators, and its purpose is to facilitate geometrical computations. Present textbooks and implementation do not always convey this geometrical flavor or the computational and representational convenience of geometric algebra, so we felt a need for a computer tutorial in which representation, computation and visualization are combined to convey both the intuition and the techniques of geometric algebra. Current software packages are either Clifford algebra only (such as CLICAL [9] and CLIFFORD [1]) or do not include graphics [6], so we decide to build our own. The result is GABLE (Geometric Algebra Learning Environment) a hands-on tutorial on geometric algebra that should be accessible to the second year student in college [3]. The GABLE tutorial explains the basics of Geometric Algebra in an accessible manner. It starts with the outer product (as a constructor of subspaces), then treats the inner product (for perpendilarity), and moves via the geometric product (for invertibility) to the more geometrical operators such as projection, rotors, meet and join, and end with to the homogeneous model of Euclidean space.
    [Show full text]
  • Geometry of Matrix Decompositions Seen Through Optimal Transport and Information Geometry
    Published in: Journal of Geometric Mechanics doi:10.3934/jgm.2017014 Volume 9, Number 3, September 2017 pp. 335{390 GEOMETRY OF MATRIX DECOMPOSITIONS SEEN THROUGH OPTIMAL TRANSPORT AND INFORMATION GEOMETRY Klas Modin∗ Department of Mathematical Sciences Chalmers University of Technology and University of Gothenburg SE-412 96 Gothenburg, Sweden Abstract. The space of probability densities is an infinite-dimensional Rie- mannian manifold, with Riemannian metrics in two flavors: Wasserstein and Fisher{Rao. The former is pivotal in optimal mass transport (OMT), whereas the latter occurs in information geometry|the differential geometric approach to statistics. The Riemannian structures restrict to the submanifold of multi- variate Gaussian distributions, where they induce Riemannian metrics on the space of covariance matrices. Here we give a systematic description of classical matrix decompositions (or factorizations) in terms of Riemannian geometry and compatible principal bundle structures. Both Wasserstein and Fisher{Rao geometries are discussed. The link to matrices is obtained by considering OMT and information ge- ometry in the category of linear transformations and multivariate Gaussian distributions. This way, OMT is directly related to the polar decomposition of matrices, whereas information geometry is directly related to the QR, Cholesky, spectral, and singular value decompositions. We also give a coherent descrip- tion of gradient flow equations for the various decompositions; most flows are illustrated in numerical examples. The paper is a combination of previously known and original results. As a survey it covers the Riemannian geometry of OMT and polar decomposi- tions (smooth and linear category), entropy gradient flows, and the Fisher{Rao metric and its geodesics on the statistical manifold of multivariate Gaussian distributions.
    [Show full text]
  • A Guided Tour to the Plane-Based Geometric Algebra PGA
    A Guided Tour to the Plane-Based Geometric Algebra PGA Leo Dorst University of Amsterdam Version 1.15{ July 6, 2020 Planes are the primitive elements for the constructions of objects and oper- ators in Euclidean geometry. Triangulated meshes are built from them, and reflections in multiple planes are a mathematically pure way to construct Euclidean motions. A geometric algebra based on planes is therefore a natural choice to unify objects and operators for Euclidean geometry. The usual claims of `com- pleteness' of the GA approach leads us to hope that it might contain, in a single framework, all representations ever designed for Euclidean geometry - including normal vectors, directions as points at infinity, Pl¨ucker coordinates for lines, quaternions as 3D rotations around the origin, and dual quaternions for rigid body motions; and even spinors. This text provides a guided tour to this algebra of planes PGA. It indeed shows how all such computationally efficient methods are incorporated and related. We will see how the PGA elements naturally group into blocks of four coordinates in an implementation, and how this more complete under- standing of the embedding suggests some handy choices to avoid extraneous computations. In the unified PGA framework, one never switches between efficient representations for subtasks, and this obviously saves any time spent on data conversions. Relative to other treatments of PGA, this text is rather light on the mathematics. Where you see careful derivations, they involve the aspects of orientation and magnitude. These features have been neglected by authors focussing on the mathematical beauty of the projective nature of the algebra.
    [Show full text]