Information Geometry and Quantum Fields

Total Page:16

File Type:pdf, Size:1020Kb

Information Geometry and Quantum Fields Information Geometry and Quantum Fields Kevin T. Grosvenor Julius-Maximilians Universität Würzburg, Germany SPIGL ’20, Les Houches, France July 30, 2020 Collaborators SciPost Phys. 8 (2020) 073 [arXiv:2001.02683] Johanna Erdmenger Ro Jefferson 1 32 Sources and References 2 32 Outline 1. Introduction to AdS/CFT 2. Fisher metric, Symmetries and Uniqueness 3. Classical metric on States: Instantons 4. Classical metric on Theories: Ising Model 5. Quantum metric on States: Coherent states 6. Connections and Curvature 7. Conclusions and Outlook 3 32 AdS/CFT Correspondence Holography: G. ’t Hooft (THU-93/26, ’93), L. Susskind (JMP 36, ’95) AdS/CFT: J. Maldacena (ATMP 2, ’97) Weak-strong duality. D R d E e d x φ0(x) O(x) = e−SSUGRA CFT φ(0;x)=φ0(x) 4 32 d;2 Embedding into R : −(X0)2 + (X1)2 + ··· + (Xd)2 − (Xd+1)2 = −L2 d(d−1) 1;:::;d Isometries: 2 rotations among X , one rotation between X0 and Xd+1, and 2d boosts mixing X0 and Xd+1 with X1;:::;d. Algebra of isometries is so(d; 2) . Anti-de Sitter Spacetime 2 L2 2 r2 µ ν Poincaré patch: ds = r2 dr + L2 ηµνdx dx . Constant negative curvature, maximally symmetric spacetime. 1 R d+1 p Vacuum of Einstein-Hilbert S = 16πG d x −g(R − 2Λ) with constant negative Λ (cosmological constant). 5 32 Anti-de Sitter Spacetime 2 L2 2 r2 µ ν Poincaré patch: ds = r2 dr + L2 ηµνdx dx . Constant negative curvature, maximally symmetric spacetime. 1 R d+1 p Vacuum of Einstein-Hilbert S = 16πG d x −g(R − 2Λ) with constant negative Λ (cosmological constant). d;2 Embedding into R : −(X0)2 + (X1)2 + ··· + (Xd)2 − (Xd+1)2 = −L2 d(d−1) 1;:::;d Isometries: 2 rotations among X , one rotation between X0 and Xd+1, and 2d boosts mixing X0 and Xd+1 with X1;:::;d. Algebra of isometries is so(d; 2) . 5 32 CFT fields form representations of the conformal algebra. Conformal Field Theory Translations Pµ = i@µ Rotations/boosts Jµν = 2ix[µ@ν] µ Dilatations D = ix @µ 2 ν SCT Kµ = i(x @µ − 2xµx @ν) Conformal algebra is so(d; 2) . In 1 + 1d, extended to infinite Virasoro algebra. 6 32 Conformal Field Theory Translations Pµ = i@µ Rotations/boosts Jµν = 2ix[µ@ν] µ Dilatations D = ix @µ 2 ν SCT Kµ = i(x @µ − 2xµx @ν) Conformal algebra is so(d; 2) . In 1 + 1d, extended to infinite Virasoro algebra. CFT fields form representations of the conformal algebra. 6 32 AdS/CFT Textbook 7 32 Quantum Information Theory in Holography Entanglement probes of the bulk. [0603011 Ryu, Takayanagi; 0705.0016 Hubeny, Rangamani, Takayanagi; 1408.3203 Engelhardt, Wall] Quantum error correction =) bulk reconstruction. [1411.7041 Almheiri, Dong, Harlow] Tensor networks =) bulk-boundary maps. [1601.01694 Hayden et. al.] Holographic distance measures. [pure: 1507.0755 Lashkari, van Raamsdonk; mixed: 1701.02319 Banerjee, Erdmenger, Sarkar] 8 32 Previous Info Geom + Physics work Ruppeiner (’95,’96) - stat phys and thermo (e.g., ideal gas) Brody & Hook (2008) - stat phys and thermo (e.g., vdW gas) Janke et. al. (2002, 2003) - 1d Ising & Spherical models Dolan, Johnston, Kenna (2002) - Potts model Amari (’97) - Neural networks Ke & Nielsen (2016) - machine learning Heckman (2013) - string theory Blau, Narain, Thompson (2001) - YM instantons in 3+1-d 9 32 R Fisher metric: gij(ξ) ≡ dx pξ(x) @i ln pξ(x) @j ln pξ(x) R R q q Fact 1: gij (ξ) = − dx pξ(x) @i@j ln pξ(x) and gij (ξ) = 4 dx @i pξ(x) @j pξ(x) Fact 2: gij is positive semi-definite. Fisher Metric x x x pξ(x) ≥ 0 p R R dx pξ(x) = 1 ξ injective @ ξ −−−−! pξ smooth, @i ≡ @ξi 10 32 R R q q Fact 1: gij (ξ) = − dx pξ(x) @i@j ln pξ(x) and gij (ξ) = 4 dx @i pξ(x) @j pξ(x) Fact 2: gij is positive semi-definite. Fisher Metric x x x pξ(x) ≥ 0 p R R dx pξ(x) = 1 ξ injective @ ξ −−−−! pξ smooth, @i ≡ @ξi R Fisher metric: gij(ξ) ≡ dx pξ(x) @i ln pξ(x) @j ln pξ(x) 10 32 Fisher Metric x x x pξ(x) ≥ 0 p R R dx pξ(x) = 1 ξ injective @ ξ −−−−! pξ smooth, @i ≡ @ξi R Fisher metric: gij(ξ) ≡ dx pξ(x) @i ln pξ(x) @j ln pξ(x) R R q q Fact 1: gij (ξ) = − dx pξ(x) @i@j ln pξ(x) and gij (ξ) = 4 dx @i pξ(x) @j pξ(x) Fact 2: gij is positive semi-definite. 10 32 dµ2 + 2dσ2 ds2 = σ2 2 2 1 σ2 2 dµ +dσ Alternative example: pµ,σ (x) = gives ds = . πσ (x−µ)2+σ2 2σ2 Comments: 1. It is not true that all free theories have flat Fisher metrics. 2. This hyperbolic space is a priori unrelated to AdS/CFT. Gaussian Distribution A simple example (1d Gaussians): Random variable space is X = R. Parameter space is Ξ = upper half-plane = (µ, σ) 2 R × R>0 (x−µ)2 1 − p (x) = p e 2σ2 µ,σ 2π σ 11 32 2 2 1 σ2 2 dµ +dσ Alternative example: pµ,σ (x) = gives ds = . πσ (x−µ)2+σ2 2σ2 Comments: 1. It is not true that all free theories have flat Fisher metrics. 2. This hyperbolic space is a priori unrelated to AdS/CFT. Gaussian Distribution A simple example (1d Gaussians): Random variable space is X = R. Parameter space is Ξ = upper half-plane = (µ, σ) 2 R × R>0 (x−µ)2 1 − p (x) = p e 2σ2 µ,σ 2π σ dµ2 + 2dσ2 ds2 = σ2 11 32 Comments: 1. It is not true that all free theories have flat Fisher metrics. 2. This hyperbolic space is a priori unrelated to AdS/CFT. Gaussian Distribution A simple example (1d Gaussians): Random variable space is X = R. Parameter space is Ξ = upper half-plane = (µ, σ) 2 R × R>0 (x−µ)2 1 − p (x) = p e 2σ2 µ,σ 2π σ dµ2 + 2dσ2 ds2 = σ2 2 2 1 σ2 2 dµ +dσ Alternative example: pµ,σ (x) = gives ds = . πσ (x−µ)2+σ2 2σ2 11 32 Gaussian Distribution A simple example (1d Gaussians): Random variable space is X = R. Parameter space is Ξ = upper half-plane = (µ, σ) 2 R × R>0 (x−µ)2 1 − p (x) = p e 2σ2 µ,σ 2π σ dµ2 + 2dσ2 ds2 = σ2 2 2 1 σ2 2 dµ +dσ Alternative example: pµ,σ (x) = gives ds = . πσ (x−µ)2+σ2 2σ2 Comments: 1. It is not true that all free theories have flat Fisher metrics. 2. This hyperbolic space is a priori unrelated to AdS/CFT. 11 32 gij(ξ) = @i@j (ξ) R Fact: gij (ξ) = h(Fi − hFiiξ)(Fj − hFj iξiξ where hfiξ ≡ dx f(x) pξ(x). Exponential Family Generalization of the Gaussian (Exponential Family): i pξ(x) = exp C(x) + ξ Fi(x) − (ξ) . R i (ξ) = ln dx exp C(x) + ξ Fi(x) is a normalization factor. Assume that fC; Fig are lin. indep. so ξ ! pξ is bijective. 12 32 Exponential Family Generalization of the Gaussian (Exponential Family): i pξ(x) = exp C(x) + ξ Fi(x) − (ξ) . R i (ξ) = ln dx exp C(x) + ξ Fi(x) is a normalization factor. Assume that fC; Fig are lin. indep. so ξ ! pξ is bijective. gij(ξ) = @i@j (ξ) R Fact: gij (ξ) = h(Fi − hFiiξ)(Fj − hFj iξiξ where hfiξ ≡ dx f(x) pξ(x). 12 32 Gaussian example: µ!µ+c −−−−−−−−! −−−−!x!x+c (µ,σ)!λ(µ,σ) −−−−−−−−! −−−−−!x!λx Symmetries ξ ! ξ~(ξ) is a symmetry of p if it can be undone by x ! x~(x) pξ~(x) dx = pξ(~x) dx~ Fact: a symmetry of p is also a symmetry of gij . 13 32 Symmetries ξ ! ξ~(ξ) is a symmetry of p if it can be undone by x ! x~(x) pξ~(x) dx = pξ(~x) dx~ Fact: a symmetry of p is also a symmetry of gij . Gaussian example: µ!µ+c −−−−−−−−! −−−−!x!x+c (µ,σ)!λ(µ,σ) −−−−−−−−! −−−−−!x!λx 13 32 Non-Uniqueness Infinitely many distributions give the same information geometry [Clingman, Murugan and Shock 2015] E.g., 1d Gaussian and Cauchy-Lorentz give 2d hyperbolic space Cauchy-Lorentz is not in the exponential family Weirder example in Clingman, Murugan, Shock (2015) of a 3d distribution that gives 2d hyperbolic space, but has none of its symmetries. 14 32 Lessons Many statistical models S can lead to the same Fisher metric (Gaussian and Cauchy-Lorentz). FM inherits the symmetries of S but can have more 15 32 R 4 µ 2 4 Simpler model: d x @µφ @ φ − g φ (Euclidean signature) 2 σ2 Exact solution to e.o.m.: φ~µ,σ(~x) = gσ j~x−~µj2+σ2 16σ2 j~x−~µj2−σ2 Lagrangian density: L~µ,σ(~x) = g2 (j~x−~µj2+σ2)4 L~µ,σ(~x) < 0 for j~x − ~µj < σ, so not a good prob. distr. 2 2 4 16 σ4 If instead we use −φ @ φ − g φ , we get p~µ,σ(~x) = g2 (j~x−~µj2+σ2)4 and the information geometry is 5d hyperbolic space. Classical metric on States: Instantons Blau, Narain, Thompson (2001) - YM instantons in 3+1-d Uses prescription of Hitchin (1990): probability distribution is taken to be the Lagrangian density evaluated on the state.
Recommended publications
  • Introduction to Information Geometry – Based on the Book “Methods of Information Geometry ”Written by Shun-Ichi Amari and Hiroshi Nagaoka
    Introduction to Information Geometry – based on the book “Methods of Information Geometry ”written by Shun-Ichi Amari and Hiroshi Nagaoka Yunshu Liu 2012-02-17 Introduction to differential geometry Geometric structure of statistical models and statistical inference Outline 1 Introduction to differential geometry Manifold and Submanifold Tangent vector, Tangent space and Vector field Riemannian metric and Affine connection Flatness and autoparallel 2 Geometric structure of statistical models and statistical inference The Fisher metric and α-connection Exponential family Divergence and Geometric statistical inference Yunshu Liu (ASPITRG) Introduction to Information Geometry 2 / 79 Introduction to differential geometry Geometric structure of statistical models and statistical inference Part I Introduction to differential geometry Yunshu Liu (ASPITRG) Introduction to Information Geometry 3 / 79 Introduction to differential geometry Geometric structure of statistical models and statistical inference Basic concepts in differential geometry Basic concepts Manifold and Submanifold Tangent vector, Tangent space and Vector field Riemannian metric and Affine connection Flatness and autoparallel Yunshu Liu (ASPITRG) Introduction to Information Geometry 4 / 79 Introduction to differential geometry Geometric structure of statistical models and statistical inference Manifold Manifold S n Manifold: a set with a coordinate system, a one-to-one mapping from S to R , n supposed to be ”locally” looks like an open subset of R ” n Elements of the set(points): points in R , probability distribution, linear system. Figure : A coordinate system ξ for a manifold S Yunshu Liu (ASPITRG) Introduction to Information Geometry 5 / 79 Introduction to differential geometry Geometric structure of statistical models and statistical inference Manifold Manifold S Definition: Let S be a set, if there exists a set of coordinate systems A for S which satisfies the condition (1) and (2) below, we call S an n-dimensional C1 differentiable manifold.
    [Show full text]
  • Arxiv:1907.11122V2 [Math-Ph] 23 Aug 2019 on M Which Are Dual with Respect to G
    Canonical divergence for flat α-connections: Classical and Quantum Domenico Felice1, ∗ and Nihat Ay2, y 1Max Planck Institute for Mathematics in the Sciences Inselstrasse 22{04103 Leipzig, Germany 2 Max Planck Institute for Mathematics in the Sciences Inselstrasse 22{04103 Leipzig, Germany Santa Fe Institute, 1399 Hyde Park Rd, Santa Fe, NM 87501, USA Faculty of Mathematics and Computer Science, University of Leipzig, PF 100920, 04009 Leipzig, Germany A recent canonical divergence, which is introduced on a smooth manifold M endowed with a general dualistic structure (g; r; r∗), is considered for flat α-connections. In the classical setting, we compute such a canonical divergence on the manifold of positive measures and prove that it coincides with the classical α-divergence. In the quantum framework, the recent canonical divergence is evaluated for the quantum α-connections on the manifold of all positive definite Hermitian operators. Also in this case we obtain that the recent canonical divergence is the quantum α-divergence. PACS numbers: Classical differential geometry (02.40.Hw), Riemannian geometries (02.40.Ky), Quantum Information (03.67.-a). I. INTRODUCTION Methods of Information Geometry (IG) [1] are ubiquitous in physical sciences and encom- pass both classical and quantum systems [2]. The natural object of study in IG is a quadruple (M; g; r; r∗) given by a smooth manifold M, a Riemannian metric g and a pair of affine connections arXiv:1907.11122v2 [math-ph] 23 Aug 2019 on M which are dual with respect to g, ∗ X g (Y; Z) = g (rX Y; Z) + g (Y; rX Z) ; (1) for all sections X; Y; Z 2 T (M).
    [Show full text]
  • Information Geometry of Orthogonal Initializations and Training
    Published as a conference paper at ICLR 2020 INFORMATION GEOMETRY OF ORTHOGONAL INITIALIZATIONS AND TRAINING Piotr Aleksander Sokół and Il Memming Park Department of Neurobiology and Behavior Departments of Applied Mathematics and Statistics, and Electrical and Computer Engineering Institutes for Advanced Computing Science and AI-driven Discovery and Innovation Stony Brook University, Stony Brook, NY 11733 {memming.park, piotr.sokol}@stonybrook.edu ABSTRACT Recently mean field theory has been successfully used to analyze properties of wide, random neural networks. It gave rise to a prescriptive theory for initializing feed-forward neural networks with orthogonal weights, which ensures that both the forward propagated activations and the backpropagated gradients are near `2 isome- tries and as a consequence training is orders of magnitude faster. Despite strong empirical performance, the mechanisms by which critical initializations confer an advantage in the optimization of deep neural networks are poorly understood. Here we show a novel connection between the maximum curvature of the optimization landscape (gradient smoothness) as measured by the Fisher information matrix (FIM) and the spectral radius of the input-output Jacobian, which partially explains why more isometric networks can train much faster. Furthermore, given that or- thogonal weights are necessary to ensure that gradient norms are approximately preserved at initialization, we experimentally investigate the benefits of maintaining orthogonality throughout training, and we conclude that manifold optimization of weights performs well regardless of the smoothness of the gradients. Moreover, we observe a surprising yet robust behavior of highly isometric initializations — even though such networks have a lower FIM condition number at initialization, and therefore by analogy to convex functions should be easier to optimize, experimen- tally they prove to be much harder to train with stochastic gradient descent.
    [Show full text]
  • Information Geometry and Optimal Transport
    Call for Papers Special Issue Affine Differential Geometry and Hesse Geometry: A Tribute and Memorial to Jean-Louis Koszul Submission Deadline: 30th November 2019 Jean-Louis Koszul (January 3, 1921 – January 12, 2018) was a French mathematician with prominent influence to a wide range of mathematical fields. He was a second generation member of Bourbaki, with several notions in geometry and algebra named after him. He made a great contribution to the fundamental theory of Differential Geometry, which is foundation of Information Geometry. The special issue is dedicated to Koszul for the mathematics he developed that bear on information sciences. Both original contributions and review articles are solicited. Topics include but are not limited to: Affine differential geometry over statistical manifolds Hessian and Kahler geometry Divergence geometry Convex geometry and analysis Differential geometry over homogeneous and symmetric spaces Jordan algebras and graded Lie algebras Pre-Lie algebras and their cohomology Geometric mechanics and Thermodynamics over homogeneous spaces Guest Editor Hideyuki Ishi (Graduate School of Mathematics, Nagoya University) e-mail: [email protected] Submit at www.editorialmanager.com/inge Please select 'S.I.: Affine differential geometry and Hessian geometry Jean-Louis Koszul (Memorial/Koszul)' Photo Author: Konrad Jacobs. Source: Archives of the Mathema�sches Forschungsins�tut Oberwolfach Editor-in-Chief Shinto Eguchi (Tokyo) Special Issue Co-Editors Information Geometry and Optimal Transport Nihat
    [Show full text]
  • Information Geometry (Part 1)
    October 22, 2010 Information Geometry (Part 1) John Baez Information geometry is the study of 'statistical manifolds', which are spaces where each point is a hypothesis about some state of affairs. In statistics a hypothesis amounts to a probability distribution, but we'll also be looking at the quantum version of a probability distribution, which is called a 'mixed state'. Every statistical manifold comes with a way of measuring distances and angles, called the Fisher information metric. In the first seven articles in this series, I'll try to figure out what this metric really means. The formula for it is simple enough, but when I first saw it, it seemed quite mysterious. A good place to start is this interesting paper: • Gavin E. Crooks, Measuring thermodynamic length. which was pointed out by John Furey in a discussion about entropy and uncertainty. The idea here should work for either classical or quantum statistical mechanics. The paper describes the classical version, so just for a change of pace let me describe the quantum version. First a lightning review of quantum statistical mechanics. Suppose you have a quantum system with some Hilbert space. When you know as much as possible about your system, then you describe it by a unit vector in this Hilbert space, and you say your system is in a pure state. Sometimes people just call a pure state a 'state'. But that can be confusing, because in statistical mechanics you also need more general 'mixed states' where you don't know as much as possible. A mixed state is described by a density matrix, meaning a positive operator with trace equal to 1: tr( ) = 1 The idea is that any observable is described by a self-adjoint operator A, and the expected value of this observable in the mixed state is A = tr( A) The entropy of a mixed state is defined by S( ) = −tr( ln ) where we take the logarithm of the density matrix just by taking the log of each of its eigenvalues, while keeping the same eigenvectors.
    [Show full text]
  • Information Geometry: Near Randomness and Near Independence
    Khadiga Arwini and C.T.J. Dodson Information Geometry: Near Randomness and Near Independence Lecture Notes in Mathematics 1953 Springer Preface The main motivation for this book lies in the breadth of applications in which a statistical model is used to represent small departures from, for example, a Poisson process. Our approach uses information geometry to provide a com- mon context but we need only rather elementary material from differential geometry, information theory and mathematical statistics. Introductory sec- tions serve together to help those interested from the applications side in making use of our methods and results. We have available Mathematica note- books to perform many of the computations for those who wish to pursue their own calculations or developments. Some 44 years ago, the second author first encountered, at about the same time, differential geometry via relativity from Weyl's book [209] during un- dergraduate studies and information theory from Tribus [200, 201] via spatial statistical processes while working on research projects at Wiggins Teape Re- search and Development Ltd|cf. the Foreword in [196] and [170, 47, 58]. Hav- ing started work there as a student laboratory assistant in 1959, this research environment engendered a recognition of the importance of international col- laboration, and a lifelong research interest in randomness and near-Poisson statistical geometric processes, persisting at various rates through a career mainly involved with global differential geometry. From correspondence in the 1960s with Gabriel Kron [4, 124, 125] on his Diakoptics, and with Kazuo Kondo who influenced the post-war Japanese schools of differential geometry and supervised Shun-ichi Amari's doctorate [6], it was clear that both had a much wider remit than traditionally pursued elsewhere.
    [Show full text]
  • Exponential Families: Dually-Flat, Hessian and Legendre Structures
    entropy Review Information Geometry of k-Exponential Families: Dually-Flat, Hessian and Legendre Structures Antonio M. Scarfone 1,* ID , Hiroshi Matsuzoe 2 ID and Tatsuaki Wada 3 ID 1 Istituto dei Sistemi Complessi, Consiglio Nazionale delle Ricerche (ISC-CNR), c/o Politecnico di Torino, 10129 Torino, Italy 2 Department of Computer Science and Engineering, Graduate School of Engineering, Nagoya Institute of Technology, Gokiso-cho, Showa-ku, Nagoya 466-8555, Japan; [email protected] 3 Region of Electrical and Electronic Systems Engineering, Ibaraki University, Nakanarusawa-cho, Hitachi 316-8511, Japan; [email protected] * Correspondence: [email protected]; Tel.: +39-011-090-7339 Received: 9 May 2018; Accepted: 1 June 2018; Published: 5 June 2018 Abstract: In this paper, we present a review of recent developments on the k-deformed statistical mechanics in the framework of the information geometry. Three different geometric structures are introduced in the k-formalism which are obtained starting from three, not equivalent, divergence functions, corresponding to the k-deformed version of Kullback–Leibler, “Kerridge” and Brègman divergences. The first statistical manifold derived from the k-Kullback–Leibler divergence form an invariant geometry with a positive curvature that vanishes in the k → 0 limit. The other two statistical manifolds are related to each other by means of a scaling transform and are both dually-flat. They have a dualistic Hessian structure endowed by a deformed Fisher metric and an affine connection that are consistent with a statistical scalar product based on the k-escort expectation. These flat geometries admit dual potentials corresponding to the thermodynamic Massieu and entropy functions that induce a Legendre structure of k-thermodynamics in the picture of the information geometry.
    [Show full text]
  • Information-Geometric Optimization Algorithms: a Unifying Picture Via Invariance Principles
    Journal of Machine Learning Research 18 (2017) 1-65 Submitted 11/14; Revised 10/16; Published 4/17 Information-Geometric Optimization Algorithms: A Unifying Picture via Invariance Principles Yann Ollivier [email protected] CNRS & LRI (UMR 8623), Université Paris-Saclay 91405 Orsay, France Ludovic Arnold [email protected] Univ. Paris-Sud, LRI 91405 Orsay, France Anne Auger [email protected] Nikolaus Hansen [email protected] Inria & CMAP, Ecole polytechnique 91128 Palaiseau, France Editor: Una-May O’Reilly Abstract We present a canonical way to turn any smooth parametric family of probability distributions on an arbitrary search space 푋 into a continuous-time black-box optimization method on 푋, the information-geometric optimization (IGO) method. Invariance as a major design principle keeps the number of arbitrary choices to a minimum. The resulting IGO flow is the flow of an ordinary differential equation conducting the natural gradient ascentofan adaptive, time-dependent transformation of the objective function. It makes no particular assumptions on the objective function to be optimized. The IGO method produces explicit IGO algorithms through time discretization. It naturally recovers versions of known algorithms and offers a systematic way to derive new ones. In continuous search spaces, IGO algorithms take a form related to natural evolution strategies (NES). The cross-entropy method is recovered in a particular case with a large time step, and can be extended into a smoothed, parametrization-independent maximum likelihood update (IGO-ML). When applied to the family of Gaussian distributions on R푑, the IGO framework recovers a version of the well-known CMA-ES algorithm and of xNES.
    [Show full text]
  • Fisher-Rao Geometry of Dirichlet Distributions Alice Le Brigant, Stephen Preston, Stéphane Puechmorel
    Fisher-Rao geometry of Dirichlet distributions Alice Le Brigant, Stephen Preston, Stéphane Puechmorel To cite this version: Alice Le Brigant, Stephen Preston, Stéphane Puechmorel. Fisher-Rao geometry of Dirich- let distributions. Differential Geometry and its Applications, Elsevier, 2021, 74, pp.101702. 10.1016/j.difgeo.2020.101702. hal-02568473v2 HAL Id: hal-02568473 https://hal.archives-ouvertes.fr/hal-02568473v2 Submitted on 7 Nov 2020 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. FISHER-RAO GEOMETRY OF DIRICHLET DISTRIBUTIONS ALICE LE BRIGANT, STEPHEN C. PRESTON, AND STEPHANE´ PUECHMOREL Abstract. In this paper, we study the geometry induced by the Fisher-Rao metric on the parameter space of Dirichlet distributions. We show that this space is a Hadamard manifold, i.e. that it is geodesically complete and has everywhere negative sectional curvature. An important consequence for applications is that the Fr´echet mean of a set of Dirichlet distributions is uniquely defined in this geometry. 1. Introduction The differential geometric approach to probability theory and statistics has met increasing interest in the past years, from the theoretical point of view as well as in applications. In this approach, probability distributions are seen as elements of a differentiable manifold, on which a metric structure is defined through the choice of a Riemannian metric.
    [Show full text]
  • Information-Geometric Optimization Algorithms: a Unifying Picture Via Invariance Principles Yann Ollivier, Ludovic Arnold, Anne Auger, Nikolaus Hansen
    Information-Geometric Optimization Algorithms: A Unifying Picture via Invariance Principles Yann Ollivier, Ludovic Arnold, Anne Auger, Nikolaus Hansen To cite this version: Yann Ollivier, Ludovic Arnold, Anne Auger, Nikolaus Hansen. Information-Geometric Optimization Algorithms: A Unifying Picture via Invariance Principles. 2011. hal-00601503v2 HAL Id: hal-00601503 https://hal.archives-ouvertes.fr/hal-00601503v2 Preprint submitted on 29 Jun 2013 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Information-Geometric Optimization Algorithms: A Unifying Picture via Invariance Principles Yann Ollivier, Ludovic Arnold, Anne Auger, Nikolaus Hansen Abstract We present a canonical way to turn any smooth parametric fam- ily of probability distributions on an arbitrary search space X into a continuous-time black-box optimization method on X, the information- geometric optimization (IGO) method. Invariance as a major design principle keeps the number of arbitrary choices to a minimum. The resulting IGO flow is the flow of an ordinary differential equation con- ducting the natural gradient ascent of an adaptive, time-dependent transformation of the objective function. It makes no particular as- sumptions on the objective function to be optimized. The IGO method produces explicit IGO algorithms through time discretization.
    [Show full text]
  • Quantum Metrology and Thermometry in Open Systems. Thermal
    ADVERTIMENT. Lʼaccés als continguts dʼaquesta tesi queda condicionat a lʼacceptació de les condicions dʼús establertes per la següent llicència Creative Commons: http://cat.creativecommons.org/?page_id=184 ADVERTENCIA. El acceso a los contenidos de esta tesis queda condicionado a la aceptación de las condiciones de uso establecidas por la siguiente licencia Creative Commons: http://es.creativecommons.org/blog/licencias/ WARNING. The access to the contents of this doctoral thesis it is limited to the acceptance of the use conditions set by the following Creative Commons license: https://creativecommons.org/licenses/?lang=en Universitat Autonoma` de Barcelona Quantum Metrology and Thermometry in Open Systems Thermal Equilibrium and Beyond by Mohammad Mehboudi under supervision of Dr. Anna Sanpera Trigueros A thesis submitted in partial fulfillment for the degree of Doctor of Philosophy in Unitat de F´ısica Teorica:` Informacio´ i Fenomens` Quantics` Departament de F´ısica Facultat de Ciencies` Bellaterra, April 2017 iii Abstract In my thesis I explore quantum metrology in open systems, with a special focus on quantum temperature estimation, or thermometry. For this aim, I categorize my study in two different regimes of thermal equilibrium and beyond thermal equilibrium. In both cases, my collabora- tors and I, raise questions of fundamental and technological interest. Among such questions, I point out the followings: What are the ultimate precision bounds on thermometry with individual (single) probes? Is it possible to improve these bounds by using quantum resources such as quantum correlations and quantum criticality? We not only find the ultimate precision bound on thermometry, posed by physical laws of na- ture, but also show how to exploit quantum resources to surpass the classical bounds on preci- sion, even at finite temperature.
    [Show full text]
  • The Fisher-Rao Geometry of Beta Distributions Applied to the Study of Canonical Moments Alice Le Brigant, Stéphane Puechmorel
    The Fisher-Rao geometry of beta distributions applied to the study of canonical moments Alice Le Brigant, Stéphane Puechmorel To cite this version: Alice Le Brigant, Stéphane Puechmorel. The Fisher-Rao geometry of beta distributions applied to the study of canonical moments. 2019. hal-02100897 HAL Id: hal-02100897 https://hal.archives-ouvertes.fr/hal-02100897 Preprint submitted on 16 Apr 2019 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. The Fisher-Rao geometry of beta distributions applied to the study of canonical moments Alice Le Brigant and Stéphane Puechmorel Abstract This paper studies the Fisher-Rao geometry on the parameter space of beta distributions. We derive the geodesic equations and the sectional curvature, and prove that it is negative. This leads to uniqueness for the Riemannian centroid in that space. We use this Riemannian structure to study canonical moments, an intrinsic representation of the moments of a probability distribution. Drawing on the fact that a uniform distribution in the regular moment space corresponds to a product of beta distributions in the canonical moment space, we propose a mapping from the space of canonical moments to the product beta manifold, allowing us to use the Fisher-Rao geometry of beta distributions to compare and analyze canonical moments.
    [Show full text]