A Course on Large Deviations with an Introduction to Gibbs Measures

Total Page:16

File Type:pdf, Size:1020Kb

A Course on Large Deviations with an Introduction to Gibbs Measures A Course on Large Deviations with an Introduction to Gibbs Measures Firas Rassoul-Agha Timo Seppäläinen Graduate Studies in Mathematics Volume 162 American Mathematical Society A Course on Large Deviations with an Introduction to Gibbs Measures https://doi.org/10.1090//gsm/162 A Course on Large Deviations with an Introduction to Gibbs Measures Firas Rassoul-Agha Timo Seppäläinen Graduate Studies in Mathematics Volume 162 American Mathematical Society Providence, Rhode Island EDITORIAL COMMITTEE Dan Abramovich Daniel S. Freed Rafe Mazzeo (Chair) Gigliola Staffilani 2010 Mathematics Subject Classification. Primary 60-01, 60F10, 60J10, 60K35, 60K37, 82B05, 82B20. For additional information and updates on this book, visit www.ams.org/bookpages/gsm-162 Library of Congress Cataloging-in-Publication Data Rassoul-Agha, Firas. A course on large deviations with an introduction to Gibbs measures / Firas Rassoul-Agha, Timo Sepp¨al¨ainen. pages cm. — (Graduate studies in mathematics ; volume 162) Includes bibliographical references and indexes. ISBN 978-0-8218-7578-0 (alk. paper) 1. Large deviations. 2. Probabilities. 3. Measure theory. I. Sepp¨al¨ainen, Timo O. II. Title. QA273.67.R375 2015 519.2—dc23 2014035275 Copying and reprinting. Individual readers of this publication, and nonprofit libraries acting for them, are permitted to make fair use of the material, such as to copy select pages for use in teaching or research. Permission is granted to quote brief passages from this publication in reviews, provided the customary acknowledgment of the source is given. Republication, systematic copying, or multiple reproduction of any material in this publication is permitted only under license from the American Mathematical Society. Permissions to reuse portions of AMS publication content are handled by Copyright Clearance Center’s RightsLink service. For more information, please visit: http://www.ams.org/rightslink. Send requests for translation rights and licensed reprints to [email protected]. Excluded from these provisions is material for which the author holds copyright. In such cases, requests for permission to reuse or reprint material should be addressed directly to the author(s). Copyright ownership is indicated on the copyright page, or on the lower right-hand corner of the first page of each article within proceedings volumes. c 2015 by the American Mathematical Society. All rights reserved. The American Mathematical Society retains all rights except those granted to the United States Government. Printed in the United States of America. ∞ The paper used in this book is acid-free and falls within the guidelines established to ensure permanence and durability. Visit the AMS home page at http://www.ams.org/ 10987654321 201918171615 To Alla, Maxim, and Kirill To Celeste, David, Ansa, and Timo Contents Preface xi Part I. Large deviations: General theory and i.i.d. processes Chapter 1. Introductory discussion 3 §1.1. Information-theoretic entropy 5 §1.2. Thermodynamic entropy 8 §1.3. Large deviations as useful estimates 12 Chapter 2. The large deviation principle 17 §2.1. Precise asymptotics on an exponential scale 17 §2.2. Lower semicontinuous and tight rate functions 20 §2.3. Weak large deviation principle 23 §2.4. Aspects of Cram´er’s theorem 26 §2.5. Limits, deviations, and fluctuations 33 Chapter 3. Large deviations and asymptotics of integrals 35 §3.1. Contraction principle 35 §3.2. Varadhan’s theorem 37 §3.3. Bryc’s theorem 41 §3.4. Curie-Weiss model of ferromagnetism 43 Chapter 4. Convex analysis in large deviation theory 49 §4.1. Some elementary convex analysis 49 §4.2. Rate function as a convex conjugate 58 §4.3. Multidimensional Cram´er theorem 61 vii viii Contents Chapter 5. Relative entropy and large deviations for empirical measures 67 §5.1. Relative entropy 67 §5.2. Sanov’s theorem 73 §5.3. Maximum entropy principle 78 Chapter 6. Process level large deviations for i.i.d. fields 83 §6.1. Setting 83 §6.2. Specific relative entropy 85 §6.3. Pressure and the large deviation principle 91 Part II. Statistical mechanics Chapter 7. Formalism for classical lattice systems 99 §7.1. Finite-volume model 99 §7.2. Potentials and Hamiltonians 101 §7.3. Specifications 103 §7.4. Phase transition 108 §7.5. Extreme Gibbs measures 110 §7.6. Uniqueness for small potentials 112 Chapter 8. Large deviations and equilibrium statistical mechanics 121 §8.1. Thermodynamic limit of the pressure 121 §8.2. Entropy and large deviations under Gibbs measures 124 §8.3. Dobrushin-Lanford-Ruelle (DLR) variational principle 127 Chapter 9. Phase transition in the Ising model 133 §9.1. One-dimensional Ising model 136 §9.2. Phase transition at low temperature 138 §9.3. Case of no external field 141 §9.4. Case of nonzero external field 146 Chapter 10. Percolation approach to phase transition 149 §10.1. Bernoulli bond percolation and random cluster measures 149 §10.2. Ising phase transition revisited 153 Contents ix Part III. Additional large deviation topics Chapter 11. Further asymptotics for i.i.d. random variables 161 §11.1. Refinement of Cram´er’s theorem 161 §11.2. Moderate deviations 164 Chapter 12. Large deviations through the limiting generating function 167 §12.1. Essential smoothness and exposed points 167 §12.2. G¨artner-Ellis theorem 175 §12.3. Large deviations for the current of particles 179 Chapter 13. Large deviations for Markov chains 187 §13.1. Relative entropy for kernels 187 §13.2. Countable Markov chains 191 §13.3. Finite Markov chains 203 Chapter 14. Convexity criterion for large deviations 213 Chapter 15. Nonstationary independent variables 221 §15.1. Generalization of relative entropy and Sanov’s theorem 221 §15.2. Proof of the large deviation principle 223 Chapter 16. Random walk in a dynamical random environment 233 §16.1. Quenched large deviation principles 234 §16.2. Proofs via the Baxter-Jain theorem 239 Appendixes Appendix A. Analysis 259 §A.1. Metric spaces and topology 259 §A.2. Measure and integral 262 §A.3. Product spaces 267 §A.4. Separation theorem 268 §A.5. Minimax theorem 269 x Contents Appendix B. Probability 273 §B.1. Independence 274 §B.2. Existence of stochastic processes 275 §B.3. Conditional expectation 276 §B.4. Weak topology of probability measures 278 §B.5. First limit theorems 282 §B.6. Ergodic theory 282 §B.7. Stochastic ordering 288 Appendix C. Inequalities from statistical mechanics 293 §C.1. Griffiths’s inequality 293 §C.2. Griffiths-Hurst-Sherman inequality 294 Appendix D. Nonnegative matrices 297 Bibliography 299 Notation index 305 Author index 311 General index 313 Preface This book arose from courses on large deviations and related topics given by the authors in the Departments of Mathematics at the Ohio State University (1993), at the University of Wisconsin-Madison (2006 and 2013), and at the University of Utah (2008 and 2013). Our goal has been to create an attractive collection of material for a semester’s course which would also serve the broader needs of students from different fields. This goal has had two implications for the book. (1) We have not aimed at anything like an encyclopedic coverage of different techniques for proving large deviation principles (LDPs). Part I of the book focuses on one classic line of reasoning: (i) upper bound by an exponential Markov-Chebyshev inequality, (ii) lower bound by a change of measure, and (iii) an argument to match the rates from (i) and (ii). Beyond this technique Part I covers Bryc’s theorem and proves Cram´er’s theorem with the subadditive method. Part III of the book covers the G¨artner-Ellis theorem and an approach based on the convexity of a local rate function due to Baxter and Jain. (2) We have not felt obligated to stay within the boundaries of large deviation theory but instead follow the trail of some interesting material. Large deviation theory is a natural gateway to statistical mechanics. A dis- cussion of statistical mechanics would be incomplete without some study of phase transitions. We prove the phase transition of the Ising model in two different ways: (i) first with classical techniques: the Peierls argument, Dobrushin’s uniqueness condition, and correlation inequalities and (ii) the second time with random cluster measures. This means leaving large de- viation theory completely behind. Along the way we have the opportunity to learn coupling methods which are central to modern probability theory xi xii Preface but do not get serious application in the typical first graduate course in probability. We give now a brief overview of the contents of the book. Part I covers core general large deviation theory, the relevant convex analysis, and the large deviations of independent and identically distributed (i.i.d.) processes on three levels: Cram´er’s theorem, Sanov’s theorem, and the process level LDP for i.i.d. variables indexed by a multidimensional square lattice. Part II introduces Gibbs measures and proves the Dobrushin-Lanford- Ruelle variational principle that characterizes translation-invariant Gibbs measures. After this we study the phase transition of the Ising model. Part II ends with a chapter on the Fortuin-Kasteleyn random cluster model and the percolation approach to Ising phase transition. Part III develops the large deviation themes of Part I in several direc- tions. Large deviations of i.i.d. variables are complemented with moderate deviations and with more precise large deviation asymptotics. The G¨artner- Ellis theorem is developed carefully, together with the necessary additional convex analysis beyond the basics covered in Part I. From large deviations of i.i.d. processes we move on to Markov chains, to nonstationary indepen- dent random variables, and finally to random walk in a dynamical random environment. The last two topics give us an opportunity to apply another approach to proving large deviation principles, namely the Baxter-Jain the- orem.
Recommended publications
  • Pdf File of Second Edition, January 2018
    Probability on Graphs Random Processes on Graphs and Lattices Second Edition, 2018 GEOFFREY GRIMMETT Statistical Laboratory University of Cambridge copyright Geoffrey Grimmett Geoffrey Grimmett Statistical Laboratory Centre for Mathematical Sciences University of Cambridge Wilberforce Road Cambridge CB3 0WB United Kingdom 2000 MSC: (Primary) 60K35, 82B20, (Secondary) 05C80, 82B43, 82C22 With 56 Figures copyright Geoffrey Grimmett Contents Preface ix 1 Random Walks on Graphs 1 1.1 Random Walks and Reversible Markov Chains 1 1.2 Electrical Networks 3 1.3 FlowsandEnergy 8 1.4 RecurrenceandResistance 11 1.5 Polya's Theorem 14 1.6 GraphTheory 16 1.7 Exercises 18 2 Uniform Spanning Tree 21 2.1 De®nition 21 2.2 Wilson's Algorithm 23 2.3 Weak Limits on Lattices 28 2.4 Uniform Forest 31 2.5 Schramm±LownerEvolutionsÈ 32 2.6 Exercises 36 3 Percolation and Self-Avoiding Walks 39 3.1 PercolationandPhaseTransition 39 3.2 Self-Avoiding Walks 42 3.3 ConnectiveConstantoftheHexagonalLattice 45 3.4 CoupledPercolation 53 3.5 Oriented Percolation 53 3.6 Exercises 56 4 Association and In¯uence 59 4.1 Holley Inequality 59 4.2 FKG Inequality 62 4.3 BK Inequalitycopyright Geoffrey Grimmett63 vi Contents 4.4 HoeffdingInequality 65 4.5 In¯uenceforProductMeasures 67 4.6 ProofsofIn¯uenceTheorems 72 4.7 Russo'sFormulaandSharpThresholds 80 4.8 Exercises 83 5 Further Percolation 86 5.1 Subcritical Phase 86 5.2 Supercritical Phase 90 5.3 UniquenessoftheIn®niteCluster 96 5.4 Phase Transition 99 5.5 OpenPathsinAnnuli 103 5.6 The Critical Probability in Two Dimensions 107
    [Show full text]
  • Probability on Graphs Random Processes on Graphs and Lattices
    Probability on Graphs Random Processes on Graphs and Lattices GEOFFREY GRIMMETT Statistical Laboratory University of Cambridge c G. R. Grimmett 1/4/10, 17/11/10, 5/7/12 Geoffrey Grimmett Statistical Laboratory Centre for Mathematical Sciences University of Cambridge Wilberforce Road Cambridge CB3 0WB United Kingdom 2000 MSC: (Primary) 60K35, 82B20, (Secondary) 05C80, 82B43, 82C22 With 44 Figures c G. R. Grimmett 1/4/10, 17/11/10, 5/7/12 Contents Preface ix 1 Random walks on graphs 1 1.1 RandomwalksandreversibleMarkovchains 1 1.2 Electrical networks 3 1.3 Flowsandenergy 8 1.4 Recurrenceandresistance 11 1.5 Polya's theorem 14 1.6 Graphtheory 16 1.7 Exercises 18 2 Uniform spanning tree 21 2.1 De®nition 21 2.2 Wilson's algorithm 23 2.3 Weak limits on lattices 28 2.4 Uniform forest 31 2.5 Schramm±LownerevolutionsÈ 32 2.6 Exercises 37 3 Percolation and self-avoiding walk 39 3.1 Percolationandphasetransition 39 3.2 Self-avoiding walks 42 3.3 Coupledpercolation 45 3.4 Orientedpercolation 45 3.5 Exercises 48 4 Association and in¯uence 50 4.1 Holley inequality 50 4.2 FKGinequality 53 4.3 BK inequality 54 4.4 Hoeffdinginequality 56 c G. R. Grimmett 1/4/10, 17/11/10, 5/7/12 vi Contents 4.5 In¯uenceforproductmeasures 58 4.6 Proofsofin¯uencetheorems 63 4.7 Russo'sformulaandsharpthresholds 75 4.8 Exercises 78 5 Further percolation 81 5.1 Subcritical phase 81 5.2 Supercritical phase 86 5.3 Uniquenessofthein®nitecluster 92 5.4 Phase transition 95 5.5 Openpathsinannuli 99 5.6 The critical probability in two dimensions 103 5.7 Cardy's formula 110 5.8 The
    [Show full text]
  • Chapter on Gibbs State
    7 Gibbs states Brook’s theorem states that a positive probability measure on a finite product may be decomposed into factors indexed by the cliquesofits dependency graph. Closely related to this is the well known fact that apositivemeasureisaspatialMarkovfieldonagraphG if and only if it is a Gibbs state. The Ising and Potts models are introduced, and the n-vector model is mentioned. 7.1 Dependency graphs Let X (X1, X2,...,Xn) be a family of random variables on a given probability= space. For i, j V 1, 2,...,n with i j,wewritei j ∈ ={ } "= ⊥ if: X and X are independent conditional on (X : k i, j).Therelation i j k "= is thus symmetric, and it gives rise to a graph G with vertex set V and edge-set⊥ E i, j : i j ,calledthedependency graph of X (or of its ={$ % "⊥ } law). We shall see that the law of X may be expressed as a product over terms corresponding to complete subgraphs of G.Acompletesubgraphof G is called a clique,andwewriteK for the set of all cliques of G.For notational simplicity later, we designate the empty subset of V to be a clique, and thus ∅ K.Acliqueismaximal if no strict superset is a clique, and ∈ we write M for the set of maximal cliques of G. Weassumeforsimplicitythatthe Xi take values in some countable subset S of the reals R.ThelawofX gives rise to a probability mass function π on Sn given by π(x) P(X x for i V ), x (x , x ,...,x ) Sn. = i = i ∈ = 1 2 n ∈ It is easily seen by the definition of independence that i j if and only if π may be factorized in the form ⊥ π(x) g(x , U)h(x , U), x Sn, = i j ∈ for some functions g and h,whereU (x : k i, j).ForK K and = k "= ∈ x Sn,wewritex (x : i K ).Wecallπ positive if π(x)>0forall ∈ K = i ∈ x Sn.
    [Show full text]
  • Physics 132- Fundamentals of Physics for Biologists II
    Physics 132- Fundamentals of Physics for Biologists II Statistical Physics and Thermodynamics QUIZ 2 Quiz 2 25 20 15 10 Numberof Students 5 AVG: 5.15 0 STDEV: 2.17 0 2 4 6 8 10 Score 1. (4 pts) A 200 g block of copper at a temperature of 55 oC is put into an insulated beaker of water at 20 oC. The two come to thermal equilibrium at a temperature of about 30 oC – much closer to the original temperature of the water than of the copper. (From the macroscopic point of view, the specific heat of water is about 1.0 J/g-oC while the specific heat of copper is only about 0.4 J/g-oC.) However, from the microscopic point of view… 1.1 (2 pts) From this you can conclude A. There are more degrees of freedom in 200 g of copper than in 200 g of water. B. There are fewer degrees of freedom in 200 g of copper than in 200 g of water. C. There are about the same number of degrees of freedom in 200 g of copper as there are in 200 g of water. D. The information given doesn’t tell you anything about the number of degrees of freedom in the matter. 1.2 (2 pts) From this you can conclude the following about the kinetic energy associated with moving in the x-direction (after the system has come to thermal equilibrium ): A. The average x-direction contribution to kinetic energy of a water molecule is greater than that of a copper atom.
    [Show full text]
  • Hidden Gibbs Models: Theory and Applications
    Hidden Gibbs Models: Theory and Applications – DRAFT – Evgeny Verbitskiy Mathematical Institute, Leiden University, The Netherlands Johann Bernoulli Institute, Groningen University, The Netherlands November 21, 2015 2 Contents 1 Introduction1 1.1 Hidden Models / Hidden Processes............... 3 1.1.1 Equivalence of “random” and “deterministic” settings. 4 1.1.2 Deterministic chaos..................... 4 1.2 Hidden Gibbs Models....................... 6 1.2.1 How realistic is the Gibbs assumption?......... 8 I Theory9 2 Gibbs states 11 2.1 Gibbs-Bolztmann Ansatz..................... 11 2.2 Gibbs states for Lattice Systems................. 12 2.2.1 Translation Invariant Gibbs states............. 13 2.2.2 Regularity of Gibbs states................. 15 2.3 Gibbs and other regular stochastic processes.......... 16 2.3.1 Markov Chains....................... 17 2.3.2 k-step Markov chains.................... 18 2.3.3 Variable Length Markov Chains.............. 19 2.3.4 Countable Mixtures of Markov Chains (CMMC’s)... 20 2.3.5 g-measures (chains with complete connections).... 20 2.4 Gibbs measures in Dynamical Systems............. 27 2.4.1 Equilibrium states...................... 28 3 Functions of Markov processes 31 3.1 Markov Chains (recap)....................... 31 3.1.1 Regular measures on Subshifts of Finite Type...... 33 3.2 Function of Markov Processes.................. 33 3.2.1 Functions of Markov Chains in Dynamical systems.. 35 3.3 Hidden Markov Models...................... 36 3.4 Functions of Markov Chains from the Thermodynamic point of view................................ 38 3 4 CONTENTS 4 Hidden Gibbs Processes 45 4.1 Functions of Gibbs Processes................... 45 4.2 Yayama................................ 45 4.3 Review: renormalization of g-measures............. 47 4.4 Cluster Expansions........................
    [Show full text]