Lecture Notes for Stat 375 Inference in Graphical Models
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Bayesian Network - Wikipedia
2019/9/16 Bayesian network - Wikipedia Bayesian network A Bayesian network, Bayes network, belief network, decision network, Bayes(ian) model or probabilistic directed acyclic graphical model is a probabilistic graphical model (a type of statistical model) that represents a set of variables and their conditional dependencies via a directed acyclic graph (DAG). Bayesian networks are ideal for taking an event that occurred and predicting the likelihood that any one of several possible known causes was the contributing factor. For example, a Bayesian network A simple Bayesian network. Rain could represent the probabilistic relationships between diseases and influences whether the sprinkler symptoms. Given symptoms, the network can be used to compute the is activated, and both rain and probabilities of the presence of various diseases. the sprinkler influence whether the grass is wet. Efficient algorithms can perform inference and learning in Bayesian networks. Bayesian networks that model sequences of variables (e.g. speech signals or protein sequences) are called dynamic Bayesian networks. Generalizations of Bayesian networks that can represent and solve decision problems under uncertainty are called influence diagrams. Contents Graphical model Example Inference and learning Inferring unobserved variables Parameter learning Structure learning Statistical introduction Introductory examples Restrictions on priors Definitions and concepts Factorization definition Local Markov property Developing Bayesian networks Markov blanket Causal networks Inference complexity and approximation algorithms Software History See also Notes References https://en.wikipedia.org/wiki/Bayesian_network 1/14 2019/9/16 Bayesian network - Wikipedia Further reading External links Graphical model Formally, Bayesian networks are DAGs whose nodes represent variables in the Bayesian sense: they may be observable quantities, latent variables, unknown parameters or hypotheses. -
The Ising Model
The Ising Model Today we will switch topics and discuss one of the most studied models in statistical physics the Ising Model • Some applications: – Magnetism (the original application) – Liquid-gas transition – Binary alloys (can be generalized to multiple components) • Onsager solved the 2D square lattice (1D is easy!) • Used to develop renormalization group theory of phase transitions in 1970’s. • Critical slowing down and “cluster methods”. Figures from Landau and Binder (LB), MC Simulations in Statistical Physics, 2000. Atomic Scale Simulation 1 The Model • Consider a lattice with L2 sites and their connectivity (e.g. a square lattice). • Each lattice site has a single spin variable: si = ±1. • With magnetic field h, the energy is: N −β H H = −∑ Jijsis j − ∑ hisi and Z = ∑ e ( i, j ) i=1 • J is the nearest neighbors (i,j) coupling: – J > 0 ferromagnetic. – J < 0 antiferromagnetic. • Picture of spins at the critical temperature Tc. (Note that connected (percolated) clusters.) Atomic Scale Simulation 2 Mapping liquid-gas to Ising • For liquid-gas transition let n(r) be the density at lattice site r and have two values n(r)=(0,1). E = ∑ vijnin j + µ∑ni (i, j) i • Let’s map this into the Ising model spin variables: 1 s = 2n − 1 or n = s + 1 2 ( ) v v + µ H s s ( ) s c = ∑ i j + ∑ i + 4 (i, j) 2 i J = −v / 4 h = −(v + µ) / 2 1 1 1 M = s n = n = M + 1 N ∑ i N ∑ i 2 ( ) i i Atomic Scale Simulation 3 JAVA Ising applet http://physics.weber.edu/schroeder/software/demos/IsingModel.html Dynamically runs using heat bath algorithm. -
An Introduction to Factor Graphs
An Introduction to Factor Graphs Hans-Andrea Loeliger MLSB 2008, Copenhagen 1 Definition A factor graph represents the factorization of a function of several variables. We use Forney-style factor graphs (Forney, 2001). Example: f(x1, x2, x3, x4, x5) = fA(x1, x2, x3) · fB(x3, x4, x5) · fC(x4). fA fB x1 x3 x5 x2 x4 fC Rules: • A node for every factor. • An edge or half-edge for every variable. • Node g is connected to edge x iff variable x appears in factor g. (What if some variable appears in more than 2 factors?) 2 Example: Markov Chain pXYZ(x, y, z) = pX(x) pY |X(y|x) pZ|Y (z|y). X Y Z pX pY |X pZ|Y We will often use capital letters for the variables. (Why?) Further examples will come later. 3 Message Passing Algorithms operate by passing messages along the edges of a factor graph: - - - 6 6 ? ? - - - - - ... 6 6 ? ? 6 6 ? ? 4 A main point of factor graphs (and similar graphical notations): A Unified View of Historically Different Things Statistical physics: - Markov random fields (Ising 1925) Signal processing: - linear state-space models and Kalman filtering: Kalman 1960. - recursive least-squares adaptive filters - Hidden Markov models: Baum et al. 1966. - unification: Levy et al. 1996. Error correcting codes: - Low-density parity check codes: Gallager 1962; Tanner 1981; MacKay 1996; Luby et al. 1998. - Convolutional codes and Viterbi decoding: Forney 1973. - Turbo codes: Berrou et al. 1993. Machine learning, statistics: - Bayesian networks: Pearl 1988; Shachter 1988; Lauritzen and Spiegelhalter 1988; Shafer and Shenoy 1990. 5 Other Notation Systems for Graphical Models Example: p(u, w, x, y, z) = p(u)p(w)p(x|u, w)p(y|x)p(z|x). -
Arxiv:1511.03031V2
submitted to acta physica slovaca 1– 133 A BRIEF ACCOUNT OF THE ISING AND ISING-LIKE MODELS: MEAN-FIELD, EFFECTIVE-FIELD AND EXACT RESULTS Jozef Streckaˇ 1, Michal Jasˇcurˇ 2 Department of Theoretical Physics and Astrophysics, Faculty of Science, P. J. Saf´arikˇ University, Park Angelinum 9, 040 01 Koˇsice, Slovakia The present article provides a tutorial review on how to treat the Ising and Ising-like models within the mean-field, effective-field and exact methods. The mean-field approach is illus- trated on four particular examples of the lattice-statistical models: the spin-1/2 Ising model in a longitudinal field, the spin-1 Blume-Capel model in a longitudinal field, the mixed-spin Ising model in a longitudinal field and the spin-S Ising model in a transverse field. The mean- field solutions of the spin-1 Blume-Capel model and the mixed-spin Ising model demonstrate a change of continuous phase transitions to discontinuous ones at a tricritical point. A con- tinuous quantum phase transition of the spin-S Ising model driven by a transverse magnetic field is also explored within the mean-field method. The effective-field theory is elaborated within a single- and two-spin cluster approach in order to demonstrate an efficiency of this ap- proximate method, which affords superior approximate results with respect to the mean-field results. The long-standing problem of this method concerned with a self-consistent deter- mination of the free energy is also addressed in detail. More specifically, the effective-field theory is adapted for the spin-1/2 Ising model in a longitudinal field, the spin-S Blume-Capel model in a longitudinal field and the spin-1/2 Ising model in a transverse field. -
Statistical Field Theory University of Cambridge Part III Mathematical Tripos
Preprint typeset in JHEP style - HYPER VERSION Michaelmas Term, 2017 Statistical Field Theory University of Cambridge Part III Mathematical Tripos David Tong Department of Applied Mathematics and Theoretical Physics, Centre for Mathematical Sciences, Wilberforce Road, Cambridge, CB3 OBA, UK http://www.damtp.cam.ac.uk/user/tong/sft.html [email protected] –1– Recommended Books and Resources There are a large number of books which cover the material in these lectures, although often from very di↵erent perspectives. They have titles like “Critical Phenomena”, “Phase Transitions”, “Renormalisation Group” or, less helpfully, “Advanced Statistical Mechanics”. Here are some that I particularly like Nigel Goldenfeld, Phase Transitions and the Renormalization Group • Agreatbook,coveringthebasicmaterialthatwe’llneedanddelvingdeeperinplaces. Mehran Kardar, Statistical Physics of Fields • The second of two volumes on statistical mechanics. It cuts a concise path through the subject, at the expense of being a little telegraphic in places. It is based on lecture notes which you can find on the web; a link is given on the course website. John Cardy, Scaling and Renormalisation in Statistical Physics • Abeautifullittlebookfromoneofthemastersofconformalfieldtheory.Itcoversthe material from a slightly di↵erent perspective than these lectures, with more focus on renormalisation in real space. Chaikin and Lubensky, Principles of Condensed Matter Physics • Shankar, Quantum Field Theory and Condensed Matter • Both of these are more all-round condensed matter books, but with substantial sections on critical phenomena and the renormalisation group. Chaikin and Lubensky is more traditional, and packed full of content. Shankar covers modern methods of QFT, with an easygoing style suitable for bedtime reading. Anumberofexcellentlecturenotesareavailableontheweb.Linkscanbefoundon the course webpage: http://www.damtp.cam.ac.uk/user/tong/sft.html. -
Conditional Random Fields
Conditional Random Fields Grant “Tractable Inference” Van Horn and Milan “No Marginal” Cvitkovic Recap: Discriminative vs. Generative Models Suppose we have a dataset drawn from Generative models try to learn Hard problem; always requires assumptions Captures all the nuance of the data distribution (modulo our assumptions) Discriminative models try to learn Simpler problem; often all we care about Typically doesn’t let us make useful modeling assumptions about Recap: Graphical Models One way to make generative modeling tractable is to make simplifying assumptions about which variables affect which other variables. These assumptions can often be represented graphically, leading to a whole zoo of models known collectively as Graphical Models. Recap: Graphical Models One zoo-member is the Bayesian Network, which you’ve met before: Naive Bayes HMM The vertices in a Bayes net are random variables. An edge from A to B means roughly: “A causes B” Or more formally: “B is independent of all other vertices when conditioned on its parents”. Figures from Sutton and McCallum, ‘12 Recap: Graphical Models Another is the Markov Random Field (MRF), an undirected graphical model Again vertices are random variables. Edges show which variables depend on each other: Thm: This implies that Where the are called “factors”. They measure how compatible an assignment of subset of random variables are. Figure from Wikipedia Recap: Graphical Models Then there are Factor Graphs, which are just MRFs with bonus vertices to remove any ambiguity about exactly how the MRF factors An MRF An MRF A factor graph for this MRF Another factor graph for this MRF Circles and squares are both vertices; circles are random variables, squares are factors. -
Lecture 8: the Ising Model Introduction
Lecture 8: The Ising model Introduction I Up to now: Toy systems with interesting properties (random walkers, cluster growth, percolation) I Common to them: No interactions I Add interactions now, with significant role I Immediate consequence: much richer structure in model, in particular: phase transitions I Simulate interactions with RNG (Monte Carlo method) I Include the impact of temperature: ideas from thermodynamics and statistical mechanics important I Simple system as example: coupled spins (see below), will use the canonical ensemble for its description The Ising model I A very interesting model for understanding some properties of magnetic materials, especially the phase transition ferromagnetic ! paramagnetic I Intrinsically, magnetism is a quantum effect, triggered by the spins of particles aligning with each other I Ising model a superb toy model to understand this dynamics I Has been invented in the 1920's by E.Ising I Ever since treated as a first, paradigmatic model The model (in 2 dimensions) I Consider a square lattice with spins at each lattice site I Spins can have two values: si = ±1 I Take into account only nearest neighbour interactions (good approximation, dipole strength falls off as 1=r 3) I Energy of the system: P E = −J si sj hiji I Here: exchange constant J > 0 (for ferromagnets), and hiji denotes pairs of nearest neighbours. I (Micro-)states α characterised by the configuration of each spin, the more aligned the spins in a state α, the smaller the respective energy Eα. I Emergence of spontaneous magnetisation (without external field): sufficiently many spins parallel Adding temperature I Without temperature T : Story over. -
Arxiv:1504.02898V2 [Cond-Mat.Stat-Mech] 7 Jun 2015 Keywords: Percolation, Explosive Percolation, SLE, Ising Model, Earth Topography
Recent advances in percolation theory and its applications Abbas Ali Saberi aDepartment of Physics, University of Tehran, P.O. Box 14395-547,Tehran, Iran bSchool of Particles and Accelerators, Institute for Research in Fundamental Sciences (IPM) P.O. Box 19395-5531, Tehran, Iran Abstract Percolation is the simplest fundamental model in statistical mechanics that exhibits phase transitions signaled by the emergence of a giant connected component. Despite its very simple rules, percolation theory has successfully been applied to describe a large variety of natural, technological and social systems. Percolation models serve as important universality classes in critical phenomena characterized by a set of critical exponents which correspond to a rich fractal and scaling structure of their geometric features. We will first outline the basic features of the ordinary model. Over the years a variety of percolation models has been introduced some of which with completely different scaling and universal properties from the original model with either continuous or discontinuous transitions depending on the control parameter, di- mensionality and the type of the underlying rules and networks. We will try to take a glimpse at a number of selective variations including Achlioptas process, half-restricted process and spanning cluster-avoiding process as examples of the so-called explosive per- colation. We will also introduce non-self-averaging percolation and discuss correlated percolation and bootstrap percolation with special emphasis on their recent progress. Directed percolation process will be also discussed as a prototype of systems displaying a nonequilibrium phase transition into an absorbing state. In the past decade, after the invention of stochastic L¨ownerevolution (SLE) by Oded Schramm, two-dimensional (2D) percolation has become a central problem in probability theory leading to the two recent Fields medals. -
Multiple Schramm-Loewner Evolutions and Statistical Mechanics Martingales
Multiple Schramm-Loewner Evolutions and Statistical Mechanics Martingales Michel Bauer a, Denis Bernard a 1, Kalle Kyt¨ol¨a b [email protected]; [email protected]; [email protected] a Service de Physique Th´eorique de Saclay CEA/DSM/SPhT, Unit´ede recherche associ´ee au CNRS CEA-Saclay, 91191 Gif-sur-Yvette, France. b Department of Mathematics, P.O. Box 68 FIN-00014 University of Helsinki, Finland. Abstract A statistical mechanics argument relating partition functions to martingales is used to get a condition under which random geometric processes can describe interfaces in 2d statistical mechanics at critical- ity. Requiring multiple SLEs to satisfy this condition leads to some natural processes, which we study in this note. We give examples of such multiple SLEs and discuss how a choice of conformal block is related to geometric configuration of the interfaces and what is the physical meaning of mixed conformal blocks. We illustrate the general arXiv:math-ph/0503024v2 21 Mar 2005 ideas on concrete computations, with applications to percolation and the Ising model. 1Member of C.N.R.S 1 Contents 1 Introduction 3 2 Basics of Schramm-L¨owner evolutions: Chordal SLE 6 3 A proposal for multiple SLEs 7 3.1 Thebasicequations ....................... 7 3.2 Archprobabilities......................... 8 4 First comments 10 4.1 Statistical mechanics interpretation . 10 4.2 SLEasaspecialcaseof2SLE. 10 4.3 Makingsense ........................... 12 4.4 A few martingales for nSLEs .................. 13 4.5 Classicallimit........................... 14 4.6 Relations with other work . 15 5 CFT background 16 6 Martingales from statistical mechanics 19 6.1 Tautological martingales . -
Markov Random Fields and Stochastic Image Models
Markov Random Fields and Stochastic Image Models Charles A. Bouman School of Electrical and Computer Engineering Purdue University Phone: (317) 494-0340 Fax: (317) 494-3358 email [email protected] Available from: http://dynamo.ecn.purdue.edu/»bouman/ Tutorial Presented at: 1995 IEEE International Conference on Image Processing 23-26 October 1995 Washington, D.C. Special thanks to: Ken Sauer Suhail Saquib Department of Electrical School of Electrical and Computer Engineering Engineering University of Notre Dame Purdue University 1 Overview of Topics 1. Introduction (b) Non-Gaussian MRF's 2. The Bayesian Approach i. Quadratic functions ii. Non-Convex functions 3. Discrete Models iii. Continuous MAP estimation (a) Markov Chains iv. Convex functions (b) Markov Random Fields (MRF) (c) Parameter Estimation (c) Simulation i. Estimation of σ (d) Parameter estimation ii. Estimation of T and p parameters 4. Application of MRF's to Segmentation 6. Application to Tomography (a) The Model (a) Tomographic system and data models (b) Bayesian Estimation (b) MAP Optimization (c) MAP Optimization (c) Parameter estimation (d) Parameter Estimation 7. Multiscale Stochastic Models (e) Other Approaches (a) Continuous models 5. Continuous Models (b) Discrete models (a) Gaussian Random Process Models 8. High Level Image Models i. Autoregressive (AR) models ii. Simultaneous AR (SAR) models iii. Gaussian MRF's iv. Generalization to 2-D 2 References in Statistical Image Modeling 1. Overview references [100, 89, 50, 54, 162, 4, 44] 4. Simulation and Stochastic Optimization Methods [118, 80, 129, 100, 68, 141, 61, 76, 62, 63] 2. Type of Random Field Model 5. Computational Methods used with MRF Models (a) Discrete Models i. -
Percolation Theory Applied to Financial Markets: a Cluster Description of Herding Behavior Leading to Bubbles and Crashes
Percolation Theory Applied to Financial Markets: A Cluster Description of Herding Behavior Leading to Bubbles and Crashes Master's Thesis Maximilian G. A. Seyrich February 20, 2015 Advisor: Professor Didier Sornette Chair of Entrepreneurial Risk Department of Physics Department MTEC ETH Zurich Abstract In this thesis, we clarify the herding effect due to cluster dynamics of traders. We provide a framework which is able to derive the crash hazard rate of the Johannsen-Ledoit-Sornette model (JLS) as a power law diverging function of the percolation scaling parameter. Using this framework, we are able to create reasonable bubbles and crashes in price time series. We present a variety of different kinds of bubbles and crashes and provide insights into the dynamics of financial mar- kets. Our simulations show that most of the time, a crash is preceded by a bubble. Yet, a bubble must not end with a crash. The time of a crash is a random variable, being simulated by a Poisson process. The crash hazard rate plays the role of the intensity function. The headstone of this thesis is a new description of the crash hazard rate. Recently dis- covered super-linear relations between groups sizes and group activi- ties, as well as percolation theory, are applied to the Cont-Bouchaud cluster dynamics model. i Contents Contents iii 1 Motivation1 2 Models & Methods3 2.1 The Johanson-Ledoit-Sornette (JLS) Model........... 3 2.2 Percolation theory.......................... 6 2.2.1 Introduction......................... 6 2.2.2 Percolation Point and Cluster Number......... 7 2.2.3 Finite Size Effects...................... 8 2.3 Ising Model ............................ -
8 Blocks, Markov Chains, Ising Model
Tel Aviv University, 2007 Large deviations 60 8 Blocks, Markov chains, Ising model 8a Introductory remarks . 60 8b Pair frequencies: combinatorial approach . 63 8c Markovchains .................... 66 8d Ising model (one-dimensional) . 68 8e Pair frequencies: linear algebra approach . 69 8f Dimension two . 70 8a Introductory remarks The following three questions are related more closely than it may seem. 8a1 Question. 100 children stay in a ring, 40 boys and 60 girls. Among the 100 pairs of neighbors, 20 pairs are heterosexual (a girl and a boy); others are not. What about the number of all such configurations? 8a2 Question. A Markov chain with two states (0 and 1) is given via its 2 2-matrix of transition probabilities. What about the probability that the × state 1 occurs 60 times among the first 100? 8a3 Question. (Ising model) A one-dimensional array of n spin-1/2 par- ticles is described by the configuration space 1, 1 n. Each configuration (s ,...,s ) 1, 1 n has its energy {− } 1 n ∈ {− } 1 H (s ,...,s )= (s s + + s s ) h(s + + s ) ; n 1 n −2 1 2 ··· n−1 n − 1 ··· n here h R is a parameter. (It is the strength of an external magnetic field, ∈ while the strength of the nearest neighbor coupling is set to 1.) What about the dependence of the energy and the mean spin (s + + s )/n on h and 1 ··· n the temperature? Tossing a fair coin n times we get a random element (β ,...,β ) of 0, 1 n, 1 n { } and may consider the n 1 pairs (β1, β2), (β2, β3),..., (βn−1, βn).