The Ising Model in Psychometrics

Total Page:16

File Type:pdf, Size:1020Kb

The Ising Model in Psychometrics Chapter 8 The Ising Model in Psychometrics Abstract This chapter provides a general introduction of network modeling in psychometrics. The chapter starts with an introduction to the statistical model formulation of pairwise Markov random fields (PMRF), followed by an introduction of the PMRF suitable for binary data: the Ising model. The Ising model is a model used in ferromagnetism to explain phase transitions in a field of particles. Following the description of the Ising model in statistical physics, the chapter continues to show that the Ising model is closely related to models used in psychometrics. The Ising model can be shown to be equivalent to certain kinds of logistic regression models, loglinear models and multi-dimensional item response theory (MIRT) models. The equivalence between the Ising model and the MIRT model puts standard psychometrics in a new light and leads to a strikingly di↵erent interpretation of well-known latent variable models. The chapter gives an overview of methods that can be used to estimate the Ising model, and concludes with a discussion on the interpretation of latent variables given the equivalence between the Ising model and MIRT. 8.1 Introduction In recent years, network models have been proposed as an alternative way of looking at psychometric problems (Van Der Maas et al., 2006; Cramer et al., 2010; Borsboom & Cramer, 2013). In these models, psychometric item responses are conceived of as proxies for variables that directly interact with each other. For example, the symptoms of depression (such as loss of energy, sleep problems, and low self esteem) are traditionally thought of as being determined by a common latent variable (depression, or the liability to become depressed; Aggen, Neale, & This chapter has been adapted from: Epskamp, S., Maris, G., Waldorp, L.J., and Borsboom, D. (in press). Network Psychometrics. In Irwing, P., Hughes, D., and Booth, T. (Eds.), Handbook of Psychometrics. New York: Wiley. 143 8. The Ising Model in Psychometrics Kendler, 2005). In network models, these symptoms are instead hypothesized to form networks of mutually reinforcing variables (e.g., sleep problems may lead to loss of energy, which may lead to low self esteem, which may cause rumination that in turn may reinforce sleep problems). On the face of it, such network models o↵er an entirely di↵erent conceptualization of why psychometric variables cluster in the way that they do. However, it has also been suggested in the literature that latent variables may somehow correspond to sets of tightly intertwined observables (e.g., see the Appendix of Van Der Maas et al., 2006). In the current chapter, we aim to make this connection explicit. As we will show, a particular class of latent variable models (namely, multidimensional Item Response Theory models) yields exactly the same probability distribution over the observed variables as a particular class of network models (namely, Ising models). In the current chapter, we exploit the consequences of this equivalence. We will first introduce the general class of models used in network analysis called Markov Random Fields. Specifically, we will discuss the Markov random field for binary data called the Ising Model, which originated from statistical physics but has since been used in many fields of science. We will show how the Ising Model relates to psychometrical practice, with a focus on the equivalence between the Ising Model and multidimensional item response theory. We will demonstrate how the Ising model can be estimated and finally, we will discuss the conceptual implications of this equivalence. Notation Throughout this chapter we will denote random variables with capital letters and possible realizations with lower case letters; vectors will be represented with bold- faced letters. For parameters, we will use boldfaced capital letters to indicate matrices instead of vectors whereas for random variables we will use boldfaced capital letters to indicate a random vector. Roman letters will be used to denote observable variables and parameters (such as the number of nodes) and Greek letters will be used to denote unobservable variables and parameters that need to be estimated. In this chapter we will mainly model the random vector X: X> = X1 X2 ... XP , containing P binary variables that⇥ take the values 1⇤ (e.g., correct, true or yes) and 1 (e.g., incorrect, false or no). We will denote a realization, or state, of − X with x> = x1 x2 ... xp . Let N be the number of observations and n(x) the number of observations that have response pattern x. Furthermore, let i denote the⇥ subscript of a random⇤ variable and j the subscript of a dif- ferent random variable (j = i). Thus, Xi is the ith random variable and xi its realization. The superscript6 (...) will indicate that elements are removed −(i) from a vector; for example, X− indicates the random vector X without Xi: (i) (i) X− = X1,...,Xi 1,Xi+1,....XP , and x− indicates its realization. Sim- − (i,j) X (i,j) ilarly, X−⇥ indicates X without X⇤ i and Xj and x− its realization. An overview of all notations used in this chapter can be seen in Appendix B. 144 8.2. Markov Random Fields X2 X1 X3 Figure 8.1: Example of a PMRF of three nodes, X1, X2 and X3 , connected by two edges, one between X1 and X2 and one between X2 and X3. 8.2 Markov Random Fields A network, also called a graph, can be encoded as a set G consisting of two sets: V , which contains the nodes in the network, and E, which contains the edges that connect these nodes. For example, the graph in Figure 8.1 contains three nodes: V = 1, 2, 3 , which are connected by two edges: E = (1, 2), (2, 3) . We will use this{ type} of network to represent a pairwise Markov{ random field} (PMRF; Lauritzen, 1996; Murphy, 2012), in which nodes represent observed ran- dom variables1 and edges represent (conditional) association between two nodes. More importantly, the absence of an edge represents the Markov property that two nodes are conditionally independent given all other nodes in the network: (i,j) (i,j) X X X− = x− (i, j) E (8.1) i ?? j | () 62 Thus, a PMRF encodes the independence structure of the system of nodes. In the case of Figure 8.1, X1 and X3 are independent given that we know X2 = x2. This could be due to several reasons; there might be a causal path from X1 to X3 or vise versa, X2 might be the common cause of X1 and X3, unobserved variables might cause the dependencies between X1 and X2 and X2 and X3, or the edges in the network might indicate actual pairwise interactions between X1 and X2 and X2 and X3. Of particular interest to psychometrics are models in which the presence of latent common causes induces associations among the observed variables. If such a common cause model holds, we cannot condition on any observed variable to completely remove the association between two nodes (Pearl, 2000). Thus, if an unobserved variable acts as a common cause to some of the observed variables, we should find a fully connected clique in the PMRF that describes the associations 1Throughout this chapter, nodes in a network designate variables, hence the terms are used interchangeably. 145 8. The Ising Model in Psychometrics among these nodes. The network in Figure 8.1, for example, cannot represent associations between three nodes that are subject to the influence of a latent common cause; if that were the case, it would be impossible to obtain conditional independence between X1 and X3 by conditioning on X2. Parameterizing Markov Random Fields A PMRF can be parameterized as a product of strictly positive potential functions φ(x) (Murphy, 2012): 1 Pr (X = x)= φ (x ) φ (x ,x ) , (8.2) Z i i ij i j i <ij> Y Y in which i takes the product over all nodes, i =1, 2,...,P, <ij> takes the product over all distinct pairs of nodes i and j (j>i), and Z is a normalizing constant suchQ that the probability function sums to unity over allQ possible patterns of observations in the sample space: Z = φi (xi) φij (xi,xj) . x i <ij> X Y Y Here, x takes the sum over all possible realizations of X. All φ(x) functions result in positive real numbers, which encode the potentials: the preference for P the relevant part of X to be in some state. The φi(xi) functions encode the node potentials of the network; the preference of node Xi to be in state xi, regardless of the state of the other nodes in the network. Thus, φi(xi) maps the potential for Xi to take the value xi regardless of the rest of the network. If φi(xi)=0, for instance, then Xi will never take the value xi, while φi(xi) = 1 indicates that there is no preference for X to take any particular value and φ (x )= indicates i i i 1 that the system always prefers Xi to take the value xi. The φij (xi,xj) functions encode the pairwise potentials of the network; the preference of nodes Xi and Xj to both be in states xi and xj. As φij (xi,xj) grows higher we would expect to observe Xj = xj whenever Xi = xi. Note that the potential functions are not identified; we can multiply both φi(xi) or φij (xi,xj) with some constant for all possible outcomes of xi, in which case this constant becomes a constant multiplier to (8.2) and is cancelled out in the normalizing constant Z. A typical identification constraint on the potential functions is to set the marginal geometric means of all outcomes equal to 1; over all possible outcomes of each argument, the logarithm of each potential function should sum to 0: ln φ (x )= ln φ (x ,x )= ln φ (x ,x )=0 x ,x (8.3) i i ij i j ij i j 8 i j x x x Xi Xi Xj in which denotes the sum over all possible realizations for X , and xi i xj denotes the sum over all possible realizations of Xj.
Recommended publications
  • The Ising Model
    The Ising Model Today we will switch topics and discuss one of the most studied models in statistical physics the Ising Model • Some applications: – Magnetism (the original application) – Liquid-gas transition – Binary alloys (can be generalized to multiple components) • Onsager solved the 2D square lattice (1D is easy!) • Used to develop renormalization group theory of phase transitions in 1970’s. • Critical slowing down and “cluster methods”. Figures from Landau and Binder (LB), MC Simulations in Statistical Physics, 2000. Atomic Scale Simulation 1 The Model • Consider a lattice with L2 sites and their connectivity (e.g. a square lattice). • Each lattice site has a single spin variable: si = ±1. • With magnetic field h, the energy is: N −β H H = −∑ Jijsis j − ∑ hisi and Z = ∑ e ( i, j ) i=1 • J is the nearest neighbors (i,j) coupling: – J > 0 ferromagnetic. – J < 0 antiferromagnetic. • Picture of spins at the critical temperature Tc. (Note that connected (percolated) clusters.) Atomic Scale Simulation 2 Mapping liquid-gas to Ising • For liquid-gas transition let n(r) be the density at lattice site r and have two values n(r)=(0,1). E = ∑ vijnin j + µ∑ni (i, j) i • Let’s map this into the Ising model spin variables: 1 s = 2n − 1 or n = s + 1 2 ( ) v v + µ H s s ( ) s c = ∑ i j + ∑ i + 4 (i, j) 2 i J = −v / 4 h = −(v + µ) / 2 1 1 1 M = s n = n = M + 1 N ∑ i N ∑ i 2 ( ) i i Atomic Scale Simulation 3 JAVA Ising applet http://physics.weber.edu/schroeder/software/demos/IsingModel.html Dynamically runs using heat bath algorithm.
    [Show full text]
  • Arxiv:1511.03031V2
    submitted to acta physica slovaca 1– 133 A BRIEF ACCOUNT OF THE ISING AND ISING-LIKE MODELS: MEAN-FIELD, EFFECTIVE-FIELD AND EXACT RESULTS Jozef Streckaˇ 1, Michal Jasˇcurˇ 2 Department of Theoretical Physics and Astrophysics, Faculty of Science, P. J. Saf´arikˇ University, Park Angelinum 9, 040 01 Koˇsice, Slovakia The present article provides a tutorial review on how to treat the Ising and Ising-like models within the mean-field, effective-field and exact methods. The mean-field approach is illus- trated on four particular examples of the lattice-statistical models: the spin-1/2 Ising model in a longitudinal field, the spin-1 Blume-Capel model in a longitudinal field, the mixed-spin Ising model in a longitudinal field and the spin-S Ising model in a transverse field. The mean- field solutions of the spin-1 Blume-Capel model and the mixed-spin Ising model demonstrate a change of continuous phase transitions to discontinuous ones at a tricritical point. A con- tinuous quantum phase transition of the spin-S Ising model driven by a transverse magnetic field is also explored within the mean-field method. The effective-field theory is elaborated within a single- and two-spin cluster approach in order to demonstrate an efficiency of this ap- proximate method, which affords superior approximate results with respect to the mean-field results. The long-standing problem of this method concerned with a self-consistent deter- mination of the free energy is also addressed in detail. More specifically, the effective-field theory is adapted for the spin-1/2 Ising model in a longitudinal field, the spin-S Blume-Capel model in a longitudinal field and the spin-1/2 Ising model in a transverse field.
    [Show full text]
  • Statistical Field Theory University of Cambridge Part III Mathematical Tripos
    Preprint typeset in JHEP style - HYPER VERSION Michaelmas Term, 2017 Statistical Field Theory University of Cambridge Part III Mathematical Tripos David Tong Department of Applied Mathematics and Theoretical Physics, Centre for Mathematical Sciences, Wilberforce Road, Cambridge, CB3 OBA, UK http://www.damtp.cam.ac.uk/user/tong/sft.html [email protected] –1– Recommended Books and Resources There are a large number of books which cover the material in these lectures, although often from very di↵erent perspectives. They have titles like “Critical Phenomena”, “Phase Transitions”, “Renormalisation Group” or, less helpfully, “Advanced Statistical Mechanics”. Here are some that I particularly like Nigel Goldenfeld, Phase Transitions and the Renormalization Group • Agreatbook,coveringthebasicmaterialthatwe’llneedanddelvingdeeperinplaces. Mehran Kardar, Statistical Physics of Fields • The second of two volumes on statistical mechanics. It cuts a concise path through the subject, at the expense of being a little telegraphic in places. It is based on lecture notes which you can find on the web; a link is given on the course website. John Cardy, Scaling and Renormalisation in Statistical Physics • Abeautifullittlebookfromoneofthemastersofconformalfieldtheory.Itcoversthe material from a slightly di↵erent perspective than these lectures, with more focus on renormalisation in real space. Chaikin and Lubensky, Principles of Condensed Matter Physics • Shankar, Quantum Field Theory and Condensed Matter • Both of these are more all-round condensed matter books, but with substantial sections on critical phenomena and the renormalisation group. Chaikin and Lubensky is more traditional, and packed full of content. Shankar covers modern methods of QFT, with an easygoing style suitable for bedtime reading. Anumberofexcellentlecturenotesareavailableontheweb.Linkscanbefoundon the course webpage: http://www.damtp.cam.ac.uk/user/tong/sft.html.
    [Show full text]
  • Lecture 8: the Ising Model Introduction
    Lecture 8: The Ising model Introduction I Up to now: Toy systems with interesting properties (random walkers, cluster growth, percolation) I Common to them: No interactions I Add interactions now, with significant role I Immediate consequence: much richer structure in model, in particular: phase transitions I Simulate interactions with RNG (Monte Carlo method) I Include the impact of temperature: ideas from thermodynamics and statistical mechanics important I Simple system as example: coupled spins (see below), will use the canonical ensemble for its description The Ising model I A very interesting model for understanding some properties of magnetic materials, especially the phase transition ferromagnetic ! paramagnetic I Intrinsically, magnetism is a quantum effect, triggered by the spins of particles aligning with each other I Ising model a superb toy model to understand this dynamics I Has been invented in the 1920's by E.Ising I Ever since treated as a first, paradigmatic model The model (in 2 dimensions) I Consider a square lattice with spins at each lattice site I Spins can have two values: si = ±1 I Take into account only nearest neighbour interactions (good approximation, dipole strength falls off as 1=r 3) I Energy of the system: P E = −J si sj hiji I Here: exchange constant J > 0 (for ferromagnets), and hiji denotes pairs of nearest neighbours. I (Micro-)states α characterised by the configuration of each spin, the more aligned the spins in a state α, the smaller the respective energy Eα. I Emergence of spontaneous magnetisation (without external field): sufficiently many spins parallel Adding temperature I Without temperature T : Story over.
    [Show full text]
  • Arxiv:1504.02898V2 [Cond-Mat.Stat-Mech] 7 Jun 2015 Keywords: Percolation, Explosive Percolation, SLE, Ising Model, Earth Topography
    Recent advances in percolation theory and its applications Abbas Ali Saberi aDepartment of Physics, University of Tehran, P.O. Box 14395-547,Tehran, Iran bSchool of Particles and Accelerators, Institute for Research in Fundamental Sciences (IPM) P.O. Box 19395-5531, Tehran, Iran Abstract Percolation is the simplest fundamental model in statistical mechanics that exhibits phase transitions signaled by the emergence of a giant connected component. Despite its very simple rules, percolation theory has successfully been applied to describe a large variety of natural, technological and social systems. Percolation models serve as important universality classes in critical phenomena characterized by a set of critical exponents which correspond to a rich fractal and scaling structure of their geometric features. We will first outline the basic features of the ordinary model. Over the years a variety of percolation models has been introduced some of which with completely different scaling and universal properties from the original model with either continuous or discontinuous transitions depending on the control parameter, di- mensionality and the type of the underlying rules and networks. We will try to take a glimpse at a number of selective variations including Achlioptas process, half-restricted process and spanning cluster-avoiding process as examples of the so-called explosive per- colation. We will also introduce non-self-averaging percolation and discuss correlated percolation and bootstrap percolation with special emphasis on their recent progress. Directed percolation process will be also discussed as a prototype of systems displaying a nonequilibrium phase transition into an absorbing state. In the past decade, after the invention of stochastic L¨ownerevolution (SLE) by Oded Schramm, two-dimensional (2D) percolation has become a central problem in probability theory leading to the two recent Fields medals.
    [Show full text]
  • Multiple Schramm-Loewner Evolutions and Statistical Mechanics Martingales
    Multiple Schramm-Loewner Evolutions and Statistical Mechanics Martingales Michel Bauer a, Denis Bernard a 1, Kalle Kyt¨ol¨a b [email protected]; [email protected]; [email protected] a Service de Physique Th´eorique de Saclay CEA/DSM/SPhT, Unit´ede recherche associ´ee au CNRS CEA-Saclay, 91191 Gif-sur-Yvette, France. b Department of Mathematics, P.O. Box 68 FIN-00014 University of Helsinki, Finland. Abstract A statistical mechanics argument relating partition functions to martingales is used to get a condition under which random geometric processes can describe interfaces in 2d statistical mechanics at critical- ity. Requiring multiple SLEs to satisfy this condition leads to some natural processes, which we study in this note. We give examples of such multiple SLEs and discuss how a choice of conformal block is related to geometric configuration of the interfaces and what is the physical meaning of mixed conformal blocks. We illustrate the general arXiv:math-ph/0503024v2 21 Mar 2005 ideas on concrete computations, with applications to percolation and the Ising model. 1Member of C.N.R.S 1 Contents 1 Introduction 3 2 Basics of Schramm-L¨owner evolutions: Chordal SLE 6 3 A proposal for multiple SLEs 7 3.1 Thebasicequations ....................... 7 3.2 Archprobabilities......................... 8 4 First comments 10 4.1 Statistical mechanics interpretation . 10 4.2 SLEasaspecialcaseof2SLE. 10 4.3 Makingsense ........................... 12 4.4 A few martingales for nSLEs .................. 13 4.5 Classicallimit........................... 14 4.6 Relations with other work . 15 5 CFT background 16 6 Martingales from statistical mechanics 19 6.1 Tautological martingales .
    [Show full text]
  • Markov Random Fields and Stochastic Image Models
    Markov Random Fields and Stochastic Image Models Charles A. Bouman School of Electrical and Computer Engineering Purdue University Phone: (317) 494-0340 Fax: (317) 494-3358 email [email protected] Available from: http://dynamo.ecn.purdue.edu/»bouman/ Tutorial Presented at: 1995 IEEE International Conference on Image Processing 23-26 October 1995 Washington, D.C. Special thanks to: Ken Sauer Suhail Saquib Department of Electrical School of Electrical and Computer Engineering Engineering University of Notre Dame Purdue University 1 Overview of Topics 1. Introduction (b) Non-Gaussian MRF's 2. The Bayesian Approach i. Quadratic functions ii. Non-Convex functions 3. Discrete Models iii. Continuous MAP estimation (a) Markov Chains iv. Convex functions (b) Markov Random Fields (MRF) (c) Parameter Estimation (c) Simulation i. Estimation of σ (d) Parameter estimation ii. Estimation of T and p parameters 4. Application of MRF's to Segmentation 6. Application to Tomography (a) The Model (a) Tomographic system and data models (b) Bayesian Estimation (b) MAP Optimization (c) MAP Optimization (c) Parameter estimation (d) Parameter Estimation 7. Multiscale Stochastic Models (e) Other Approaches (a) Continuous models 5. Continuous Models (b) Discrete models (a) Gaussian Random Process Models 8. High Level Image Models i. Autoregressive (AR) models ii. Simultaneous AR (SAR) models iii. Gaussian MRF's iv. Generalization to 2-D 2 References in Statistical Image Modeling 1. Overview references [100, 89, 50, 54, 162, 4, 44] 4. Simulation and Stochastic Optimization Methods [118, 80, 129, 100, 68, 141, 61, 76, 62, 63] 2. Type of Random Field Model 5. Computational Methods used with MRF Models (a) Discrete Models i.
    [Show full text]
  • Percolation Theory Applied to Financial Markets: a Cluster Description of Herding Behavior Leading to Bubbles and Crashes
    Percolation Theory Applied to Financial Markets: A Cluster Description of Herding Behavior Leading to Bubbles and Crashes Master's Thesis Maximilian G. A. Seyrich February 20, 2015 Advisor: Professor Didier Sornette Chair of Entrepreneurial Risk Department of Physics Department MTEC ETH Zurich Abstract In this thesis, we clarify the herding effect due to cluster dynamics of traders. We provide a framework which is able to derive the crash hazard rate of the Johannsen-Ledoit-Sornette model (JLS) as a power law diverging function of the percolation scaling parameter. Using this framework, we are able to create reasonable bubbles and crashes in price time series. We present a variety of different kinds of bubbles and crashes and provide insights into the dynamics of financial mar- kets. Our simulations show that most of the time, a crash is preceded by a bubble. Yet, a bubble must not end with a crash. The time of a crash is a random variable, being simulated by a Poisson process. The crash hazard rate plays the role of the intensity function. The headstone of this thesis is a new description of the crash hazard rate. Recently dis- covered super-linear relations between groups sizes and group activi- ties, as well as percolation theory, are applied to the Cont-Bouchaud cluster dynamics model. i Contents Contents iii 1 Motivation1 2 Models & Methods3 2.1 The Johanson-Ledoit-Sornette (JLS) Model........... 3 2.2 Percolation theory.......................... 6 2.2.1 Introduction......................... 6 2.2.2 Percolation Point and Cluster Number......... 7 2.2.3 Finite Size Effects...................... 8 2.3 Ising Model ............................
    [Show full text]
  • 8 Blocks, Markov Chains, Ising Model
    Tel Aviv University, 2007 Large deviations 60 8 Blocks, Markov chains, Ising model 8a Introductory remarks . 60 8b Pair frequencies: combinatorial approach . 63 8c Markovchains .................... 66 8d Ising model (one-dimensional) . 68 8e Pair frequencies: linear algebra approach . 69 8f Dimension two . 70 8a Introductory remarks The following three questions are related more closely than it may seem. 8a1 Question. 100 children stay in a ring, 40 boys and 60 girls. Among the 100 pairs of neighbors, 20 pairs are heterosexual (a girl and a boy); others are not. What about the number of all such configurations? 8a2 Question. A Markov chain with two states (0 and 1) is given via its 2 2-matrix of transition probabilities. What about the probability that the × state 1 occurs 60 times among the first 100? 8a3 Question. (Ising model) A one-dimensional array of n spin-1/2 par- ticles is described by the configuration space 1, 1 n. Each configuration (s ,...,s ) 1, 1 n has its energy {− } 1 n ∈ {− } 1 H (s ,...,s )= (s s + + s s ) h(s + + s ) ; n 1 n −2 1 2 ··· n−1 n − 1 ··· n here h R is a parameter. (It is the strength of an external magnetic field, ∈ while the strength of the nearest neighbor coupling is set to 1.) What about the dependence of the energy and the mean spin (s + + s )/n on h and 1 ··· n the temperature? Tossing a fair coin n times we get a random element (β ,...,β ) of 0, 1 n, 1 n { } and may consider the n 1 pairs (β1, β2), (β2, β3),..., (βn−1, βn).
    [Show full text]
  • A Course in Interacting Particle Systems
    A Course in Interacting Particle Systems J.M. Swart January 14, 2020 arXiv:1703.10007v2 [math.PR] 13 Jan 2020 2 Contents 1 Introduction 7 1.1 General set-up . .7 1.2 The voter model . .9 1.3 The contact process . 11 1.4 Ising and Potts models . 14 1.5 Phase transitions . 17 1.6 Variations on the voter model . 20 1.7 Further models . 22 2 Continuous-time Markov chains 27 2.1 Poisson point sets . 27 2.2 Transition probabilities and generators . 30 2.3 Poisson construction of Markov processes . 31 2.4 Examples of Poisson representations . 33 3 The mean-field limit 35 3.1 Processes on the complete graph . 35 3.2 The mean-field limit of the Ising model . 36 3.3 Analysis of the mean-field model . 38 3.4 Functions of Markov processes . 42 3.5 The mean-field contact process . 47 3.6 The mean-field voter model . 49 3.7 Exercises . 51 4 Construction and ergodicity 53 4.1 Introduction . 53 4.2 Feller processes . 54 4.3 Poisson construction . 63 4.4 Generator construction . 72 4.5 Ergodicity . 79 4.6 Application to the Ising model . 81 4.7 Further results . 85 5 Monotonicity 89 5.1 The stochastic order . 89 5.2 The upper and lower invariant laws . 94 5.3 The contact process . 97 5.4 Other examples . 100 3 4 CONTENTS 5.5 Exercises . 101 6 Duality 105 6.1 Introduction . 105 6.2 Additive systems duality . 106 6.3 Cancellative systems duality . 113 6.4 Other dualities .
    [Show full text]
  • The Potts Model and Tutte Polynomial, and Associated Connections Between Statistical Mechanics and Graph Theory
    The Potts Model and Tutte Polynomial, and Associated Connections Between Statistical Mechanics and Graph Theory Robert Shrock C. N. Yang Institute for Theoretical Physics, Stony Brook University Lecture 1 at Indiana University - Purdue University, Indianapolis (IUPUI) Workshop on Connections Between Complex Dynamics, Statistical Physics, and Limiting Spectra of Self-similar Group Actions, Aug. 2016 Outline • Introduction • Potts model in statistical mechanics and equivalence of Potts partition function with Tutte polynomial in graph theory; special cases • Some easy cases • Calculational methods and some simple examples • Partition functions zeros and their accumulation sets as n → ∞ • Chromatic polynomials and ground state entropy of Potts antiferromagnet • Historical note on graph coloring • Conclusions Introduction Some Results from Statistical Physics Statistical physics deals with properties of many-body systems. Description of such systems uses a specified form for the interaction between the dynamical variables. An example is magnetic systems, where the dynamical variables are spins located at sites of a regular lattice Λ; these interact with each other by an energy function called a Hamiltonian . H Simple example: a model with integer-valued (effectively classical) spins σ = 1 at i ± sites i on a lattice Λ, with Hamiltonian = J σ σ H σ H − I i j − i Xeij Xi where JI is the spin-spin interaction constant, eij refers to a bond on Λ joining sites i and j, and H is a possible external magnetic field. This is the Ising model. Let T denote the temperature and define β = 1/(kBT ), where k = 1.38 10 23 J/K=0.862 10 4 eV/K is the Boltzmann constant.
    [Show full text]
  • Universality and Conformal Invariance for the Ising Model in Domains With
    Universality and conformal invariance for the Ising model in domains with boundary∗ by R. P. Langlands, Marc-Andre´ Lewis and Yvan Saint-Aubin Abstract. The partition function with boundary conditions for various two•dimensional Ising models is examined and previ• ously unobserved properties of conformal invariance and univer• sality are established numerically. ∗ First appeared in J. Stat. Physics 98, Nos. 1/2, pp. 131–244, 2000 1. Introduction. Although the experiments of this paper, statistical and numerical, were undertaken in pursuit of a goal not widely shared, they may be of general interest since they reveal a number of curious properties of the two•dimensional Ising model that had not been previously observed. The goal is not difficult to state. Although planar lattice models of statistical mechanics are in many respects well understood physically, their mathematical investigation lags far behind. Since these models are purely mathematical, this is regrettable. It seems to us that the problem is not simply to introduce mathematical standards into arguments otherwise well understood; rather the statistical•mechanical consequences of the notion of renormalization remain obscure. Our experiments were undertaken to support the view that the fixed point (or points) of the renormalization procedure can be realized as concrete mathematical objects and that a first step in any attempt to come to terms with renormalization is to understand what they are. We have resorted to numerical studies because a frontal mathematical attack without any clear notion of the possible conclusions has little chance of success. We are dealing with a domain in which the techniques remain to be developed.
    [Show full text]