Bayesian Generalised Ensemble Markov Chain Monte Carlo

Bayesian Generalised Ensemble Markov Chain Monte Carlo

Bayesian generalised ensemble Markov chain Monte Carlo Jes Frellsen Ole Winther University of Cambridge Technical University of Denmark Zoubin Ghahramani Jesper Ferkinghoff-Borg University of Cambridge University of Copenhagen Abstract likelihood values, which makes the method inapplica- ble for calculating key multivariate integrals, in partic- Bayesian generalised ensemble (BayesGE) ular the partition function (evidence, marginal likeli- is a new method that addresses two ma- hood) or the density of states associated with p(x) (Iba jor drawbacks of standard Markov chain 2001; Bishop 2006; Ferkinghoff-Borg 2012). Secondly, Monte Carlo algorithms for inference in high- canonical sampling is often hampered by a high degree dimensional probability models: inapplica- of correlations between the generated states for stan- bility to estimate the partition function and dard choices of the transition kernel. This property, poor mixing properties. BayesGE uses a which is referred to as poor mixing or slow relaxation Bayesian approach to iteratively update the of the Markov chain, reduces the effective number of belief about the density of states (distribu- samples and may lead to results which are erroneously tion of the log likelihood under the prior) sensitive to the arbitrary initialisation of the chain. for the model, with the dual purpose of en- In the past few decades, a variety of MCMC methods hancing the sampling efficiency and mak- known as extended ensembles have been proposed to ing the estimation of the partition function alleviate these two deficiencies (see review and refer- tractable. We benchmark BayesGE on Ising ence in Gelman and Meng (1998), Iba (2001), Mur- and Potts systems and show that it compares ray (2007), and Ferkinghoff-Borg (2012)). The un- favourably to existing state-of-the-art meth- derlying idea of this approach is to build a \bridge" ods. from the part of the probability distribution where the Markov chain suffers from slow relaxation to the part where the sampling is free from such problems. These 1 INTRODUCTION methods include simulated tempering (Marinari and Parisi 1992; Lyubartsev et al. 1992; Irb¨ack and Pot- For most probabilistic models, p(x), exact infer- thast 1995), parallel tempering (Swendsen and Wang ence is intractable, and one has to resort to some 1986; Geyer 1991) and generalised ensembles (Berg form of approximation (Bishop 2006). Markov chain and Neuhaus 1992; Lee 1993; Hesselbo and Stinch- Monte Carlo (MCMC) methods constitute a partic- combe 1995). In extended ensembles the transition ularly powerful and versatile approach for this pur- kernel is extended in such way that it at the same pose (Metropolis and Ulam 1949; Metropolis et al. time allows for the calculation of the partition function 1953; Hastings 1970). In standard MCMC methods, as well as for the reconstruction of the desired statis- p(x) is sampled through a Markov chain with a fixed tics for the original target distribution p(x). However transition kernel that is constructed to have the unique this kernel depends on integral quantities of the model invariant distribution p(x). In the following we will which are not a priori known. Therefore these tech- broadly refer to this as canonical sampling. niques rely on an iterative approach, where estimates While the canonical MCMC method has been the of these quantities obtained from previous iteration(s) workhorse in statistics and physics for the last 50 are used to define the transition probability kernel years or so, it suffers from two primary deficiencies. for the next iteration (Murray 2007; Ferkinghoff-Borg First, it only samples from a narrow interval of log- 2012). At present, however, extended ensemble techniques th Appearing in Proceedings of the 19 International Con- use rather heuristic approaches to define the required ference on Artificial Intelligence and Statistics (AISTATS) iteration procedure and are all based on frequentist 2016, Cadiz, Spain. JMLR: W&CP volume 41. Copyright 2016 by the authors. estimators. Consequently, whilst they have shown 408 Bayesian generalised ensemble Markov chain Monte Carlo promising results in a wide range of problems in ma- due to its link to the thermodynamical free energy. chine learning and statistical physics (Salakhutdinov Here, p0 = pβ=0 represents the normalised integra- 1 2010; Landau and Binder 2014), these aspects compro- tion measure and consequently Z0 ≡ Zβ=0 = 1. In mise both the robustness and speed of the algorithms the non-thermal context of Bayesian statistics, we set and limit their applicability for inference in more com- β = 1 in which case the \energy" equals minus log plex systems. likelihood E(x) = − log p(datajx), x are model pa- rameters and latent variables, p is the prior distri- In this paper we focus on the generalised ensemble 0 bution and Z ≡ Z is known as the model evidence (GE) subclass of methods, due to its larger domain of β=1 or marginal likelihood. Another central object is the application compared to the tempering based coun- density of states which is defined as terparts (Hansmann and Okamoto 1997). We pro- pose a novel Bayesian generalised ensemble (BayesGE) Z g(E) = δ(E − E(x))p (x) dx (3) method to address the problems pertaining the tra- 0 ditional GE techniques. The proposed method is and measures the prior mass associated with energy applicable to both discrete and continuous distribu- E. In statistical physics, g(E) is related to the micro- tions but requires discretisation of the log-likelihood canonical entropy s(E) through Boltzmann's formula in the current formulation. We test the algorithm s(E) ≡ log g(E), from which all thermodynamic po- on two discrete 2D spin systems: an Ising model tentials can be calculated. Typically, the energy can and a Potts model. Both models are canonical ex- be evaluated for any x but the partition function Zβ amples of systems displaying cooperative transitions and g are unknown. and slow relaxation, for which tempering based in- ference methods in general are inadequate. Infer- A Markov chain with unique invariant distribution ence in Potts models is further complicated by their pβ can be constructed using the Metropolis{Hastings generic multimodal nature as opposed to the essen- (MH) algorithm (Metropolis et al. 1953; Hastings tial bimodal nature of Ising systems. We demonstrate 1970), in which a sequence of states fxig is gener- the robustness and accuracy of the algorithm in esti- ated by sampling a trial state from a proposal distri- mating the full density of states and partition func- bution x0 ∼ q(·|xi) and accepting the state with prob- pβ (x0)q(x x0) tion of the two models for different size and/or com- j ability a(x0jxi) = min 1; p (x)q(x x) at each time plexity and show that BayesGE outperforms existing β 0j step. Typically, pβ only occupies an exponentially state-of-the-art methods: the Wang{Landau (WL) al- small part of the volume under p0, as illustrated in gorithm (Wang and Landau 2001), annealed impor- figure S1. For the Metropolis{Hastings algorithm, this tance sampling (AIS) (Neal 2001) and nested sam- implies that the normalisation constant Zβ (as well as pling (Skilling 2006; Murray et al. 2005). other multivariate-integrals) is not tractable and fur- In section 2, we outline the basic methodology of gen- thermore that the sampling for more complex models eralised ensembles. In section 3, we detail the ele- may suffer from slow mixing of the Markov chain. ments of the BayesGE method. The inference results In the GE procedure, an artificial target distribution of the algorithm in comparison to existing advanced is constructed that facilitates a \bridging" between sampling methods on the two spin systems are pre- p0 and pβ. The starting point is to replace the log- sented in section 4. We conclude by discussing relevant Boltzmann weights −βE with a different weight func- extensions of the method for future work. tion w(y(x)), where y(x) represent a set of functions of x of particular interest (\reaction coordinates" or 2 GENERALISED ENSEMBLES \order parameters"). The most common choice is to set y = E in which case the GE target distribution Consider a posterior distribution of x 2 X on the form takes the form exp(w(E(x)))p0(x) exp(−βE(x))p (x) pGE(x) = p(xjw) ≡ (4) 0 Zw pβ(x) = (1) Zβ where where Z Z Zw = exp(w(E(x)))p0(x) dx : (5) Zβ = exp(−βE(x))p0(x) dx : (2) 1 Note, that the physical partition function Z0 is nor- In the context of statistical physics, β represents the mally identified with the total volume of X , corresponding to p0 being an actual integration measure as opposed to inverse temperature times the Boltzmann constant, a normalised one. However, for most statistical considera- E(x) is the energy and the normaliser Zβ is known tions it is the ratio Zβ =Z0 that is of primary importance, as the partition function, which plays a central role so this normalisation convention poses no loss of generality. 409 Jes Frellsen, Ole Winther, Zoubin Ghahramani, Jesper Ferkinghoff-Borg Sampling from p(xjw) is realised with MH by replacing 3 BAYESIAN GENERALISED pβ(x) with p(xjw) in a(x0jxi). ENSEMBLE In order to ensure the target distribution has the de- sired properties, GE methods make use of the den- We wish to address the problems of traditional GE ap- sity of states to define the weights. Using the defi- proaches in a principled manner, by treating the en- nition from equation (3), the marginal distribution of tropy estimation as an inference problem in a Bayesian the energy p(Ejw) = R δ(E − E(x))p(xjw) dx can be framework. In the following we will assume that expressed as we are considering a discrete or discretised system having J possible energy values.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    9 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us