Modern Monte Carlo Methods and Their Application In
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
The No-U-Turn Sampler: Adaptively Setting Path Lengths in Hamiltonian Monte Carlo
Journal of Machine Learning Research 15 (2014) 1593-1623 Submitted 11/11; Revised 10/13; Published 4/14 The No-U-Turn Sampler: Adaptively Setting Path Lengths in Hamiltonian Monte Carlo Matthew D. Hoffman [email protected] Adobe Research 601 Townsend St. San Francisco, CA 94110, USA Andrew Gelman [email protected] Departments of Statistics and Political Science Columbia University New York, NY 10027, USA Editor: Anthanasios Kottas Abstract Hamiltonian Monte Carlo (HMC) is a Markov chain Monte Carlo (MCMC) algorithm that avoids the random walk behavior and sensitivity to correlated parameters that plague many MCMC methods by taking a series of steps informed by first-order gradient information. These features allow it to converge to high-dimensional target distributions much more quickly than simpler methods such as random walk Metropolis or Gibbs sampling. However, HMC's performance is highly sensitive to two user-specified parameters: a step size and a desired number of steps L. In particular, if L is too small then the algorithm exhibits undesirable random walk behavior, while if L is too large the algorithm wastes computation. We introduce the No-U-Turn Sampler (NUTS), an extension to HMC that eliminates the need to set a number of steps L. NUTS uses a recursive algorithm to build a set of likely candidate points that spans a wide swath of the target distribution, stopping automatically when it starts to double back and retrace its steps. Empirically, NUTS performs at least as efficiently as (and sometimes more efficiently than) a well tuned standard HMC method, without requiring user intervention or costly tuning runs. -
Introduction to Ggplot2
Introduction to ggplot2 Dawn Koffman Office of Population Research Princeton University January 2014 1 Part 1: Concepts and Terminology 2 R Package: ggplot2 Used to produce statistical graphics, author = Hadley Wickham "attempt to take the good things about base and lattice graphics and improve on them with a strong, underlying model " based on The Grammar of Graphics by Leland Wilkinson, 2005 "... describes the meaning of what we do when we construct statistical graphics ... More than a taxonomy ... Computational system based on the underlying mathematics of representing statistical functions of data." - does not limit developer to a set of pre-specified graphics adds some concepts to grammar which allow it to work well with R 3 qplot() ggplot2 provides two ways to produce plot objects: qplot() # quick plot – not covered in this workshop uses some concepts of The Grammar of Graphics, but doesn’t provide full capability and designed to be very similar to plot() and simple to use may make it easy to produce basic graphs but may delay understanding philosophy of ggplot2 ggplot() # grammar of graphics plot – focus of this workshop provides fuller implementation of The Grammar of Graphics may have steeper learning curve but allows much more flexibility when building graphs 4 Grammar Defines Components of Graphics data: in ggplot2, data must be stored as an R data frame coordinate system: describes 2-D space that data is projected onto - for example, Cartesian coordinates, polar coordinates, map projections, ... geoms: describe type of geometric objects that represent data - for example, points, lines, polygons, ... aesthetics: describe visual characteristics that represent data - for example, position, size, color, shape, transparency, fill scales: for each aesthetic, describe how visual characteristic is converted to display values - for example, log scales, color scales, size scales, shape scales, .. -
Regression Models by Gretl and R Statistical Packages for Data Analysis in Marine Geology Polina Lemenkova
Regression Models by Gretl and R Statistical Packages for Data Analysis in Marine Geology Polina Lemenkova To cite this version: Polina Lemenkova. Regression Models by Gretl and R Statistical Packages for Data Analysis in Marine Geology. International Journal of Environmental Trends (IJENT), 2019, 3 (1), pp.39 - 59. hal-02163671 HAL Id: hal-02163671 https://hal.archives-ouvertes.fr/hal-02163671 Submitted on 3 Jul 2019 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Distributed under a Creative Commons Attribution| 4.0 International License International Journal of Environmental Trends (IJENT) 2019: 3 (1),39-59 ISSN: 2602-4160 Research Article REGRESSION MODELS BY GRETL AND R STATISTICAL PACKAGES FOR DATA ANALYSIS IN MARINE GEOLOGY Polina Lemenkova 1* 1 ORCID ID number: 0000-0002-5759-1089. Ocean University of China, College of Marine Geo-sciences. 238 Songling Rd., 266100, Qingdao, Shandong, P. R. C. Tel.: +86-1768-554-1605. Abstract Received 3 May 2018 Gretl and R statistical libraries enables to perform data analysis using various algorithms, modules and functions. The case study of this research consists in geospatial analysis of Accepted the Mariana Trench, a hadal trench located in the Pacific Ocean. -
September 2010
THE ISBA BULLETIN Vol. 17 No. 3 September 2010 The official bulletin of the International Society for Bayesian Analysis AMESSAGE FROM THE membership renewal (there will be special provi- PRESIDENT sions for members who hold multi-year member- ships). Peter Muller¨ The Bayesian Nonparametrics Section (IS- ISBA President, 2010 BA/BNP) is already up and running, and plan- [email protected] ning the 2011 BNP workshop. Please see the News from the World section in this Bulletin First some sad news. In August we lost two and our homepage (select “business” and “mee- big Bayesians. Julian Besag passed away on Au- tings”). gust 6, and Arnold Zellner passed away on Au- gust 11. Arnold was one of the founding ISBA ISBA/SBSS Educational Initiative. Jointly presidents and was instrumental to get Bayesian with ASA/SBSS (American Statistical Associa- started. Obituaries on this issue of the Analysis tion, Section on Bayesian Statistical Science) we Bulletin and on our homepage acknowledge Juli- launched a new joint educational initiative. The an and Arnold’s pathbreaking contributions and initiative formalized a long standing history of their impact on the lives of many people in our collaboration of ISBA and ASA/SBSS in related research community. They will be missed dearly! matters. Continued on page 2. ISBA Elections 2010. Please check out the elec- tion statements of the candidates for the upco- In this issue ming ISBA elections. We have an amazing slate ‰ A MESSAGE FROM THE BA EDITOR of candidates. Thanks to the nominating commit- *Page 2 tee, Mike West (chair), Renato Martins Assuncao,˜ Jennifer Hill, Beatrix Jones, Jaeyong Lee, Yasuhi- ‰ 2010 ISBA ELECTION *Page 3 ro Omori and Gareth Robert! ‰ SAVAGE AWARD AND MITCHELL PRIZE 2010 *Page 8 Sections. -
IMS Bulletin 39(4)
Volume 39 • Issue 4 IMS1935–2010 Bulletin May 2010 Meet the 2010 candidates Contents 1 IMS Elections 2–3 Members’ News: new ISI members; Adrian Raftery; It’s time for the 2010 IMS elections, and Richard Smith; we introduce this year’s nominees who are IMS Collections vol 5 standing for IMS President-Elect and for IMS Council. You can read all the candi- 4 IMS Election candidates dates’ statements, starting on page 4. 9 Amendments to This year there are also amendments Constitution and Bylaws to the Constitution and Bylaws to vote Letter to the Editor 11 on: they are listed The candidate for IMS President-Elect is Medallion Preview: Laurens on page 9. 13 Ruth Williams de Haan Voting is open until June 26, so 14 COPSS Fisher lecture: Bruce https://secure.imstat.org/secure/vote2010/vote2010.asp Lindsay please visit to cast your vote! 15 Rick’s Ramblings: March Madness 16 Terence’s Stuff: And ANOVA thing 17 IMS meetings 27 Other meetings 30 Employment Opportunities 31 International Calendar of Statistical Events The ten Council candidates, clockwise from top left, are: 35 Information for Advertisers Krzysztof Burdzy, Francisco Cribari-Neto, Arnoldo Frigessi, Peter Kim, Steve Lalley, Neal Madras, Gennady Samorodnitsky, Ingrid Van Keilegom, Yazhen Wang and Wing H Wong Abstract submission deadline extended to April 30 IMS Bulletin 2 . IMs Bulletin Volume 39 . Issue 4 Volume 39 • Issue 4 May 2010 IMS members’ news ISSN 1544-1881 International Statistical Institute elects new members Contact information Among the 54 new elected ISI members are several IMS members. We congratulate IMS IMS Bulletin Editor: Xuming He Fellow Jon Wellner, and IMS members: Subhabrata Chakraborti, USA; Liliana Forzani, Assistant Editor: Tati Howell Argentina; Ronald D. -
Arxiv:1801.02309V4 [Stat.ML] 11 Dec 2019
Journal of Machine Learning Research 20 (2019) 1-42 Submitted 4/19; Revised 11/19; Published 12/19 Log-concave sampling: Metropolis-Hastings algorithms are fast Raaz Dwivedi∗;y [email protected] Yuansi Chen∗;} [email protected] Martin J. Wainwright};y;z [email protected] Bin Yu};y [email protected] Department of Statistics} Department of Electrical Engineering and Computer Sciencesy University of California, Berkeley Voleon Groupz, Berkeley Editor: Suvrit Sra Abstract We study the problem of sampling from a strongly log-concave density supported on Rd, and prove a non-asymptotic upper bound on the mixing time of the Metropolis-adjusted Langevin algorithm (MALA). The method draws samples by simulating a Markov chain obtained from the discretization of an appropriate Langevin diffusion, combined with an accept-reject step. Relative to known guarantees for the unadjusted Langevin algorithm (ULA), our bounds show that the use of an accept-reject step in MALA leads to an ex- ponentially improved dependence on the error-tolerance. Concretely, in order to obtain samples with TV error at most δ for a density with condition number κ, we show that MALA requires κd log(1/δ) steps from a warm start, as compared to the κ2d/δ2 steps establishedO in past work on ULA. We also demonstrate the gains of a modifiedO version of MALA over ULA for weakly log-concave densities. Furthermore, we derive mixing time bounds for the Metropolized random walk (MRW) and obtain (κ) mixing time slower than MALA. We provide numerical examples that support ourO theoretical findings, and demonstrate the benefits of Metropolis-Hastings adjustment for Langevin-type sampling algorithms. -
Julia: a Fresh Approach to Numerical Computing∗
SIAM REVIEW c 2017 Society for Industrial and Applied Mathematics Vol. 59, No. 1, pp. 65–98 Julia: A Fresh Approach to Numerical Computing∗ Jeff Bezansony Alan Edelmanz Stefan Karpinskix Viral B. Shahy Abstract. Bridging cultures that have often been distant, Julia combines expertise from the diverse fields of computer science and computational science to create a new approach to numerical computing. Julia is designed to be easy and fast and questions notions generally held to be \laws of nature" by practitioners of numerical computing: 1. High-level dynamic programs have to be slow. 2. One must prototype in one language and then rewrite in another language for speed or deployment. 3. There are parts of a system appropriate for the programmer, and other parts that are best left untouched as they have been built by the experts. We introduce the Julia programming language and its design|a dance between special- ization and abstraction. Specialization allows for custom treatment. Multiple dispatch, a technique from computer science, picks the right algorithm for the right circumstance. Abstraction, which is what good computation is really about, recognizes what remains the same after differences are stripped away. Abstractions in mathematics are captured as code through another technique from computer science, generic programming. Julia shows that one can achieve machine performance without sacrificing human con- venience. Key words. Julia, numerical, scientific computing, parallel AMS subject classifications. 68N15, 65Y05, 97P40 DOI. 10.1137/141000671 Contents 1 Scientific Computing Languages: The Julia Innovation 66 1.1 Julia Architecture and Language Design Philosophy . 67 ∗Received by the editors December 18, 2014; accepted for publication (in revised form) December 16, 2015; published electronically February 7, 2017. -
Tea: a High-Level Language and Runtime System for Automating
Tea: A High-level Language and Runtime System for Automating Statistical Analysis Eunice Jun1, Maureen Daum1, Jared Roesch1, Sarah Chasins2, Emery Berger34, Rene Just1, Katharina Reinecke1 1University of Washington, 2University of California, 3University of Seattle, WA Berkeley, CA Massachusetts Amherst, femjun, mdaum, jroesch, [email protected] 4Microsoft Research, rjust, [email protected] Redmond, WA [email protected] ABSTRACT Since the development of modern statistical methods (e.g., Though statistical analyses are centered on research questions Student’s t-test, ANOVA, etc.), statisticians have acknowl- and hypotheses, current statistical analysis tools are not. Users edged the difficulty of identifying which statistical tests people must first translate their hypotheses into specific statistical should use to answer their specific research questions. Almost tests and then perform API calls with functions and parame- a century later, choosing appropriate statistical tests for eval- ters. To do so accurately requires that users have statistical uating a hypothesis remains a challenge. As a consequence, expertise. To lower this barrier to valid, replicable statistical errors in statistical analyses are common [26], especially given analysis, we introduce Tea, a high-level declarative language that data analysis has become a common task for people with and runtime system. In Tea, users express their study de- little to no statistical expertise. sign, any parametric assumptions, and their hypotheses. Tea A wide variety of tools (such as SPSS [55], SAS [54], and compiles these high-level specifications into a constraint satis- JMP [52]), programming languages (e.g., R [53]), and libraries faction problem that determines the set of valid statistical tests (including numpy [40], scipy [23], and statsmodels [45]), en- and then executes them to test the hypothesis. -
Revolution R Enterprise™ 7.1 Getting Started Guide
Revolution R Enterprise™ 7.1 Getting Started Guide The correct bibliographic citation for this manual is as follows: Revolution Analytics, Inc. 2014. Revolution R Enterprise 7.1 Getting Started Guide. Revolution Analytics, Inc., Mountain View, CA. Revolution R Enterprise 7.1 Getting Started Guide Copyright © 2014 Revolution Analytics, Inc. All rights reserved. No part of this publication may be reproduced, stored in a retrieval system, or transmitted, in any form or by any means, electronic, mechanical, photocopying, recording, or otherwise, without the prior written permission of Revolution Analytics. U.S. Government Restricted Rights Notice: Use, duplication, or disclosure of this software and related documentation by the Government is subject to restrictions as set forth in subdivision (c) (1) (ii) of The Rights in Technical Data and Computer Software clause at 52.227-7013. Revolution R, Revolution R Enterprise, RPE, RevoScaleR, RevoDeployR, RevoTreeView, and Revolution Analytics are trademarks of Revolution Analytics. Other product names mentioned herein are used for identification purposes only and may be trademarks of their respective owners. Revolution Analytics. 2570 W. El Camino Real Suite 222 Mountain View, CA 94040 USA. Revised on March 3, 2014 We want our documentation to be useful, and we want it to address your needs. If you have comments on this or any Revolution document, send e-mail to [email protected]. We’d love to hear from you. Contents Chapter 1. What Is Revolution R Enterprise? .................................................................... -
Julia: a Modern Language for Modern ML
Julia: A modern language for modern ML Dr. Viral Shah and Dr. Simon Byrne www.juliacomputing.com What we do: Modernize Technical Computing Today’s technical computing landscape: • Develop new learning algorithms • Run them in parallel on large datasets • Leverage accelerators like GPUs, Xeon Phis • Embed into intelligent products “Business as usual” will simply not do! General Micro-benchmarks: Julia performs almost as fast as C • 10X faster than Python • 100X faster than R & MATLAB Performance benchmark relative to C. A value of 1 means as fast as C. Lower values are better. A real application: Gillespie simulations in systems biology 745x faster than R • Gillespie simulations are used in the field of drug discovery. • Also used for simulations of epidemiological models to study disease propagation • Julia package (Gillespie.jl) is the state of the art in Gillespie simulations • https://github.com/openjournals/joss- papers/blob/master/joss.00042/10.21105.joss.00042.pdf Implementation Time per simulation (ms) R (GillespieSSA) 894.25 R (handcoded) 1087.94 Rcpp (handcoded) 1.31 Julia (Gillespie.jl) 3.99 Julia (Gillespie.jl, passing object) 1.78 Julia (handcoded) 1.2 Those who convert ideas to products fastest will win Computer Quants develop Scientists prepare algorithms The last 25 years for production (Python, R, SAS, DEPLOY (C++, C#, Java) Matlab) Quants and Computer Compress the Scientists DEPLOY innovation cycle collaborate on one platform - JULIA with Julia Julia offers competitive advantages to its users Julia is poised to become one of the Thank you for Julia. Yo u ' v e k i n d l ed leading tools deployed by developers serious excitement. -
An Introduction to Data Analysis Using the Pymc3 Probabilistic Programming Framework: a Case Study with Gaussian Mixture Modeling
An introduction to data analysis using the PyMC3 probabilistic programming framework: A case study with Gaussian Mixture Modeling Shi Xian Liewa, Mohsen Afrasiabia, and Joseph L. Austerweila aDepartment of Psychology, University of Wisconsin-Madison, Madison, WI, USA August 28, 2019 Author Note Correspondence concerning this article should be addressed to: Shi Xian Liew, 1202 West Johnson Street, Madison, WI 53706. E-mail: [email protected]. This work was funded by a Vilas Life Cycle Professorship and the VCRGE at University of Wisconsin-Madison with funding from the WARF 1 RUNNING HEAD: Introduction to PyMC3 with Gaussian Mixture Models 2 Abstract Recent developments in modern probabilistic programming have offered users many practical tools of Bayesian data analysis. However, the adoption of such techniques by the general psychology community is still fairly limited. This tutorial aims to provide non-technicians with an accessible guide to PyMC3, a robust probabilistic programming language that allows for straightforward Bayesian data analysis. We focus on a series of increasingly complex Gaussian mixture models – building up from fitting basic univariate models to more complex multivariate models fit to real-world data. We also explore how PyMC3 can be configured to obtain significant increases in computational speed by taking advantage of a machine’s GPU, in addition to the conditions under which such acceleration can be expected. All example analyses are detailed with step-by-step instructions and corresponding Python code. Keywords: probabilistic programming; Bayesian data analysis; Markov chain Monte Carlo; computational modeling; Gaussian mixture modeling RUNNING HEAD: Introduction to PyMC3 with Gaussian Mixture Models 3 1 Introduction Over the last decade, there has been a shift in the norms of what counts as rigorous research in psychological science. -
The Semantics of Subroutines and Iteration in the Bayesian Programming Language Probt
1 The semantics of Subroutines and Iteration in the Bayesian Programming language ProBT R. LAURENT∗, K. MEKHNACHA∗, E. MAZERy and P. BESSIERE` z ∗ProbaYes S.A.S, Grenoble, France yUniversity of Grenoble-Alpes, CNRS/LIG, Grenoble, France zUniversity Pierre et Marie Curie, CNRS/ISIR, Paris, France Abstract—Bayesian models are tools of choice when solving 8 8 8 problems with incomplete information. Bayesian networks pro- > > >Variables > > <> vide a first but limited approach to address such problems. > <> Decomposition <> Specification (π) For real world applications, additional semantics is needed to Description > P arametric construct more complex models, especially those with repetitive > :>F orms > > P rogram structures or substructures. ProBT, a Bayesian a programming > :> > Identification (based on δ) language, provides a set of constructs for developing and applying :> complex models with substructures and repetitive structures. Question The goal of this paper is to present and discuss the semantics associated to these constructs. Figure 1. A Bayesian program is constructed from a Description and a Question. The Description is given by the programmer who writes a Index Terms—Probabilistic Programming semantics , Bayesian Specification of a model π and an Identification of its parameter values, Programming, ProBT which can be set in the program or obtained through a learning process from a data set δ. The Specification is constructed from a set of relevant variables, a decomposition of the joint probability distribution over these variables and a set of parametric forms (mathematical models) for the terms I. INTRODUCTION of this decomposition. ProBT [1] was designed to translate the ideas of E.T. Jaynes [2] into an actual programming language.