Introduction to Monte Carlo Simulation”

Total Page:16

File Type:pdf, Size:1020Kb

Introduction to Monte Carlo Simulation” Natalia A. Humphreys. Monte Carlo Simulation_Intro.docx Date: Friday, April 5, 2013 Time: 10:30 am Place: Career Center Seminar Room, SSB 3.107. Title: “Introduction to Monte Carlo Simulation” Source: This presentation is based on the article “Microsoft Excel Data Analysis and Business Modeling” by Wayne L. Winston, a well-known statistician and business professor who specializes in creative, practical applications of Excel. General description: We would like to be able to accurately estimate the probabilities of uncertain events. For example, what is the probability that a new product’s cash flows will have a positive net present value (NPV)? What is the riskiness of our investment portfolio? What is the optimal plant capacity that should be built for each drug by a pharmaceutical company? Monte Carlo simulation enables us to model situations that present uncertainty and play them out thousands of times on a computer. A little bit of history: During the 1930s and 1940s, many computer simulations were performed to estimate the probability that the chain reaction needed for the atom bomb would work successfully. The Monte Carlo method was coined then by the physicists John von Neumann, Stanislaw Ulam and Nicholas Metropolis, while they were working on this and other nuclear weapon projects (Manhattan Project) in the Los Alamos National Laboratory. It was named in homage to the Monte Carlo Casino, a famous casino where Ulam's uncle would often gamble away his money. Objectives: 1. Learn about who uses Monte Carlo simulation. 2. Refresh Excel basic skills: learn what happens when we enter =RAND() in a cell. 3. Understand how to simulate values of a discrete random variable. 4. Understand how to simulate values of a normal random variable. 5. Learn about how Monte Carlo simulation could be used to determine how many cards a greeting card company should produce. 6. Understand advantages of the Monte Carlo simulation over deterministic, or “single-point estimate” analysis. Page 1 of 1 .
Recommended publications
  • Monte Carlo Simulation in Statistics with a View to Recent Ideas
    (sfi)2 Statistics for Innovation Monte Carlo simulation in statistics, with a view to recent ideas Arnoldo Frigessi [email protected] THE JOURNAL OF CHEMICAL PHYSICS VOLUME 21, NUMBER 6 JUNE, 1953 Equation of State Calculations by Fast Computing Machines NICHOLAS METROPOLIS, ARIANNA W. ROSENBLUTH, MARSHALL N. ROSENBLUTH, AND AUGUSTA H. TELLER, Los Alamos Scientific Laboratory, Los Alamos, New Mexico AND EDWARD TELLER, * Department of Physics, University of Chicago, Chicago, Illinois (Received March 6, 1953) A general method, suitable for fast computing machines, for investigating such properties as equations of state for substances consisting of interacting individual molecules is described. The method consists of a modified Monte Carlo integration over configuration space. Results for the two-dimensional rigid-sphere system have been obtained on the Los Alamos MANIAC and are presented here. These results are compared to the free volume equation of state and to a four-term virial coefficient expansion. The Metropolis algorithm from1953 has been cited as among the top 10 algorithms having the "greatest influence on the development and practice of science and engineering in the 20th century." MANIAC The MANIAC II 1952 Multiplication took a milli- second. (Metropolis choose the name MANIAC in the hope of stopping the rash of ”silly” acronyms for machine names.) Some MCMC milestones 1953 Metropolis Heat bath, dynamic Monte Carlo 1970 Hastings Political Analysis, 10:3 (2002) Society for Political Methodology 1984 Geman & Geman Gibbs
    [Show full text]
  • The Development of Military Nuclear Strategy And
    The Development of Military Nuclear Strategy and Anglo-American Relations, 1939 – 1958 Submitted by: Geoffrey Charles Mallett Skinner to the University of Exeter as a thesis for the degree of Doctor of Philosophy in History, July 2018 This thesis is available for Library use on the understanding that it is copyright material and that no quotation from the thesis may be published without proper acknowledgement. I certify that all material in this thesis which is not my own work has been identified and that no material has previously been submitted and approved for the award of a degree by this or any other University. (Signature) ……………………………………………………………………………… 1 Abstract There was no special governmental partnership between Britain and America during the Second World War in atomic affairs. A recalibration is required that updates and amends the existing historiography in this respect. The wartime atomic relations of those countries were cooperative at the level of science and resources, but rarely that of the state. As soon as it became apparent that fission weaponry would be the main basis of future military power, America decided to gain exclusive control over the weapon. Britain could not replicate American resources and no assistance was offered to it by its conventional ally. America then created its own, closed, nuclear system and well before the 1946 Atomic Energy Act, the event which is typically seen by historians as the explanation of the fracturing of wartime atomic relations. Immediately after 1945 there was insufficient systemic force to create change in the consistent American policy of atomic monopoly. As fusion bombs introduced a new magnitude of risk, and as the nuclear world expanded and deepened, the systemic pressures grew.
    [Show full text]
  • A Selected Bibliography of Publications By, and About, J
    A Selected Bibliography of Publications by, and about, J. Robert Oppenheimer Nelson H. F. Beebe University of Utah Department of Mathematics, 110 LCB 155 S 1400 E RM 233 Salt Lake City, UT 84112-0090 USA Tel: +1 801 581 5254 FAX: +1 801 581 4148 E-mail: [email protected], [email protected], [email protected] (Internet) WWW URL: http://www.math.utah.edu/~beebe/ 17 March 2021 Version 1.47 Title word cross-reference $1 [Duf46]. $12.95 [Edg91]. $13.50 [Tho03]. $14.00 [Hug07]. $15.95 [Hen81]. $16.00 [RS06]. $16.95 [RS06]. $17.50 [Hen81]. $2.50 [Opp28g]. $20.00 [Hen81, Jor80]. $24.95 [Fra01]. $25.00 [Ger06]. $26.95 [Wol05]. $27.95 [Ger06]. $29.95 [Goo09]. $30.00 [Kev03, Kle07]. $32.50 [Edg91]. $35 [Wol05]. $35.00 [Bed06]. $37.50 [Hug09, Pol07, Dys13]. $39.50 [Edg91]. $39.95 [Bad95]. $8.95 [Edg91]. α [Opp27a, Rut27]. γ [LO34]. -particles [Opp27a]. -rays [Rut27]. -Teilchen [Opp27a]. 0-226-79845-3 [Guy07, Hug09]. 0-8014-8661-0 [Tho03]. 0-8047-1713-3 [Edg91]. 0-8047-1714-1 [Edg91]. 0-8047-1721-4 [Edg91]. 0-8047-1722-2 [Edg91]. 0-9672617-3-2 [Bro06, Hug07]. 1 [Opp57f]. 109 [Con05, Mur05, Nas07, Sap05a, Wol05, Kru07]. 112 [FW07]. 1 2 14.99/$25.00 [Ber04a]. 16 [GHK+96]. 1890-1960 [McG02]. 1911 [Meh75]. 1945 [GHK+96, Gow81, Haw61, Bad95, Gol95a, Hew66, She82, HBP94]. 1945-47 [Hew66]. 1950 [Ano50]. 1954 [Ano01b, GM54, SZC54]. 1960s [Sch08a]. 1963 [Kuh63]. 1967 [Bet67a, Bet97, Pun67, RB67]. 1976 [Sag79a, Sag79b]. 1981 [Ano81]. 20 [Goe88]. 2005 [Dre07]. 20th [Opp65a, Anoxx, Kai02].
    [Show full text]
  • The Convergence of Markov Chain Monte Carlo Methods 3
    The Convergence of Markov chain Monte Carlo Methods: From the Metropolis method to Hamiltonian Monte Carlo Michael Betancourt From its inception in the 1950s to the modern frontiers of applied statistics, Markov chain Monte Carlo has been one of the most ubiquitous and successful methods in statistical computing. The development of the method in that time has been fueled by not only increasingly difficult problems but also novel techniques adopted from physics. In this article I will review the history of Markov chain Monte Carlo from its inception with the Metropolis method to the contemporary state-of-the-art in Hamiltonian Monte Carlo. Along the way I will focus on the evolving interplay between the statistical and physical perspectives of the method. This particular conceptual emphasis, not to mention the brevity of the article, requires a necessarily incomplete treatment. A complementary, and entertaining, discussion of the method from the statistical perspective is given in Robert and Casella (2011). Similarly, a more thorough but still very readable review of the mathematics behind Markov chain Monte Carlo and its implementations is given in the excellent survey by Neal (1993). I will begin with a discussion of the mathematical relationship between physical and statistical computation before reviewing the historical introduction of Markov chain Monte Carlo and its first implementations. Then I will continue to the subsequent evolution of the method with increasing more sophisticated implementations, ultimately leading to the advent of Hamiltonian Monte Carlo. 1. FROM PHYSICS TO STATISTICS AND BACK AGAIN At the dawn of the twentieth-century, physics became increasingly focused on under- arXiv:1706.01520v2 [stat.ME] 10 Jan 2018 standing the equilibrium behavior of thermodynamic systems, especially ensembles of par- ticles.
    [Show full text]
  • Monte Carlo Sampling Methods
    [1] Monte Carlo Sampling Methods Jasmina L. Vujic Nuclear Engineering Department University of California, Berkeley Email: [email protected] phone: (510) 643-8085 fax: (510) 643-9685 UCBNE, J. Vujic [2] Monte Carlo Monte Carlo is a computational technique based on constructing a random process for a problem and carrying out a NUMERICAL EXPERIMENT by N-fold sampling from a random sequence of numbers with a PRESCRIBED probability distribution. x - random variable N 1 xˆ = ---- x N∑ i i = 1 Xˆ - the estimated or sample mean of x x - the expectation or true mean value of x If a problem can be given a PROBABILISTIC interpretation, then it can be modeled using RANDOM NUMBERS. UCBNE, J. Vujic [3] Monte Carlo • Monte Carlo techniques came from the complicated diffusion problems that were encountered in the early work on atomic energy. • 1772 Compte de Bufon - earliest documented use of random sampling to solve a mathematical problem. • 1786 Laplace suggested that π could be evaluated by random sampling. • Lord Kelvin used random sampling to aid in evaluating time integrals associated with the kinetic theory of gases. • Enrico Fermi was among the first to apply random sampling methods to study neutron moderation in Rome. • 1947 Fermi, John von Neuman, Stan Frankel, Nicholas Metropolis, Stan Ulam and others developed computer-oriented Monte Carlo methods at Los Alamos to trace neutrons through fissionable materials UCBNE, J. Vujic Monte Carlo [4] Monte Carlo methods can be used to solve: a) The problems that are stochastic (probabilistic) by nature: - particle transport, - telephone and other communication systems, - population studies based on the statistics of survival and reproduction.
    [Show full text]
  • The Markov Chain Monte Carlo Approach to Importance Sampling
    The Markov Chain Monte Carlo Approach to Importance Sampling in Stochastic Programming by Berk Ustun B.S., Operations Research, University of California, Berkeley (2009) B.A., Economics, University of California, Berkeley (2009) Submitted to the School of Engineering in partial fulfillment of the requirements for the degree of Master of Science in Computation for Design and Optimization at the MASSACHUSETTS INSTITUTE OF TECHNOLOGY September 2012 c Massachusetts Institute of Technology 2012. All rights reserved. Author.................................................................... School of Engineering August 10, 2012 Certified by . Mort Webster Assistant Professor of Engineering Systems Thesis Supervisor Certified by . Youssef Marzouk Class of 1942 Associate Professor of Aeronautics and Astronautics Thesis Reader Accepted by . Nicolas Hadjiconstantinou Associate Professor of Mechanical Engineering Director, Computation for Design and Optimization 2 The Markov Chain Monte Carlo Approach to Importance Sampling in Stochastic Programming by Berk Ustun Submitted to the School of Engineering on August 10, 2012, in partial fulfillment of the requirements for the degree of Master of Science in Computation for Design and Optimization Abstract Stochastic programming models are large-scale optimization problems that are used to fa- cilitate decision-making under uncertainty. Optimization algorithms for such problems need to evaluate the expected future costs of current decisions, often referred to as the recourse function. In practice, this calculation is computationally difficult as it involves the evaluation of a multidimensional integral whose integrand is an optimization problem. Accordingly, the recourse function is estimated using quadrature rules or Monte Carlo methods. Although Monte Carlo methods present numerous computational benefits over quadrature rules, they require a large number of samples to produce accurate results when they are embedded in an optimization algorithm.
    [Show full text]
  • The Los Alamos Thermonuclear Weapon Project, 1942-1952
    Igniting The Light Elements: The Los Alamos Thermonuclear Weapon Project, 1942-1952 by Anne Fitzpatrick Dissertation submitted to the Faculty of Virginia Polytechnic Institute and State University in partial fulfillment of the requirements for the degree of DOCTOR OF PHILOSOPHY in SCIENCE AND TECHNOLOGY STUDIES Approved: Joseph C. Pitt, Chair Richard M. Burian Burton I. Kaufman Albert E. Moyer Richard Hirsh June 23, 1998 Blacksburg, Virginia Keywords: Nuclear Weapons, Computing, Physics, Los Alamos National Laboratory Igniting the Light Elements: The Los Alamos Thermonuclear Weapon Project, 1942-1952 by Anne Fitzpatrick Committee Chairman: Joseph C. Pitt Science and Technology Studies (ABSTRACT) The American system of nuclear weapons research and development was conceived and developed not as a result of technological determinism, but by a number of individual architects who promoted the growth of this large technologically-based complex. While some of the technological artifacts of this system, such as the fission weapons used in World War II, have been the subject of many historical studies, their technical successors -- fusion (or hydrogen) devices -- are representative of the largely unstudied highly secret realms of nuclear weapons science and engineering. In the postwar period a small number of Los Alamos Scientific Laboratory’s staff and affiliates were responsible for theoretical work on fusion weapons, yet the program was subject to both the provisions and constraints of the U. S. Atomic Energy Commission, of which Los Alamos was a part. The Commission leadership’s struggle to establish a mission for its network of laboratories, least of all to keep them operating, affected Los Alamos’s leaders’ decisions as to the course of weapons design and development projects.
    [Show full text]
  • Oral History Interview with Nicolas C. Metropolis
    An Interview with NICOLAS C. METROPOLIS OH 135 Conducted by William Aspray on 29 May 1987 Los Alamos, NM Charles Babbage Institute The Center for the History of Information Processing University of Minnesota, Minneapolis Copyright, Charles Babbage Institute 1 Nicholas C. Metropolis Interview 29 May 1987 Abstract Metropolis, the first director of computing services at Los Alamos National Laboratory, discusses John von Neumann's work in computing. Most of the interview concerns activity at Los Alamos: how von Neumann came to consult at the laboratory; his scientific contacts there, including Metropolis, Robert Richtmyer, and Edward Teller; von Neumann's first hands-on experience with punched card equipment; his contributions to shock-fitting and the implosion problem; interactions between, and comparisons of von Neumann and Enrico Fermi; and the development of Monte Carlo techniques. Other topics include: the relationship between Turing and von Neumann; work on numerical methods for non-linear problems; and the ENIAC calculations done for Los Alamos. 2 NICHOLAS C. METROPOLIS INTERVIEW DATE: 29 May 1987 INTERVIEWER: William Aspray LOCATION: Los Alamos, NM ASPRAY: This is an interview on the 29th of May, 1987 in Los Alamos, New Mexico with Dr. Nicholas Metropolis. Why don't we begin by having you tell me something about when you first met von Neumann and what your relationship with him was over a period of time? METROPOLIS: Well, he first came to Los Alamos on the invitation of Dr. Oppenheimer. I don't remember quite when the date was that he came, but it was rather early on. It may have been the fall of 1943.
    [Show full text]
  • Doing Mathematics on the ENIAC. Von Neumann's and Lehmer's
    Doing Mathematics on the ENIAC. Von Neumann's and Lehmer's different visions.∗ Liesbeth De Mol Center for Logic and Philosophy of Science University of Ghent, Blandijnberg 2, 9000 Gent, Belgium [email protected] Abstract In this paper we will study the impact of the computer on math- ematics and its practice from a historical point of view. We will look at what kind of mathematical problems were implemented on early electronic computing machines and how these implementations were perceived. By doing so, we want to stress that the computer was in fact, from its very beginning, conceived as a mathematical instru- ment per se, thus situating the contemporary usage of the computer in mathematics in its proper historical background. We will focus on the work by two computer pioneers: Derrick H. Lehmer and John von Neumann. They were both involved with the ENIAC and had strong opinions about how these new machines might influence (theoretical and applied) mathematics. 1 Introduction The impact of the computer on society can hardly be underestimated: it affects almost every aspect of our lives. This influence is not restricted to ev- eryday activities like booking a hotel or corresponding with friends. Science has been changed dramatically by the computer { both in its form and in ∗The author is currently a postdoctoral research fellow of the Fund for Scientific Re- search { Flanders (FWO) and a fellow of the Kunsthochschule f¨urMedien, K¨oln. 1 its content. Also mathematics did not escape this influence of the computer. In fact, the first computer applications were mathematical in nature, i.e., the first electronic general-purpose computing machines were used to solve or study certain mathematical (applied as well as theoretical) problems.
    [Show full text]
  • The New Nuclear Forensics: Analysis of Nuclear Material for Security
    THE NEW NUCLEAR FORENSICS Analysis of Nuclear Materials for Security Purposes edited by vitaly fedchenko The New Nuclear Forensics Analysis of Nuclear Materials for Security Purposes STOCKHOLM INTERNATIONAL PEACE RESEARCH INSTITUTE SIPRI is an independent international institute dedicated to research into conflict, armaments, arms control and disarmament. Established in 1966, SIPRI provides data, analysis and recommendations, based on open sources, to policymakers, researchers, media and the interested public. The Governing Board is not responsible for the views expressed in the publications of the Institute. GOVERNING BOARD Sven-Olof Petersson, Chairman (Sweden) Dr Dewi Fortuna Anwar (Indonesia) Dr Vladimir Baranovsky (Russia) Ambassador Lakhdar Brahimi (Algeria) Jayantha Dhanapala (Sri Lanka) Ambassador Wolfgang Ischinger (Germany) Professor Mary Kaldor (United Kingdom) The Director DIRECTOR Dr Ian Anthony (United Kingdom) Signalistgatan 9 SE-169 70 Solna, Sweden Telephone: +46 8 655 97 00 Fax: +46 8 655 97 33 Email: [email protected] Internet: www.sipri.org The New Nuclear Forensics Analysis of Nuclear Materials for Security Purposes EDITED BY VITALY FEDCHENKO OXFORD UNIVERSITY PRESS 2015 1 Great Clarendon Street, Oxford OX2 6DP, United Kingdom Oxford University Press is a department of the University of Oxford. It furthers the University’s objective of excellence in research, scholarship, and education by publishing worldwide. Oxford is a registered trade mark of Oxford University Press in the UK and in certain other countries © SIPRI 2015 The moral rights of the authors have been asserted All rights reserved. No part of this publication may be reproduced, stored in a retrieval system, or transmitted, in any form or by any means, without the prior permission in writing of SIPRI, or as expressly permitted by law, or under terms agreed with the appropriate reprographics rights organizations.
    [Show full text]
  • 75 YEARS Trinity Test the Dawn of America’S Scientific Innovation CONTENTS
    75 YEARS Trinity Test The Dawn of America’s Scientific Innovation CONTENTS 1 THE MANHATTAN PROJECT ........ 4 2 TRINITY TEST - JULY 16, 1945 ...... 6 3 1940s .............................................. 10 4 1950s ..........................................12 5 1960s ..........................................14 1970s ..........................................16 Beyond the advances in nuclear physics 6 and chemistry that made the “ 7 1980s ..........................................18 “ first functional atomic device possible, Trinity was arguably the greatest 1990s ..........................................20 scientific experiment ever conducted. 8 9 2000s ..........................................22 Lisa E. Gordon-Hagerty U.S. Under Secretary of Energy for Nuclear Security Administrator of the National Nuclear Security Administration 10 2010s ..........................................24 11 2020 and FORWARD ..................... 26 1 Manhattan Project: The Origin of the Trinity Test In the 1920s-1930s, a young Hungarian-German physicist Roosevelt responded by launching The Manhattan Project, Leo Szilard led the field of nuclear research, submitting a nationwide network of laboratories and manufacturing patents for a linear accelerator (1928) and cyclotron (1929), facilities designed to collaboratively assist in the before collaborating with Albert Einstein to develop the manufacture of a new atomic weapon. Einstein refrigerator. But when Hitler came into power in 1933, Szilard fled to England, encouraging his friends and family to do the same. In England, he first described the nuclear chain reaction (1933) and patented an early design for a nuclear fission reactor (1934). In 1938, Szilard joined Einstein in the United States, but the rumor that a group of Berlin chemists had split the uranium atom made them so concerned that in 1939, they sent an urgent letter to President Franklin D. Roosevelt, warning him that that Axis scientists were working to turn new nuclear discoveries into a superweapon.
    [Show full text]
  • Teller's Technical Nemeses: the American Hydrogen Bomb and Its Development Within a Technological Infrastructure
    PHIL & TECH 3:3 Spring 1998 Fitzpatrick, Teller's Technical Nemeses/10 TELLER'S TECHNICAL NEMESES: THE AMERICAN HYDROGEN BOMB AND ITS DEVELOPMENT WITHIN A TECHNOLOGICAL INFRASTRUCTURE Anne Fitzpatrick, George Washington University In World War II the U.S. Army contracted the University of Pennsylvania's Moore School of Engineering to develop a new, large electronic computer—among the first of its kind—in hopes that the machine would be able to perform ballistics calculations for the war effort. The machine was not completed before the end of the war, however, and the Army was not even the first group to utilize the machine. The first calculation ever run on the Electronic Numeric Integrator and Calculator (or ENIAC, as it was known), was for the Los Alamos nuclear weapons laboratory. The "Super problem" was the first attempt to calculate the feasibility of a thermonuclear bomb. The problem, however, was too complicated for the ENIAC with its 1000 bits of memory and 18,000 vacuum tubes, and only a very simplified version of the calculation was run, revealing very little about how such a weapon might work. Although Los Alamos was exploring hydrogen weapons during and right after the Second World War, why did the U.S. not successfully test a thermonuclear bomb until 1952? I will argue that the American thermonuclear weapons program was, early on, entrenched in a technological infrastructure which affected the pace and initial results of the project, demonstrating how one particular aspect of this infrastructure—computing—influenced the practice of nuclear weapons research, design, and development.
    [Show full text]