Selected Topics on Data Analysis in Particle Physics Lecture Notes

Total Page:16

File Type:pdf, Size:1020Kb

Selected Topics on Data Analysis in Particle Physics Lecture Notes Selected Topics on Data Analysis in Particle Physics Lecture Notes V.Karim¨aki HEP Summer School, Lammi Biological Station March 1992 Printed version edited from hand made transparencies, Spring, 2010 Contents 1 Introduction 1 2 Propagation of measurement errors 2 2.1 Error propagation in case of linear transformation ............ 2 2.2 Error propagation in case of non-linear transformation ......... 4 3 Fitting parameterized model with experimental data 7 3.1 The method of maximum likelihood .................... 7 3.2 Least squares method as a special case of maximum likelihood ..... 8 4 Least squares solutions with error estimates 10 4.1 Linear χ2 solution .............................. 10 4.2 Non-linear least squares fit ......................... 11 4.2.1 Convergence criteria ........................ 13 4.3 Least squares fit with constraints ..................... 13 4.4 Fit quality tests ............................... 16 4.4.1 Pull or stretch values of fitted parameters ............ 16 4.4.2 Fit probability distribution ..................... 16 4.5 Maximum Likelihood and Poisson statistics ................ 18 4.5.1 Maximum Likelihood and parameter errors ............ 20 5 Introduction to Monte Carlo simulation 21 5.1 Generation of non-uniform distribution by inversion method ...... 21 5.2 Inversion method for discrete distribution ................. 22 5.3 Approximate inversion method using tabulation ............. 23 5.4 Hit-or-Miss method ............................. 24 5.5 Hit-or-Miss method by comparison function ............... 24 5.6 Composition method ............................ 25 i 6 Generation from common continuous distributions 27 6.1 Uniform distribution ............................ 27 6.2 Gaussian distribution ........................... 27 6.2.1 Polar or Box-Muller method .................... 29 6.2.2 Faster variation of the polar method ............... 30 6.2.3 Gaussian generation by summation method ........... 31 6.2.4 Kahn’s method ........................... 32 6.3 Multidimensional Gaussian distribution .................. 32 6.4 Exponential distribution .......................... 35 6.5 χ2 distribution ............................... 37 6.6 Cauchy or Breit-Wigner distribution ................... 39 6.7 Landau distribution ............................. 41 7 Monte Carlo integration 43 7.1 Basics of the MC integration ........................ 43 7.2 Convergence and precision of the MC integration ............ 45 7.3 Methods to accelerate the convergence and precision ........... 46 7.3.1 Stratified sampling ......................... 47 7.3.2 Importance sampling ........................ 48 7.3.3 Control variates ........................... 49 7.3.4 Antithetic variates ......................... 50 7.4 Limits of integration in the MC method ................. 51 7.5 Comparison of convergence with other methods ............. 52 A Basic statistics for simulation A53 A.1 Introduction ................................. A53 A.2 Definition of probability .......................... A53 A.3 Combined probabilities ........................... A54 A.4 Probability density function ........................ A54 A.5 Cumulative distribution function ..................... A55 A.6 Marginal and conditional distribution ................... A55 A.7 Expectation value, mean value and variance ............... A56 A.8 Covariance and correlation ......................... A56 A.9 Independent variates, covariance matrix .................. A57 A.10 Change of variables ............................. A57 A.11 Distribution of a function of random variates ............... A58 A.12 Moments of a distribution ......................... A59 A.13 Characteristic function ........................... A60 ii A.14 Cumulants of a distribution ........................ A61 A.15 Probability generating function ...................... A61 B Exercises B62 B.1 Basic statistics exercises .......................... B62 B.2 Uniform random number generation exercises .............. B63 B.3 Generation from arbitrary distributions exercises ............. B64 B.4 Wellknown continuous distributions exercises ............... B65 B.5 Discrete distributions exercises ....................... B68 B.6 Monte Carlo integration exercises ..................... B69 B.7 Application examples ............................ B70 iii 1 Introduction BlaBla ... 1 2 Propagation of measurement errors Error propagation is needed when quantities to be studied must be calculated from directly measurable quantities with given covariance matrix. Let us denote a set of measurements as follows: • Set of measurements: x = (x1, x2, . , xn) • Covariance matrix: Vx, with vij = cov(xi, xj) The components xi represent individual measurements which may be correlated. In case of uncorrelated xi the covariance matrix is diagonal. Furthermore, let us suppose that we are interested in a set of transformed variables which are derived from the measured variables x by some transformation formula: x0 = x0(x) (1) 0 0 0 0 or: x = (x1, x2, . , xm), m ≤ n (2) 0 where the components xi are functions of x. For example: The measured variables could be the Cartesian coordinates x = (x, y, z) and the transformed variables could be the spherical coordinates x0 = (r, ϕ, θ). 0 Now the question is: What is the covariance matrix Vx0 of the new variables x ? This is the problem of the error propagation. The answer is derived by recalling first the definition of the covariance matrix: Z (Vx)ij ≡ cov(xi, xj) ≡ cov(x)ij ≡ σij ≡ vij = (xi − hxii)(xj − hxji)f(x)dx (3) where f(x) is the probability density and where we have listed various notations for the covariance. From the definition it readily follows that cov(axi, bxj) = ab cov(xi, xj) (4) or more generally: X X X cov( aixi, bjxj) = aibjcov(xi, xj) (5) i j i,j 2.1 Error propagation in case of linear transformation We first consider linear transformations x → x0. A linear transformation can be written in matrix notation: x0 = J x (6) where J is an m × n matrix independent of x. According to the definition (3) we can derive the expression for the covariance matrix element i, j: X X (Vx0 )ij = cov( Jikxk , Jjlxl) k l X = JikJjl cov(xk, xl) k,l T = (JVxJ )ij 2 where we have used equation (5). The above result implies the following transformation law for the covariance matrix: T Vx0 = JVx J (7) This the error propagation formula in case of linear transformation of variables x0 = Jx. Examples Pn Example 1: Error estimate of sum of measured quantities s = 1 xi. Error ∆s = ? In matrix notation we write s = Jx where J = (1 1 ··· 1) and it follows that n 2 2 T X X (∆s) ≡ σs = (1 1 ··· 1)Vx(1 1 ··· 1) = σii + 2 σij (8) 1 i>j 2 where we use the fact that covariance matrix is symmetric and the notation (∆xi) = σii. If Vx is diagonal (uncorrelated measurements xi), we get: n 2 X 2 σs = σi (9) 1 2 where our notation is: σii = σi . Example 2: Error estimate of weighted mean P i wixi µ = P . i wi of N quantities. Error estimate σµ = ? Here we assume that xi are independent of each other so that cov(xi, xj) = 0 ↔ i 6= j and 2 σ1 σ2 (0) 2 Vx = .. (10) (0) . 2 σN where all the off-diagonal elements are zero. Now the J matrix is 1 by N: X −1 J = ( wi) (w1 w2 ··· wN ) i and we get 2 X −2 T X −2 X 2 σµ = ( wi) (w1 w2 ··· wN )Vx(w1 w2 ··· wN ) = ( wi) (wiσi) (11) i i i −2 Normally the weights are inverse variances i.e. wi = σi and inserting to the above result we obtain: 1 X 1 = . (12) σ2 σ2 µ i i 3 A special case is that the weights are equal: σi = ∆x for all i. In this case we get N 1 X ∆x µ = x ; σ = √ (13) N i µ 1 N which are the formulae for a simple (unweighted) mean and its error estimation. Example 3: Error estimate of simple sums of three quantities x1, x2 and x3: u1 = x1 + x2 (14) u2 = x2 + x3. (15) Here the transformation is from 3D to 2D so that n = 3 and m = 2. The transformation matrix is: 1 1 0 J = 0 1 1 so that the covariance matrix of u = (u1, u2) is: σ σ σ 1 0 1 1 0 11 12 13 V = σ σ σ 1 1 (16) u 0 1 1 12 22 23 σ13 σ23 σ33 0 1 where the middle most matrix is the covariance matrix Vx of the variables x = (x1, x2, x3). 2.2 Error propagation in case of non-linear transformation We define again a transformation of a set of measurements x as: x0 = x0(x) x = (x1, . , xn) 0 0 0 x = (x1, . , xm) where x0 is under-constrained i.e. m ≤ n. Here x is again an array of measurements with known covariance matrix Vx which is symmetric n × n matrix. The problem is now how to calculate the covariance matrix of quantities x0 i.e. the covariance matrix 0 0 Vx0 of the m quantities x1, . , xm. We expand x0 in Taylor series around the expectation value of x: x = hxi. Each 0 component xi is expanded as: 0 0 0 xi = xi(hxi) + ∇xxi(x = hxi)(x − hxi) + ··· (17) so that the expansion in matrix form reads: x0 = x0(hxi) + J(x − hxi) + ··· (18) where J is now the Jacobian derivative matrix of the transformation: ∂x0 ∂x0 ∇x0 1 ··· 1 1 ∂x1 ∂xn . . .. J = . = . (19) 0 0 0 ∂xm ∂xm ∇xm ··· ∂x1 ∂xn 4 computed at x = hxi. Neglecting higher order terms we have the expansion: x0 = x0(hxi) + Jx − Jhxi. (20) The first and the third terms are constant, because they are calculated at a fixed point x = hxi, so that their covariances vanish and we have cov(x0) = cov(Jx) = J cov(x) JT . Using the V notation for the covariance matrix, the error propagation formula in case of non-linear transfromation reads: T Vx0 = JVx J (21) which is the same formula as in case of linear transformation except that the matrix J is now the Jacobian derivative matrix of the non-linear transformation computed at x = hxi. Examples Example 1: Error estimate of a product u = x1x2. Given the covariance matrix of (x1, x2) what is the error estimate ∆u? The Jacobian is J = ∂u/∂x1 ∂u/∂x2 = x2 x1 so that 2 2 2 σ1 σ12 x2 2 2 2 2 (∆u) ≡ σu = x2 x1 2 = x2σ1 + 2x1x2σ12 + x1σ2.
Recommended publications
  • Duality for Real and Multivariate Exponential Families
    Duality for real and multivariate exponential families a, G´erard Letac ∗ aInstitut de Math´ematiques de Toulouse, 118 route de Narbonne 31062 Toulouse, France. Abstract n Consider a measure µ on R generating a natural exponential family F(µ) with variance function VF(µ)(m) and Laplace transform exp(ℓµ(s)) = exp( s, x )µ(dx). ZRn −h i A dual measure µ∗ satisfies ℓ′ ( ℓ′ (s)) = s. Such a dual measure does not always exist. One important property µ∗ µ 1 − − is ℓ′′ (m) = (VF(µ)(m))− , leading to the notion of duality among exponential families (or rather among the extended µ∗ notion of T exponential families TF obtained by considering all translations of a given exponential family F). Keywords: Dilogarithm distribution, Landau distribution, large deviations, quadratic and cubic real exponential families, Tweedie scale, Wishart distributions. 2020 MSC: Primary 62H05, Secondary 60E10 1. Introduction One can be surprized by the explicit formulas that one gets from large deviations in one dimension. Consider iid random variables X1,..., Xn,... and the empirical mean Xn = (X1 + + Xn)/n. Then the Cram´er theorem [6] says that the limit ··· 1/n Pr(Xn > m) n α(m) → →∞ 1 does exist for m > E(X ). For instance, the symmetric Bernoulli case Pr(Xn = 1) = leads, for 0 < m < 1, to 1 ± 2 1 m 1+m α(m) = (1 + m)− − (1 m)− . (1) − 1 x For the bilateral exponential distribution Xn e dx we get, for m > 0, ∼ 2 −| | 1 1 √1+m2 2 α(m) = e − (1 + √1 + m ). 2 What strange formulas! Other ones can be found in [17], Problem 410.
    [Show full text]
  • Luria-Delbruck Experiment
    Statistical mechanics approaches for studying microbial growth (Lecture 1) Ariel Amir, 10/2019 In the following are included: 1) PDF of lecture slides 2) Calculations related to the Luria-Delbruck experiment. (See also the original paper referenced therein) 3) Calculation of the fixation probability in a serial dilution protocol, taken from SI of: Guo, Vucelja and Amir, Science Advances, 5(7), eaav3842 (2019). Statistical mechanics approaches for studying microbial growth Ariel Amir E. coli Daughter Mother cell cells Time Doubling time/ generation time/growth/replication Taheri-Araghi et al. (2015) Microbial growth 25μm Escherichia coli Saccharomyces cerevisiae Halobacterium salinarum Stewart et al., Soifer al., Eun et al., PLoS Biol (2005), Current Biology (2016) Nat. Micro (2018) Doubling time ~ tens of minutes to several hours (credit: wikipedia) Time Outline (Lecture I) • Why study microbes? Luria-Delbruck experiment, Evolution experiments • Introduction to microbial growth, with focus on cell size regulation (Lecture II) • Size control and correlations across different domains of life • Going from single-cell variability to the population growth (Lecture III) • Bet-hedging • Optimal partitioning of cellular resources Why study microbes? Luria-Delbruck experiment Protocol • Grow culture to ~108 cells. • Expose to bacteriophage. • Count survivors (plating). What to do when the experiment is not reproducible? (variance, standard deviation >> mean) Luria and Delbruck (1943) Why study microbes? Luria-Delbruck experiment Model 1: adaptation
    [Show full text]
  • A Tail Quantile Approximation Formula for the Student T and the Symmetric Generalized Hyperbolic Distribution
    A Service of Leibniz-Informationszentrum econstor Wirtschaft Leibniz Information Centre Make Your Publications Visible. zbw for Economics Schlüter, Stephan; Fischer, Matthias J. Working Paper A tail quantile approximation formula for the student t and the symmetric generalized hyperbolic distribution IWQW Discussion Papers, No. 05/2009 Provided in Cooperation with: Friedrich-Alexander University Erlangen-Nuremberg, Institute for Economics Suggested Citation: Schlüter, Stephan; Fischer, Matthias J. (2009) : A tail quantile approximation formula for the student t and the symmetric generalized hyperbolic distribution, IWQW Discussion Papers, No. 05/2009, Friedrich-Alexander-Universität Erlangen-Nürnberg, Institut für Wirtschaftspolitik und Quantitative Wirtschaftsforschung (IWQW), Nürnberg This Version is available at: http://hdl.handle.net/10419/29554 Standard-Nutzungsbedingungen: Terms of use: Die Dokumente auf EconStor dürfen zu eigenen wissenschaftlichen Documents in EconStor may be saved and copied for your Zwecken und zum Privatgebrauch gespeichert und kopiert werden. personal and scholarly purposes. Sie dürfen die Dokumente nicht für öffentliche oder kommerzielle You are not to copy documents for public or commercial Zwecke vervielfältigen, öffentlich ausstellen, öffentlich zugänglich purposes, to exhibit the documents publicly, to make them machen, vertreiben oder anderweitig nutzen. publicly available on the internet, or to distribute or otherwise use the documents in public. Sofern die Verfasser die Dokumente unter Open-Content-Lizenzen (insbesondere CC-Lizenzen) zur Verfügung gestellt haben sollten, If the documents have been made available under an Open gelten abweichend von diesen Nutzungsbedingungen die in der dort Content Licence (especially Creative Commons Licences), you genannten Lizenz gewährten Nutzungsrechte. may exercise further usage rights as specified in the indicated licence. www.econstor.eu IWQW Institut für Wirtschaftspolitik und Quantitative Wirtschaftsforschung Diskussionspapier Discussion Papers No.
    [Show full text]
  • Algorithms for Operations on Probability Distributions in a Computer Algebra System
    W&M ScholarWorks Dissertations, Theses, and Masters Projects Theses, Dissertations, & Master Projects 2001 Algorithms for operations on probability distributions in a computer algebra system Diane Lynn Evans College of William & Mary - Arts & Sciences Follow this and additional works at: https://scholarworks.wm.edu/etd Part of the Mathematics Commons, and the Statistics and Probability Commons Recommended Citation Evans, Diane Lynn, "Algorithms for operations on probability distributions in a computer algebra system" (2001). Dissertations, Theses, and Masters Projects. Paper 1539623382. https://dx.doi.org/doi:10.21220/s2-bath-8582 This Dissertation is brought to you for free and open access by the Theses, Dissertations, & Master Projects at W&M ScholarWorks. It has been accepted for inclusion in Dissertations, Theses, and Masters Projects by an authorized administrator of W&M ScholarWorks. For more information, please contact [email protected]. Reproduced with with permission permission of the of copyright the copyright owner. owner.Further reproductionFurther reproduction prohibited without prohibited permission. without permission. ALGORITHMS FOR OPERATIONS ON PROBABILITY DISTRIBUTIONS IN A COMPUTER ALGEBRA SYSTEM A Dissertation Presented to The Faculty of the Department of Applied Science The College of William & Mary in Virginia In Partial Fulfillment Of the Requirements for the Degree of Doctor of Philosophy by Diane Lynn Evans July 2001 Reproduced with permission of the copyright owner. Further reproduction prohibited without permission. UMI Number: 3026405 Copyright 2001 by Evans, Diane Lynn All rights reserved. ___ ® UMI UMI Microform 3026405 Copyright 2001 by Bell & Howell Information and Learning Company. All rights reserved. This microform edition is protected against unauthorized copying under Title 17, United States Code.
    [Show full text]
  • Generating Random Samples from User-Defined Distributions
    The Stata Journal (2011) 11, Number 2, pp. 299–304 Generating random samples from user-defined distributions Katar´ına Luk´acsy Central European University Budapest, Hungary lukacsy [email protected] Abstract. Generating random samples in Stata is very straightforward if the distribution drawn from is uniform or normal. With any other distribution, an inverse method can be used; but even in this case, the user is limited to the built- in functions. For any other distribution functions, their inverse must be derived analytically or numerical methods must be used if analytical derivation of the inverse function is tedious or impossible. In this article, I introduce a command that generates a random sample from any user-specified distribution function using numeric methods that make this command very generic. Keywords: st0229, rsample, random sample, user-defined distribution function, inverse method, Monte Carlo exercise 1 Introduction In statistics, a probability distribution identifies the probability of a random variable. If the random variable is discrete, it identifies the probability of each value of this vari- able. If the random variable is continuous, it defines the probability that this variable’s value falls within a particular interval. The probability distribution describes the range of possible values that a random variable can attain, further referred to as the sup- port interval, and the probability that the value of the random variable is within any (measurable) subset of that range. Random sampling refers to taking a number of independent observations from a probability distribution. Typically, the parameters of the probability distribution (of- ten referred to as true parameters) are unknown, and the aim is to retrieve them using various estimation methods on the random sample generated from this probability dis- tribution.
    [Show full text]
  • Random Variables Generation
    Random Variables Generation Revised version of the slides based on the book Discrete-Event Simulation: a first course L.L. Leemis & S.K. Park Section(s) 6.1, 6.2, 7.1, 7.2 c 2006 Pearson Ed., Inc. 0-13-142917-5 Discrete-Event Simulation Random Variables Generation 1/80 Introduction Monte Carlo Simulators differ from Trace Driven simulators because of the use of Random Number Generators to represent the variability that affects the behaviors of real systems. Uniformly distributed random variables are the most elementary representations that we can use in Monte Carlo simulation, but are not enough to capture the complexity of real systems. We must thus devise methods for generating instances (variates) of arbitrary random variables Properly using uniform random numbers, it is possible to obtain this result. In the sequel we will first recall some basic properties of Discrete and Continuous random variables and then we will discuss several methods to obtain their variates Discrete-Event Simulation Random Variables Generation 2/80 Basic Probability Concepts Empirical Probability, derives from performing an experiment many times n and counting the number of occurrences na of an event A The relative frequency of occurrence of event is na/n The frequency theory of probability asserts thatA the relative frequency converges as n → ∞ n Pr( )= lim a A n→∞ n Axiomatic Probability is a formal, set-theoretic approach Mathematically construct the sample space and calculate the number of events A The two are complementary! Discrete-Event Simulation Random
    [Show full text]
  • On Half-Cauchy Distribution and Process
    International Journal of Statistika and Mathematika, ISSN: 2277- 2790 E-ISSN: 2249-8605, Volume 3, Issue 2, 2012 pp 77-81 On Half-Cauchy Distribution and Process Elsamma Jacob1 , K Jayakumar2 1Malabar Christian College, Calicut, Kerala, INDIA. 2Department of Statistics, University of Calicut, Kerala, INDIA. Corresponding addresses: [email protected], [email protected] Research Article Abstract: A new form of half- Cauchy distribution using sin ξ cos ξ Marshall-Olkin transformation is introduced. The properties of () = − ξ, () = ξ, ≥ 0 the new distribution such as density, cumulative distribution ξ ξ function, quantiles, measure of skewness and distribution of the extremes are obtained. Time series models with half-Cauchy Remark 1.1. For the HC distribution the moments do distribution as stationary marginal distributions are not not exist. developed so far. We develop first order autoregressive process Remark 1.2. The HC distribution is infinitely divisible with the new distribution as stationary marginal distribution and (Bondesson (1987)) and self decomposable (Diedhiou the properties of the process are studied. Application of the (1998)). distribution in various fields is also discussed. Keywords: Autoregressive Process, Geometric Extreme Stable, Relationship with other distributions: Half-Cauchy Distribution, Skewness, Stationarity, Quantiles. 1. Let Y be a folded t variable with pdf given by 1. Introduction: ( ) > 0, (1.5) The half Cauchy (HC) distribution is derived from 2Γ( ) () = 1 + , ν ∈ , the standard Cauchy distribution by folding the ν ()√νπ curve on the origin so that only positive values can be observed. A continuous random variable X is said When ν = 1 , (1.5) reduces to 2 1 > 0 to have the half Cauchy distribution if its survival () = , function is given by 1 + (x)=1 − tan , > 0 (1.1) Thus, HC distribution coincides with the folded t distribution with = 1 degree of freedom.
    [Show full text]
  • Hand-Book on STATISTICAL DISTRIBUTIONS for Experimentalists
    Internal Report SUF–PFY/96–01 Stockholm, 11 December 1996 1st revision, 31 October 1998 last modification 10 September 2007 Hand-book on STATISTICAL DISTRIBUTIONS for experimentalists by Christian Walck Particle Physics Group Fysikum University of Stockholm (e-mail: [email protected]) Contents 1 Introduction 1 1.1 Random Number Generation .............................. 1 2 Probability Density Functions 3 2.1 Introduction ........................................ 3 2.2 Moments ......................................... 3 2.2.1 Errors of Moments ................................ 4 2.3 Characteristic Function ................................. 4 2.4 Probability Generating Function ............................ 5 2.5 Cumulants ......................................... 6 2.6 Random Number Generation .............................. 7 2.6.1 Cumulative Technique .............................. 7 2.6.2 Accept-Reject technique ............................. 7 2.6.3 Composition Techniques ............................. 8 2.7 Multivariate Distributions ................................ 9 2.7.1 Multivariate Moments .............................. 9 2.7.2 Errors of Bivariate Moments .......................... 9 2.7.3 Joint Characteristic Function .......................... 10 2.7.4 Random Number Generation .......................... 11 3 Bernoulli Distribution 12 3.1 Introduction ........................................ 12 3.2 Relation to Other Distributions ............................. 12 4 Beta distribution 13 4.1 Introduction .......................................
    [Show full text]
  • NAG Library Chapter Introduction G01 – Simple Calculations on Statistical Data
    g01 – Simple Calculations on Statistical Data Introduction – g01 NAG Library Chapter Introduction g01 – Simple Calculations on Statistical Data Contents 1 Scope of the Chapter.................................................... 2 2 Background to the Problems ............................................ 2 2.1 Summary Statistics .................................................... 2 2.2 Statistical Distribution Functions and Their Inverses........................ 2 2.3 Testing for Normality and Other Distributions ............................. 3 2.4 Distribution of Quadratic Forms......................................... 3 2.5 Energy Loss Distributions .............................................. 3 2.6 Vectorized Functions .................................................. 4 3 Recommendations on Choice and Use of Available Functions ............ 4 3.1 Working with Streamed or Extremely Large Datasets ....................... 7 4 Auxiliary Functions Associated with Library Function Arguments ....... 7 5 Functions Withdrawn or Scheduled for Withdrawal ..................... 7 6 References............................................................... 7 Mark 24 g01.1 Introduction – g01 NAG Library Manual 1 Scope of the Chapter This chapter covers three topics: summary statistics statistical distribution functions and their inverses; testing for Normality and other distributions. 2 Background to the Problems 2.1 Summary Statistics The summary statistics consist of two groups. The first group are those based on moments; for example mean, standard
    [Show full text]
  • Luca Lista Statistical Methods for Data Analysis in Particle Physics Lecture Notes in Physics
    Lecture Notes in Physics 909 Luca Lista Statistical Methods for Data Analysis in Particle Physics Lecture Notes in Physics Volume 909 Founding Editors W. Beiglböck J. Ehlers K. Hepp H. Weidenmüller Editorial Board M. Bartelmann, Heidelberg, Germany B.-G. Englert, Singapore, Singapore P. Hänggi, Augsburg, Germany M. Hjorth-Jensen, Oslo, Norway R.A.L. Jones, Sheffield, UK M. Lewenstein, Barcelona, Spain H. von Löhneysen, Karlsruhe, Germany J.-M. Raimond, Paris, France A. Rubio, Donostia, San Sebastian, Spain S. Theisen, Potsdam, Germany D. Vollhardt, Augsburg, Germany J.D. Wells, Ann Arbor, USA G.P. Zank, Huntsville, USA The Lecture Notes in Physics The series Lecture Notes in Physics (LNP), founded in 1969, reports new devel- opments in physics research and teaching-quickly and informally, but with a high quality and the explicit aim to summarize and communicate current knowledge in an accessible way. Books published in this series are conceived as bridging material between advanced graduate textbooks and the forefront of research and to serve three purposes: • to be a compact and modern up-to-date source of reference on a well-defined topic • to serve as an accessible introduction to the field to postgraduate students and nonspecialist researchers from related areas • to be a source of advanced teaching material for specialized seminars, courses and schools Both monographs and multi-author volumes will be considered for publication. Edited volumes should, however, consist of a very limited number of contributions only. Proceedings will not be considered for LNP. Volumes published in LNP are disseminated both in print and in electronic for- mats, the electronic archive being available at springerlink.com.
    [Show full text]
  • Handbook on Probability Distributions
    R powered R-forge project Handbook on probability distributions R-forge distributions Core Team University Year 2009-2010 LATEXpowered Mac OS' TeXShop edited Contents Introduction 4 I Discrete distributions 6 1 Classic discrete distribution 7 2 Not so-common discrete distribution 27 II Continuous distributions 34 3 Finite support distribution 35 4 The Gaussian family 47 5 Exponential distribution and its extensions 56 6 Chi-squared's ditribution and related extensions 75 7 Student and related distributions 84 8 Pareto family 88 9 Logistic distribution and related extensions 108 10 Extrem Value Theory distributions 111 3 4 CONTENTS III Multivariate and generalized distributions 116 11 Generalization of common distributions 117 12 Multivariate distributions 133 13 Misc 135 Conclusion 137 Bibliography 137 A Mathematical tools 141 Introduction This guide is intended to provide a quite exhaustive (at least as I can) view on probability distri- butions. It is constructed in chapters of distribution family with a section for each distribution. Each section focuses on the tryptic: definition - estimation - application. Ultimate bibles for probability distributions are Wimmer & Altmann (1999) which lists 750 univariate discrete distributions and Johnson et al. (1994) which details continuous distributions. In the appendix, we recall the basics of probability distributions as well as \common" mathe- matical functions, cf. section A.2. And for all distribution, we use the following notations • X a random variable following a given distribution, • x a realization of this random variable, • f the density function (if it exists), • F the (cumulative) distribution function, • P (X = k) the mass probability function in k, • M the moment generating function (if it exists), • G the probability generating function (if it exists), • φ the characteristic function (if it exists), Finally all graphics are done the open source statistical software R and its numerous packages available on the Comprehensive R Archive Network (CRAN∗).
    [Show full text]
  • Field Guide to Continuous Probability Distributions
    Field Guide to Continuous Probability Distributions Gavin E. Crooks v 1.0.0 2019 G. E. Crooks – Field Guide to Probability Distributions v 1.0.0 Copyright © 2010-2019 Gavin E. Crooks ISBN: 978-1-7339381-0-5 http://threeplusone.com/fieldguide Berkeley Institute for Theoretical Sciences (BITS) typeset on 2019-04-10 with XeTeX version 0.99999 fonts: Trump Mediaeval (text), Euler (math) 271828182845904 2 G. E. Crooks – Field Guide to Probability Distributions Preface: The search for GUD A common problem is that of describing the probability distribution of a single, continuous variable. A few distributions, such as the normal and exponential, were discovered in the 1800’s or earlier. But about a century ago the great statistician, Karl Pearson, realized that the known probabil- ity distributions were not sufficient to handle all of the phenomena then under investigation, and set out to create new distributions with useful properties. During the 20th century this process continued with abandon and a vast menagerie of distinct mathematical forms were discovered and invented, investigated, analyzed, rediscovered and renamed, all for the purpose of de- scribing the probability of some interesting variable. There are hundreds of named distributions and synonyms in current usage. The apparent diver- sity is unending and disorienting. Fortunately, the situation is less confused than it might at first appear. Most common, continuous, univariate, unimodal distributions can be orga- nized into a small number of distinct families, which are all special cases of a single Grand Unified Distribution. This compendium details these hun- dred or so simple distributions, their properties and their interrelations.
    [Show full text]