ACTS 4304 FORMULA SUMMARY Lesson 1: Basic Probability Summary of Probability Concepts Probability Functions

Total Page:16

File Type:pdf, Size:1020Kb

ACTS 4304 FORMULA SUMMARY Lesson 1: Basic Probability Summary of Probability Concepts Probability Functions ACTS 4304 FORMULA SUMMARY Lesson 1: Basic Probability Summary of Probability Concepts Probability Functions F (x) = P r(X ≤ x) S(x) = 1 − F (x) dF (x) f(x) = dx H(x) = − ln S(x) dH(x) f(x) h(x) = = dx S(x) Functions of random variables Z 1 Expected Value E[g(x)] = g(x)f(x)dx −∞ 0 n n-th raw moment µn = E[X ] n n-th central moment µn = E[(X − µ) ] Variance σ2 = E[(X − µ)2] = E[X2] − µ2 µ µ0 − 3µ0 µ + 2µ3 Skewness γ = 3 = 3 2 1 σ3 σ3 µ µ0 − 4µ0 µ + 6µ0 µ2 − 3µ4 Kurtosis γ = 4 = 4 3 2 2 σ4 σ4 Moment generating function M(t) = E[etX ] Probability generating function P (z) = E[zX ] More concepts • Standard deviation (σ) is positive square root of variance • Coefficient of variation is CV = σ/µ • 100p-th percentile π is any point satisfying F (π−) ≤ p and F (π) ≥ p. If F is continuous, it is the unique point satisfying F (π) = p • Median is 50-th percentile; n-th quartile is 25n-th percentile • Mode is x which maximizes f(x) (n) n (n) • MX (0) = E[X ], where M is the n-th derivative (n) PX (0) • n! = P r(X = n) (n) • PX (1) is the n-th factorial moment of X. Bayes' Theorem P r(BjA)P r(A) P r(AjB) = P r(B) fY (yjx)fX (x) fX (xjy) = fY (y) Law of total probability 2 If Bi is a set of exhaustive (in other words, P r([iBi) = 1) and mutually exclusive (in other words P r(Bi \ Bj) = 0 for i 6= j) events, then for any event A, X X P r(A) = P r(A \ Bi) = P r(Bi)P r(AjBi) i i Correspondingly, for continuous distributions, Z P r(A) = P r(Ajx)f(x)dx Conditional Expectation Formula EX [X] = EY [EX [XjY ]] 3 Lesson 2: Parametric Distributions Forms of probability density functions for common distributions Distribution Probability density function f(x) Uniform c; x 2 [d; u] Beta cxa−1(θ − x)b−1; x 2 [0; θ] − x Exponential ce θ ; x ≥ 0 Weibull cxτ−1e−xτ /θτ ; x ≥ 0 α−1 − x Gamma cx e θ ; x ≥ 0 c Pareto (x+θ)α+1 ; x ≥ 0 c Single-parameter Pareto xα+1 ; x ≥ θ 2 2 ce−(ln x−µ) =2σ Lognormal x ; x > 0 Summary of Parametric Distribution Concepts • If X is a member of a scale family with scale parameter θ with value s, then cX is the same family and has the same parameter values as X except that the scale parameter θ has value cs. • All distributions in the tables are scale families with scale parameter θ except for lognormal and inverse Gaussian. • If X is lognormal with parameters µ and σ, then cX is lognormal with parameters µ + ln c and σ. • See the above table to learn the forms of commonly occurring distributions. Useful facts are d + u (u − d)2 Uniform on [d, u] E[X] = ; V arX = 2 12 u2 Uniform on [0, u] E[X2] = 3 Gamma V arX = αθ2 • If Y is single-parameter Pareto with parameters α and θ, then Y − θ is two-parameter Pareto with parameters. • X is in the linear exponential family if its probability density function can be expressed as p(x)er(θ)x f(x; θ) = q(θ) 4 Lesson 3: Variance For any random variables X and Y; E[aX + bY ] = aE[X] + bE[Y ] V ar(aX + bY ) = a2V ar(X) + 2abCov(X; Y ) + b2V ar(Y ) n ! n X X For independent random variables X1;X2; ··· ;Xn; V ar Xi = V ar(Xi) i=1 i=1 For independent identically distributed random variables (i.i.d.) X1;X2; ··· ;Xn; n ! X V ar Xi = nV ar(X) i=1 n 1 X The sample mean X¯ = X n i i=1 1 The variance of the sample mean V ar(X¯) = V ar(X) n Double expectation EX [X] = EY [EX [XjY ]] Conditional variance V arX [X] = EY [V arX [XjY ]] + V arY (EX [XjY ]) 5 Lesson 4: Mixtures and Splices Pn • If X is a mixture of n random variables with weights wi such that i=1 wi = 1, then the following can be expressed as a weighted average: n X Cumulative distribution function: FX (x) = wiFXi (x) i=1 n X Probability density function: fX (x) = wifXi (x) i=1 n k X k k-th raw moment: E[X ] = wiE[Xi ] i=1 • Conditional variance: V arX [X] = EI [V arX [XjI]] + V arI (EX [XjI]) • Splices: For a spliced distribution, the sum of the probabilities of being in each splice must add up to 1. 6 Lesson 5: Policy Limits All formulas assume P r(X < 0) = 0. Z 1 E[X] = S(x) dx 0 Z u Z u E[X ^ u] = xf(x) dx + u (1 − F (u)) = S(x) dx 0 0 Z 1 E[Xk] = kxk−1S(x) dx 0 Z u Z u E[(X ^ u)k] = xkf(x) dx + uk (1 − F (u)) = kxk−1S(x) dx 0 0 For inflation, if Y = (1 + r)X, then u E[Y ^ u] = (1 + r)E X ^ 1 + r 7 Lesson 6: Deductibles Payment per Loss: L FY L (x) = FX (x + d); if Y = (X − d)+ Z 1 E[(X − d)+] = (x − d)f(x) dx d Z 1 E[(X − d)+] = S(x) dx d E[X] = E[X ^ d] + E[(X − d)+] Payment per Payment: FX (x + d) − FX (d) P FY P (x) = ; if Y = (X − d)+jX > d 1 − FX (d) SX (x + d) P SY P (x) = ; if Y = (X − d)+jX > d SX (d) E[(X − d) ] E[X] − E[X ^ d] e (d) = + = − mean excess loss X S(d) S(d) R 1(x − d)f(x) dx e (d) = d X S(d) R 1 S(x) dx e (d) = d X S(d) E[X] = E[X ^ d] + e(d) (1 − F (d)) Mean excess loss for different distributions: eX (d) = θ for exponential θ − d e (d) = ; d < θ for uniform on [0; θ] X 2 θ − d e (d) = ; d < θ for beta with parameters a = 1; b; θ X 1 + b θ + d e (d) = for two-parameter Pareto X α − 1 ( d α−1 d ≥ θ eX (d) = α(θ−d)+d for single-parameter Pareto α−1 d ≤ θ If Y L;Y P are loss and payment random variables for franchise deductible of d, and XL;XP are loss and payment random variables for ordinary deductible of d, then E[Y L] = E[XL] + dS(d) E[Y P ] = E[XP ] + d 8 Lesson 7: Loss Elimination Ratio The Loss Elimination Ratio is defined as the proportion of the expected loss which the insurer doesn't pay as a result of an ordinary deductible d: E[X ^ d] E[(X − d) ] LER(d) = = 1 − + E[X] E[X] Loss Elimination Ratio for Certain Distributions: LER(d) = 1 − e−d/θ for an exponential θ α−1 LER(d) = 1 − for a Pareto with α > 1 d + θ (θ=d)α−1 LER(d) = 1 − for a single-parameter Pareto with α > 1; d ≥ θ α 9 Lesson 8: Risk Measures and Tail Weight −1 Value-at-Risk: V aRp(X) = πp = FX (p) Tail-Value-at-Risk: R 1 xf(x) dx V aRp(X) T V aRp(X) = E [XjX > V aRp(X)] = = 1 − F (V aRp(X)) R 1 V aRy(X) dy = p = V aR (X) + e (V aR (X)) = 1 − p p X p E[X] − E[X ^ V aR (X)] = V aR (X) + p p 1 − p Value-at-Risk and Tail-Value-at-Risk measures for some distributions: Distribution V aRp(X) T V aRp(X) Exponential −θ ln(1 − p) θ (1 − ln(1 − p)) − 1 − 1 Pareto θ (1 − p) α − 1 E[X] 1 + α (1 − p) α − 1 2 −zp=2 Normal µ + z σ µ + σ · ep p 1−p 2π µ+zpσ Φ(σ−zp) Lognormal e E[X] · 1−p 10 Lesson 9: Other Topics in Severity Coverage Modifications Policy limit - the maximum amount that the coverage will pay. In the presence of a deductible or other modifications, perform the other modifications first, then the policy limit. Maximum coverage loss is the stipulated amount considered in calculating the payment. Apply this limit first, and then the deductible. If u is the maximum coverage loss and d - the deductible, then Y L = X ^ u − X ^ d Coinsurance of α is the portion of each loss reimbursed by insurance. In the presence of the three modifications, E[Y L] = α (E[X ^ u] − E[X ^ d]) If r is the inflation factor, u d E[Y L] = α(1 + r) E X ^ − E X ^ 1 + r 1 + r 11 Lesson 10: Bonuses A typical bonus is a portion of the excess of r% of premiums over losses. If c is the portion of the excess, r is the loss ratio, P is earned premium, and X is losses, then B = max (0; c(rP − X)) = crP − c min(rP; X) = crP − c (X ^ rP ) For a two-parameter Pareto distribution with α = 2 and θ, θd E[X ^ d] = d + θ 12 Lesson 11: Discrete Distributions For a (a; b; 0) class distributions, pk b = a + ; pk = P r(X = k) pk−1 k Poisson Binomial Negative binomial Geometric n r n λ m n+r−1 β βn −λ· n! n m−n 1 pn e n q (1 − q) n 1+β 1+β (1+β)n+1 Mean λ mq rβ β Variance λ mq(1 − q) rβ(1 + β) β(1 + β) q β β a 0 − 1−q 1+β 1+β q β b λ (m + 1) 1−q (r − 1) 1+β 0 For a (a; b; 1) class distributions, p0 is arbitrary and p b k = a + for k = 2; 3; 4; ··· pk−1 k Zero-truncated distributions: T pn pn = ; n > 0 1 − p0 Zero-modified distributions: M M T pn = (1 − p0 )pn E[N] = cm V ar(N) = c(1 − c)m2 + cv; where M • c is 1 − p0 • m is the mean of the corresponding zero-truncated distribution • v is the variance of the corresponding zero-truncated distribution 13 Lesson 12: Poisson/Gamma Assume that in a portfolio of insureds, loss frequency follows a Poisson distribution with parameter λ, but λ is not fixed but varies by insured.
Recommended publications
  • Modelling Censored Losses Using Splicing: a Global Fit Strategy With
    Modelling censored losses using splicing: a global fit strategy with mixed Erlang and extreme value distributions Tom Reynkens∗a, Roel Verbelenb, Jan Beirlanta,c, and Katrien Antoniob,d aLStat and LRisk, Department of Mathematics, KU Leuven. Celestijnenlaan 200B, 3001 Leuven, Belgium. bLStat and LRisk, Faculty of Economics and Business, KU Leuven. Naamsestraat 69, 3000 Leuven, Belgium. cDepartment of Mathematical Statistics and Actuarial Science, University of the Free State. P.O. Box 339, Bloemfontein 9300, South Africa. dFaculty of Economics and Business, University of Amsterdam. Roetersstraat 11, 1018 WB Amsterdam, The Netherlands. August 14, 2017 Abstract In risk analysis, a global fit that appropriately captures the body and the tail of the distribution of losses is essential. Modelling the whole range of the losses using a standard distribution is usually very hard and often impossible due to the specific characteristics of the body and the tail of the loss distribution. A possible solution is to combine two distributions in a splicing model: a light-tailed distribution for the body which covers light and moderate losses, and a heavy-tailed distribution for the tail to capture large losses. We propose a splicing model with a mixed Erlang (ME) distribution for the body and a Pareto distribution for the tail. This combines the arXiv:1608.01566v4 [stat.ME] 11 Aug 2017 flexibility of the ME distribution with the ability of the Pareto distribution to model extreme values. We extend our splicing approach for censored and/or truncated data. Relevant examples of such data can be found in financial risk analysis. We illustrate the flexibility of this splicing model using practical examples from risk measurement.
    [Show full text]
  • A New Parameter Estimator for the Generalized Pareto Distribution Under the Peaks Over Threshold Framework
    mathematics Article A New Parameter Estimator for the Generalized Pareto Distribution under the Peaks over Threshold Framework Xu Zhao 1,*, Zhongxian Zhang 1, Weihu Cheng 1 and Pengyue Zhang 2 1 College of Applied Sciences, Beijing University of Technology, Beijing 100124, China; [email protected] (Z.Z.); [email protected] (W.C.) 2 Department of Biomedical Informatics, College of Medicine, The Ohio State University, Columbus, OH 43210, USA; [email protected] * Correspondence: [email protected] Received: 1 April 2019; Accepted: 30 April 2019 ; Published: 7 May 2019 Abstract: Techniques used to analyze exceedances over a high threshold are in great demand for research in economics, environmental science, and other fields. The generalized Pareto distribution (GPD) has been widely used to fit observations exceeding the tail threshold in the peaks over threshold (POT) framework. Parameter estimation and threshold selection are two critical issues for threshold-based GPD inference. In this work, we propose a new GPD-based estimation approach by combining the method of moments and likelihood moment techniques based on the least squares concept, in which the shape and scale parameters of the GPD can be simultaneously estimated. To analyze extreme data, the proposed approach estimates the parameters by minimizing the sum of squared deviations between the theoretical GPD function and its expectation. Additionally, we introduce a recently developed stopping rule to choose the suitable threshold above which the GPD asymptotically fits the exceedances. Simulation studies show that the proposed approach performs better or similar to existing approaches, in terms of bias and the mean square error, in estimating the shape parameter.
    [Show full text]
  • Independent Approximates Enable Closed-Form Estimation of Heavy-Tailed Distributions
    Independent Approximates enable closed-form estimation of heavy-tailed distributions Kenric P. Nelson Abstract – Independent Approximates (IAs) are proven to enable a closed-form estimation of heavy-tailed distributions with an analytical density such as the generalized Pareto and Student’s t distributions. A broader proof using convolution of the characteristic function is described for future research. (IAs) are selected from independent, identically distributed samples by partitioning the samples into groups of size n and retaining the median of the samples in those groups which have approximately equal samples. The marginal distribution along the diagonal of equal values has a density proportional to the nth power of the original density. This nth-power-density, which the IAs approximate, has faster tail decay enabling closed-form estimation of its moments and retains a functional relationship with the original density. Computational experiments with between 1000 to 100,000 Student’s t samples are reported for over a range of location, scale, and shape (inverse of degree of freedom) parameters. IA pairs are used to estimate the location, IA triplets for the scale, and the geometric mean of the original samples for the shape. With 10,000 samples the relative bias of the parameter estimates is less than 0.01 and a relative precision is less than ±0.1. The theoretical bias is zero for the location and the finite bias for the scale can be subtracted out. The theoretical precision has a finite range when the shape is less than 2 for the location estimate and less than 3/2 for the scale estimate.
    [Show full text]
  • STAT 830 the Basics of Nonparametric Models The
    STAT 830 The basics of nonparametric models The Empirical Distribution Function { EDF The most common interpretation of probability is that the probability of an event is the long run relative frequency of that event when the basic experiment is repeated over and over independently. So, for instance, if X is a random variable then P (X ≤ x) should be the fraction of X values which turn out to be no more than x in a long sequence of trials. In general an empirical probability or expected value is just such a fraction or average computed from the data. To make this precise, suppose we have a sample X1;:::;Xn of iid real valued random variables. Then we make the following definitions: Definition: The empirical distribution function, or EDF, is n 1 X F^ (x) = 1(X ≤ x): n n i i=1 This is a cumulative distribution function. It is an estimate of F , the cdf of the Xs. People also speak of the empirical distribution of the sample: n 1 X P^(A) = 1(X 2 A) n i i=1 ^ This is the probability distribution whose cdf is Fn. ^ Now we consider the qualities of Fn as an estimate, the standard error of the estimate, the estimated standard error, confidence intervals, simultaneous confidence intervals and so on. To begin with we describe the best known summaries of the quality of an estimator: bias, variance, mean squared error and root mean squared error. Bias, variance, MSE and RMSE There are many ways to judge the quality of estimates of a parameter φ; all of them focus on the distribution of the estimation error φ^−φ.
    [Show full text]
  • Identifying Probability Distributions of Key Variables in Sow Herds
    Iowa State University Capstones, Theses and Creative Components Dissertations Summer 2020 Identifying Probability Distributions of Key Variables in Sow Herds Hao Tong Iowa State University Follow this and additional works at: https://lib.dr.iastate.edu/creativecomponents Part of the Veterinary Preventive Medicine, Epidemiology, and Public Health Commons Recommended Citation Tong, Hao, "Identifying Probability Distributions of Key Variables in Sow Herds" (2020). Creative Components. 617. https://lib.dr.iastate.edu/creativecomponents/617 This Creative Component is brought to you for free and open access by the Iowa State University Capstones, Theses and Dissertations at Iowa State University Digital Repository. It has been accepted for inclusion in Creative Components by an authorized administrator of Iowa State University Digital Repository. For more information, please contact [email protected]. 1 Identifying Probability Distributions of Key Variables in Sow Herds Hao Tong, Daniel C. L. Linhares, and Alejandro Ramirez Iowa State University, College of Veterinary Medicine, Veterinary Diagnostic and Production Animal Medicine Department, Ames, Iowa, USA Abstract Figuring out what probability distribution your data fits is critical for data analysis as statistical assumptions must be met for specific tests used to compare data. However, most studies with swine populations seldom report information about the distribution of the data. In most cases, sow farm production data are treated as having a normal distribution even when they are not. We conducted this study to describe the most common probability distributions in sow herd production data to help provide guidance for future data analysis. In this study, weekly production data from January 2017 to June 2019 were included involving 47 different sow farms.
    [Show full text]
  • Fixed-K Asymptotic Inference About Tail Properties
    Journal of the American Statistical Association ISSN: 0162-1459 (Print) 1537-274X (Online) Journal homepage: http://www.tandfonline.com/loi/uasa20 Fixed-k Asymptotic Inference About Tail Properties Ulrich K. Müller & Yulong Wang To cite this article: Ulrich K. Müller & Yulong Wang (2017) Fixed-k Asymptotic Inference About Tail Properties, Journal of the American Statistical Association, 112:519, 1334-1343, DOI: 10.1080/01621459.2016.1215990 To link to this article: https://doi.org/10.1080/01621459.2016.1215990 Accepted author version posted online: 12 Aug 2016. Published online: 13 Jun 2017. Submit your article to this journal Article views: 206 View related articles View Crossmark data Full Terms & Conditions of access and use can be found at http://www.tandfonline.com/action/journalInformation?journalCode=uasa20 JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION , VOL. , NO. , –, Theory and Methods https://doi.org/./.. Fixed-k Asymptotic Inference About Tail Properties Ulrich K. Müller and Yulong Wang Department of Economics, Princeton University, Princeton, NJ ABSTRACT ARTICLE HISTORY We consider inference about tail properties of a distribution from an iid sample, based on extreme value the- Received February ory. All of the numerous previous suggestions rely on asymptotics where eventually, an infinite number of Accepted July observations from the tail behave as predicted by extreme value theory, enabling the consistent estimation KEYWORDS of the key tail index, and the construction of confidence intervals using the delta method or other classic Extreme quantiles; tail approaches. In small samples, however, extreme value theory might well provide good approximations for conditional expectations; only a relatively small number of tail observations.
    [Show full text]
  • Package 'Renext'
    Package ‘Renext’ July 4, 2016 Type Package Title Renewal Method for Extreme Values Extrapolation Version 3.1-0 Date 2016-07-01 Author Yves Deville <[email protected]>, IRSN <[email protected]> Maintainer Lise Bardet <[email protected]> Depends R (>= 2.8.0), stats, graphics, evd Imports numDeriv, splines, methods Suggests MASS, ismev, XML Description Peaks Over Threshold (POT) or 'methode du renouvellement'. The distribution for the ex- ceedances can be chosen, and heterogeneous data (including histori- cal data or block data) can be used in a Maximum-Likelihood framework. License GPL (>= 2) LazyData yes NeedsCompilation no Repository CRAN Date/Publication 2016-07-04 20:29:04 R topics documented: Renext-package . .3 anova.Renouv . .5 barplotRenouv . .6 Brest . .8 Brest.years . 10 Brest.years.missing . 11 CV2............................................. 11 CV2.test . 12 Dunkerque . 13 EM.mixexp . 15 expplot . 16 1 2 R topics documented: fgamma . 17 fGEV.MAX . 19 fGPD ............................................ 22 flomax . 24 fmaxlo . 26 fweibull . 29 Garonne . 30 gev2Ren . 32 gof.date . 34 gofExp.test . 37 GPD............................................. 38 gumbel2Ren . 40 Hpoints . 41 ini.mixexp2 . 42 interevt . 43 Jackson . 45 Jackson.test . 46 logLik.Renouv . 47 Lomax . 48 LRExp . 50 LRExp.test . 51 LRGumbel . 53 LRGumbel.test . 54 Maxlo . 55 MixExp2 . 57 mom.mixexp2 . 58 mom2par . 60 NBlevy . 61 OT2MAX . 63 OTjitter . 66 parDeriv . 67 parIni.MAX . 70 pGreenwood1 . 72 plot.Rendata . 73 plot.Renouv . 74 PPplot . 78 predict.Renouv . 79 qStat . 80 readXML . 82 Ren2gev . 84 Ren2gumbel . 85 Renouv . 87 RenouvNoEst . 94 RLlegend . 96 RLpar . 98 RLplot . 99 roundPred . 101 rRendata . 102 Renext-package 3 SandT . 105 skip2noskip .
    [Show full text]
  • Multivariate Phase Type Distributions - Applications and Parameter Estimation
    Downloaded from orbit.dtu.dk on: Sep 28, 2021 Multivariate phase type distributions - Applications and parameter estimation Meisch, David Publication date: 2014 Document Version Publisher's PDF, also known as Version of record Link back to DTU Orbit Citation (APA): Meisch, D. (2014). Multivariate phase type distributions - Applications and parameter estimation. Technical University of Denmark. DTU Compute PHD-2014 No. 331 General rights Copyright and moral rights for the publications made accessible in the public portal are retained by the authors and/or other copyright owners and it is a condition of accessing publications that users recognise and abide by the legal requirements associated with these rights. Users may download and print one copy of any publication from the public portal for the purpose of private study or research. You may not further distribute the material or use it for any profit-making activity or commercial gain You may freely distribute the URL identifying the publication in the public portal If you believe that this document breaches copyright please contact us providing details, and we will remove access to the work immediately and investigate your claim. Ph.D. Thesis Multivariate phase type distributions Applications and parameter estimation David Meisch DTU Compute Department of Applied Mathematics and Computer Science Technical University of Denmark Kongens Lyngby PHD-2014-331 Preface This thesis has been submitted as a partial fulfilment of the requirements for the Danish Ph.D. degree at the Technical University of Denmark (DTU). The work has been carried out during the period from June 1st, 2010, to January 31th 2014, in the Section of Statistic in the Department of Applied Mathematics and Computer Science at DTU.
    [Show full text]
  • Estimation of Pareto Distribution Functions from Samples Contaminated by Measurement Errors
    UNIVERSITY of the WESTERN CAPE Estimation of Pareto Distribution Functions from Samples Contaminated by Measurement Errors Lwando Orbet Kondlo A thesis submitted in partial fulfillment of the requirements for the degree of Magister Scientiae in Statistics in the Faculty of Natural Sciences at the University of the Western Cape. Supervisor : Professor Chris Koen March 1, 2010 Keywords Deconvolution Distribution functions Error-Contaminated samples Errors-in-variables Jackknife Maximum likelihood method Measurement errors Nonparametric estimation Pareto distribution ABSTRACT MSc Statistics Thesis, Department of Statistics, University of the Western Cape. Estimation of population distributions, from samples which are contaminated by measurement errors, is a common problem. This study considers the prob- lem of estimating the population distribution of independent random variables Xj, from error-contaminated samples Yj (j = 1, . , n) such that Yj = Xj +j, where is the measurement error, which is assumed independent of X. The measurement error is also assumed to be normally distributed. Since the observed distribution function is a convolution of the error distribution with the true underlying distribution, estimation of the latter is often referred to as a deconvolution problem. A thorough study of the relevant deconvolution literature in statistics is reported. We also deal with the specific case when X is assumed to follow a truncated Pareto form. If observations are subject to Gaussian errors, then the observed Y is distributed as the convolution of the finite-support Pareto and Gaus- sian error distributions. The convolved probability density function (PDF) and cumulative distribution function (CDF) of the finite-support Pareto and Gaussian distributions are derived. The intention is to draw more specific connections between certain deconvolu- tion methods and also to demonstrate the application of the statistical theory of estimation in the presence of measurement error.
    [Show full text]
  • Bayesian Methods for Function Estimation
    hs25 v.2005/04/21 Prn:19/05/2005; 15:11 F:hs25013.tex; VTEX/VJ p. 1 aid: 25013 pii: S0169-7161(05)25013-7 docsubty: REV Handbook of Statistics, Vol. 25 ISSN: 0169-7161 13 © 2005 Elsevier B.V. All rights reserved. 1 1 DOI 10.1016/S0169-7161(05)25013-7 2 2 3 3 4 4 5 5 6 Bayesian Methods for Function Estimation 6 7 7 8 8 9 9 10 Nidhan Choudhuri, Subhashis Ghosal and Anindya Roy 10 11 11 12 12 Abstract 13 13 14 14 15 Keywords: consistency; convergence rate; Dirichlet process; density estimation; 15 16 Markov chain Monte Carlo; posterior distribution; regression function; spectral den- 16 17 sity; transition density 17 18 18 19 19 20 20 21 1. Introduction 21 22 22 23 Nonparametric and semiparametric statistical models are increasingly replacing para- 23 24 metric models, for the latter’s lack of sufficient flexibility to address a wide vari- 24 25 ety of data. A nonparametric or semiparametric model involves at least one infinite- 25 26 dimensional parameter, usually a function, and hence may also be referred to as an 26 27 infinite-dimensional model. Functions of common interest, among many others, include 27 28 the cumulative distribution function, density function, regression function, hazard rate, 28 29 transition density of a Markov process, and spectral density of a time series. While fre- 29 30 quentist methods for nonparametric estimation have been flourishing for many of these 30 31 problems, nonparametric Bayesian estimation methods had been relatively less devel- 31 32 oped.
    [Show full text]
  • Handbook on Probability Distributions
    R powered R-forge project Handbook on probability distributions R-forge distributions Core Team University Year 2009-2010 LATEXpowered Mac OS' TeXShop edited Contents Introduction 4 I Discrete distributions 6 1 Classic discrete distribution 7 2 Not so-common discrete distribution 27 II Continuous distributions 34 3 Finite support distribution 35 4 The Gaussian family 47 5 Exponential distribution and its extensions 56 6 Chi-squared's ditribution and related extensions 75 7 Student and related distributions 84 8 Pareto family 88 9 Logistic distribution and related extensions 108 10 Extrem Value Theory distributions 111 3 4 CONTENTS III Multivariate and generalized distributions 116 11 Generalization of common distributions 117 12 Multivariate distributions 133 13 Misc 135 Conclusion 137 Bibliography 137 A Mathematical tools 141 Introduction This guide is intended to provide a quite exhaustive (at least as I can) view on probability distri- butions. It is constructed in chapters of distribution family with a section for each distribution. Each section focuses on the tryptic: definition - estimation - application. Ultimate bibles for probability distributions are Wimmer & Altmann (1999) which lists 750 univariate discrete distributions and Johnson et al. (1994) which details continuous distributions. In the appendix, we recall the basics of probability distributions as well as \common" mathe- matical functions, cf. section A.2. And for all distribution, we use the following notations • X a random variable following a given distribution, • x a realization of this random variable, • f the density function (if it exists), • F the (cumulative) distribution function, • P (X = k) the mass probability function in k, • M the moment generating function (if it exists), • G the probability generating function (if it exists), • φ the characteristic function (if it exists), Finally all graphics are done the open source statistical software R and its numerous packages available on the Comprehensive R Archive Network (CRAN∗).
    [Show full text]
  • CONTINUOUS DISTRIBUTIONS Laplace Transform
    J. Virtamo 38.3143 Queueing Theory / Continuous Distributions 1 CONTINUOUS DISTRIBUTIONS Laplace transform (Laplace-Stieltjes transform) Definition The Laplace transform of a non-negative random variable X 0 with the probability density ≥ function f(x) is defined as st sX st f ∗(s)= ∞ e− f(t)dt = E[e− ] = ∞ e− dF (t) alsodenotedas (s) Z0 Z0 LX Mathematically it is the Laplace transform of the pdf function. • In dealing with continuous random variables the Laplace transform has the same role as • the generating function has in the case of discrete random variables. s – if X is a discrete integer-valued ( 0) r.v., then f ∗(s)= (e− ) ≥ G J. Virtamo 38.3143 Queueing Theory / Continuous Distributions 2 Laplace transform of a sum Let X and Y be independent random variables with L-transforms fX∗ (s) and fY∗ (s). s(X+Y ) fX∗ +Y (s) = E[e− ] sX sY = E[e− e− ] sX sY = E[e− ]E[e− ] (independence) = fX∗ (s)fY∗ (s) fX∗ +Y (s)= fX∗ (s)fY∗ (s) J. Virtamo 38.3143 Queueing Theory / Continuous Distributions 3 Calculating moments with the aid of Laplace transform By derivation one sees d sX sX f ∗0(s)= E[e− ]=E[ Xe− ] ds − Similarly, the nth derivative is n (n) d sX n sX f ∗ (s)= E[e− ] = E[( X) e− ] dsn − Evaluating these at s = 0 one gets E[X] = f ∗0(0) − 2 E[X ] =+f ∗00(0) . n n (n) E[X ] = ( 1) f ∗ (0) − J. Virtamo 38.3143 Queueing Theory / Continuous Distributions 4 Laplace transform of a random sum Consider the random sum Y = X + + X 1 ··· N where the Xi are i.i.d.
    [Show full text]