Distributions Reference (Pdf)

Total Page:16

File Type:pdf, Size:1020Kb

Distributions Reference (Pdf) MLEtools.nb 1 Reference: some basic distributions An excellent reference for probability distributions is: Evans, Merran, Nicholas Hastings, and Brian Peacock. 2000. Statistical Distributions, 3rd Edition. John Wiley & Sons. New York. Continuous Expected, or mean E Z = zf z „ z Variance VAR Z = E Z - E Z 2 = z - E z 2 f z „ z @ D Ÿ H L For continuous variables, the probability density function is the probability of the value z given the parameters H L @H @ DL D Ÿ H @ DL H L ü Uniform Uniform Distribution: a uniform distribution on the interval [a,b] where a < b probability density function: 1 f z = ÅÅÅÅÅÅÅÅÅÅb-a , where a § z § b b+a mean: ÅÅÅÅÅÅÅÅÅÅ2 b-a 2 variance: ÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅ12 H L H L ü r dunif z, min = a, max = b # generate 1 uniform random number over the range 1.4 to 2.4 runif H1, 1.4, 2.4 L ü mathematicaH L 1 f z_, a_, b_ := b − a ∗ random number over range 0,1 ∗ @ D Random H0.0407265 L @D ∗ random number of specified range ∗ H L MLEtools.nb 2 Random Real, 3, 6 4.87348 @ 8 <D ü Normal Normal Distribution: has two parameters: m is the location parameter, in this case the mean; s is the scale parameter, in this case the standard deviation probability density function: 2 f z = ÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅ1 exp - ÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅz-m 2 ps2 2 s2 mean: m H L variance: s2 è!!!!!!!!!!! !!! H L I M note: plotting f z will give you the familiar bell curve ü r H L probability density function dnorm z, mean, sd # plot the pdf curve dnorm x, 4, 1 ,0,7 H L random number generation, where n is the number random numbers to generation. In other words, n = 1 will create 1 random number. H H L L rnorm n, mean, sd ü mathematicaH L 1 z − m 2 f z_, m_, σ_ := Exp − 2 2 π σ2 2 σ Plot f z, 4, 1 , z, 0, 7 ; H L @ D è!!!!!!!!!!!!! A E 0.4 @ @ D 8 <D 0.3 0.2 0.1 1 2 3 4 5 6 7 MLEtools.nb 3 Show that pdf integrates to 1 Integrate f z, 1, 1 , z, −∞, ∞ 1 @ @ D 8 <D << Statistics`ContinuousDistributions` ∗ normal random number with mean of 3 and standard deviation 2 ∗ Random NormalDistribution 3, 2 H1.87451 L @ @ DD ü Lognormal Lognormal Distribution: has two parameters: m is the location parameter; s is the shape parameter probability density function: 1 ln z -log m 2 f z = ÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅ exp - ÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅ2 , where z > 0, m >0, and s >0 z 2 ps2 2 s 2 mean: m exp ÅÅÅÅÅÅÅs 2 H H L H LL median: m è!!!!!!!!!!! !!! H L I M s2 is the variance of the ln(z) variance: m2I expMs2 exp s2 - 1 note: This describes a variable in which the logarithm of the variable is a normal distribution. Whereas the normal distribu- tion is additive, the lognormal distribution is multiplicative. H L H H L L ü r dlnorm z, meanlog = m, sdlog =σ R uses a slightly different form of the lognormal model. Use the following form so that m is defined in the same way as described in the f(zH) equation above. L dlnorm z, log m , σ # generate random number rlnorm H1, meanlogH L = 1,L sdlog = 1.2 ü mathematica H L 1 Log z − Log m 2 f z_, m_, σ_ := Exp − 2 z 2 π σ2 2 σ H @ D @ DL @ D è!!!!!!!!!!!!! A E MLEtools.nb 4 Plot f z, 1.7, 1 , z, 0, 20 ; @ @ D 8 <D 0.3 0.2 0.1 5 10 15 20 Show that pdf integrates to 1 Integrate f z, 1, 1 , z, 0, ∞ 1 @ @ D 8 <D Derive expected value Integrate zf z, m, σ , z, 0, ∞ , Assumptions → σ>0, m > 0, σ∈Reals, m ∈ Reals 2 σ 2 m @ @ D 8 < 8 <D Derive variance Integrate z − Integrate zf z, m, σ , z, 0, ∞ , Assumptions → σ>0, m > 0, σ∈Reals, m ∈ Reals 2 f z, m, σ , z, 0, ∞ , Assumptions → σ>0, m > 0, σ∈Reals, m ∈ Reals @ 2 2 Hσ −1 +σ m2 @ @ D 8 < 8 <DL @ D 8 < 8 <D << Statistics`ContinuousDistributions` ∗ IlognormalM random number with the first number being the Log m and the second number being σ. In the case below, m = 3 and σ=2 ∗ Random LogNormalDistribution Log 3 ,2 H @ D 6.94065 L @ @ @ D DD MLEtools.nb 5 ü Exponential Exponential Distribution: has three parameters: m is the location parameter; s is the shape parameter probability density function: f z = a Exp -az, where z > 0 1 mean: ÅÅÅÅa 1 variance: ÅÅÅÅÅÅÅa2 note: ContinuousH L distribution@ D analog of Poisson. Can be used for dispersal distance. Widely used in Baysian Analysis ü r dexp z, rate = a # generate random number rexp H1, rate = aL ü mathematicaH L f z_, a_ := a Exp −az Plot f z, 0.5 , z, 0, 15 , PlotRange → All ; @ D @ D 0.5 @ @ D 8 < D 0.4 0.3 0.2 0.1 2 4 6 8 10 12 14 << Statistics`ContinuousDistributions` ∗ exponential random number with a = 0.5 ∗ Random ExponentialDistribution 0.5 H2.1514 L @ @ DD MLEtools.nb 6 ü Gamma Gamma Distribution: has two parameters: a is the rate parameter; n is the shape parameter probability density function: an n-1 f z = ÅÅÅÅÅÅÅÅÅÅÅG n Exp -az z , where z > 0 n mean: ÅÅÅÅa n variance: ÅÅÅÅÅÅÅ2 H L H aL @ D note: can have a long tail. When n = 1, the gamma distribution becomes the exponential distribution. When n = degrees of freedom / 2 and a = 2, the gamma distribution becomes the chi-square distribution. ü r dgamma z, shape = n , rate = a # generate random number rgamma H1, shape = n, rate = a L ü mathematica H L an f z_, a_, n_ := Exp −az zn−1 Gamma n Plot f z, 1, 1 , z, 0, 12 , PlotRange → All ; @ D @ D 1 @ D @ @ D 8 < D 0.8 0.6 0.4 0.2 2 4 6 8 10 12 MLEtools.nb 7 Plot f z, 1, 2 , z, 0, 12 , PlotRange → All ; 0.35 @ @ D 8 < D 0.3 0.25 0.2 0.15 0.1 0.05 2 4 6 8 10 12 Plot f z, 0.5, 2 , z, 0, 12 , PlotRange → All ; 0.175 @ @ D 8 < D 0.15 0.125 0.1 0.075 0.05 0.025 2 4 6 8 10 12 Derive expected value Integrate zf z, a, n , z, 0, ∞ , Assumptions → a > 0, n > 0, a ∈ Reals, n ∈ Reals n a @ @ D 8 < 8 <D Derive variance Integrate z − Integrate zf z, a, n , z, 0, ∞ , Assumptions → a > 0, n > 0, a ∈ Reals, n ∈ Reals 2 f z, a, n , z, 0, ∞ , Assumptions → a > 0, n > 0, a ∈ Reals, n ∈ Reals @ n H @ @ D 8 < 8 <DL a2 @ D 8 < 8 <D << Statistics`ContinuousDistributions` ∗ gamma random number with n = 2 and a = 0.5 ∗ Random GammaDistribution 2.0 , 1 0.5 H2.82494 L @ @ ê DD MLEtools.nb 8 ü Chi-Square Note: this section is a bit incomplete Chi-Square Distribution: A squared normal distribution is a chi-square distribution. probability density function for 1 degree of freedom: f z = ÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅ1 ÅÅÅÅÅÅÅÅ exp - ÅÅÅÅÅÅÅÅÅÅÅz 2 ps2 z 2 s2 probability densityè!!!!!!! function!!!! !!!!!! for n degrees of freedom: H L 1 Hn 2 -1 L z f z = ÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅn n 2ÅÅÅÅÅÅÅÅÅÅÅÅÅÅn z exp - ÅÅÅÅÅÅÅÅÅÅÅ2 , for z ¥ 0 s 2 G ÅÅÅÅ2 2 s mean: n s2 H ê L 4 H ê L variance: 2 n s H L H L H L notes: - n can also be thought of as a shape parameter in the context of a general distribution - often assume unit variance, s = 1 1 - G ÅÅÅÅ2 = p è!!!! ü rH L dchisq ü mathematica HL 1 z f z_, σ_ := Exp − 2 2 π σ2 z 2 σ Plot f z, 2.1 , z, 0, 7 ; @ D è!!!!!!!!!!!!!!!!! A E 1.75 1.5 @ @ D 8 <D 1.25 1 0.75 0.5 0.25 1 2 3 4 5 6 7 Show that pdf integrates to 1 MLEtools.nb 9 Integrate f z, 2.1 , z, 0, ∞ 1. @ @ D 8 <D << Statistics`ContinuousDistributions` ∗ Chi−Square distributed random number with 1 degree of freedom ∗ Random ChiSquareDistribution 1 H0.111756 L @ @ DD Add non-centrality, or location, parameter with a single degree of freedom, although no scale parameter mean: 1 + m Two forms are given below −1 z +μ 1 z 4 f z_, μ_ := Exp − BesselI −1 2, μ z 2 2 μ 1 1 1 è!!!!!!! z +μ f@z_, μ_D := Hypergeometric0F1A E J N , z μ A ê ExpE− 2 4 2 π z 2 NIntegrate f z, 2.0 , z, 0, 100000 @ D A E è!!!!!!!!!!!! A E 1. @ @ D 8 <D NIntegrate zf z, 2.0 , z, 0, 100000 3. @ @ D 8 <D Integrate zf z, μ , z, 0, ∞ , Assumptions → μ>0, μ∈Reals 1 +μ @ @ D 8 < 8 <D 1 1 Hypergeometric0F1 , z μ 2 4 Cosh z μ A E 1 1 z +μ è!!!!!!! μ + − μ − ExpA zE Exp z Exp 2 2 π z 2 è!!!!!!! è!!!!!!! 1 z +μ ExpI Az μ +EExp − A z μ EMè!!!!!!!!!!!! Exp −A E 2 2 π z 2 1è!!!!!!! è!!!!!!!z +μ 1 z +μ IA ExpE zAμ−EM +è!!!!!!!!!!!! AExp − zEμ− 2 2 π z 2 2 2 π z 2 è!!!!!!! è!!!!!!! è!!!!!!!!!!!! A E è!!!!!!!!!!!! A E MLEtools.nb 10 ü Bimodal Normal Normal Distribution: has two parameters: m is the location parameter, in this case the mean; s is the scale parameter, in this case the standard deviation probability density function: 2 f z = ÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅ1 exp - ÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅÅz-m 2 ps2 2 s2 mean: m H L variance: s2 è!!!!!!!!!!! !!! H L I M note: plotting f z will give you the familiar bell curve Here I extend that to a bimodal normal distribution.
Recommended publications
  • Effect of Probability Distribution of the Response Variable in Optimal Experimental Design with Applications in Medicine †
    mathematics Article Effect of Probability Distribution of the Response Variable in Optimal Experimental Design with Applications in Medicine † Sergio Pozuelo-Campos *,‡ , Víctor Casero-Alonso ‡ and Mariano Amo-Salas ‡ Department of Mathematics, University of Castilla-La Mancha, 13071 Ciudad Real, Spain; [email protected] (V.C.-A.); [email protected] (M.A.-S.) * Correspondence: [email protected] † This paper is an extended version of a published conference paper as a part of the proceedings of the 35th International Workshop on Statistical Modeling (IWSM), Bilbao, Spain, 19–24 July 2020. ‡ These authors contributed equally to this work. Abstract: In optimal experimental design theory it is usually assumed that the response variable follows a normal distribution with constant variance. However, some works assume other probability distributions based on additional information or practitioner’s prior experience. The main goal of this paper is to study the effect, in terms of efficiency, when misspecification in the probability distribution of the response variable occurs. The elemental information matrix, which includes information on the probability distribution of the response variable, provides a generalized Fisher information matrix. This study is performed from a practical perspective, comparing a normal distribution with the Poisson or gamma distribution. First, analytical results are obtained, including results for the linear quadratic model, and these are applied to some real illustrative examples. The nonlinear 4-parameter Hill model is next considered to study the influence of misspecification in a Citation: Pozuelo-Campos, S.; dose-response model. This analysis shows the behavior of the efficiency of the designs obtained in Casero-Alonso, V.; Amo-Salas, M.
    [Show full text]
  • The Beta-Binomial Distribution Introduction Bayesian Derivation
    In Danish: 2005-09-19 / SLB Translated: 2008-05-28 / SLB Bayesian Statistics, Simulation and Software The beta-binomial distribution I have translated this document, written for another course in Danish, almost as is. I have kept the references to Lee, the textbook used for that course. Introduction In Lee: Bayesian Statistics, the beta-binomial distribution is very shortly mentioned as the predictive distribution for the binomial distribution, given the conjugate prior distribution, the beta distribution. (In Lee, see pp. 78, 214, 156.) Here we shall treat it slightly more in depth, partly because it emerges in the WinBUGS example in Lee x 9.7, and partly because it possibly can be useful for your project work. Bayesian Derivation We make n independent Bernoulli trials (0-1 trials) with probability parameter π. It is well known that the number of successes x has the binomial distribution. Considering the probability parameter π unknown (but of course the sample-size parameter n is known), we have xjπ ∼ bin(n; π); where in generic1 notation n p(xjπ) = πx(1 − π)1−x; x = 0; 1; : : : ; n: x We assume as prior distribution for π a beta distribution, i.e. π ∼ beta(α; β); 1By this we mean that p is not a fixed function, but denotes the density function (in the discrete case also called the probability function) for the random variable the value of which is argument for p. 1 with density function 1 p(π) = πα−1(1 − π)β−1; 0 < π < 1: B(α; β) I remind you that the beta function can be expressed by the gamma function: Γ(α)Γ(β) B(α; β) = : (1) Γ(α + β) In Lee, x 3.1 is shown that the posterior distribution is a beta distribution as well, πjx ∼ beta(α + x; β + n − x): (Because of this result we say that the beta distribution is conjugate distribution to the binomial distribution.) We shall now derive the predictive distribution, that is finding p(x).
    [Show full text]
  • 9. Binomial Distribution
    J. K. SHAH CLASSES Binomial Distribution 9. Binomial Distribution THEORETICAL DISTRIBUTION (Exists only in theory) 1. Theoretical Distribution is a distribution where the variables are distributed according to some definite mathematical laws. 2. In other words, Theoretical Distributions are mathematical models; where the frequencies / probabilities are calculated by mathematical computation. 3. Theoretical Distribution are also called as Expected Variance Distribution or Frequency Distribution 4. THEORETICAL DISTRIBUTION DISCRETE CONTINUOUS BINOMIAL POISSON Distribution Distribution NORMAL OR Student’s ‘t’ Chi-square Snedecor’s GAUSSIAN Distribution Distribution F-Distribution Distribution A. Binomial Distribution (Bernoulli Distribution) 1. This distribution is a discrete probability Distribution where the variable ‘x’ can assume only discrete values i.e. x = 0, 1, 2, 3,....... n 2. This distribution is derived from a special type of random experiment known as Bernoulli Experiment or Bernoulli Trials , which has the following characteristics (i) Each trial must be associated with two mutually exclusive & exhaustive outcomes – SUCCESS and FAILURE . Usually the probability of success is denoted by ‘p’ and that of the failure by ‘q’ where q = 1-p and therefore p + q = 1. (ii) The trials must be independent under identical conditions. (iii) The number of trial must be finite (countably finite). (iv) Probability of success and failure remains unchanged throughout the process. : 442 : J. K. SHAH CLASSES Binomial Distribution Note 1 : A ‘trial’ is an attempt to produce outcomes which is neither sure nor impossible in nature. Note 2 : The conditions mentioned may also be treated as the conditions for Binomial Distributions. 3. Characteristics or Properties of Binomial Distribution (i) It is a bi parametric distribution i.e.
    [Show full text]
  • Chapter 6 Continuous Random Variables and Probability
    EF 507 QUANTITATIVE METHODS FOR ECONOMICS AND FINANCE FALL 2019 Chapter 6 Continuous Random Variables and Probability Distributions Chap 6-1 Probability Distributions Probability Distributions Ch. 5 Discrete Continuous Ch. 6 Probability Probability Distributions Distributions Binomial Uniform Hypergeometric Normal Poisson Exponential Chap 6-2/62 Continuous Probability Distributions § A continuous random variable is a variable that can assume any value in an interval § thickness of an item § time required to complete a task § temperature of a solution § height in inches § These can potentially take on any value, depending only on the ability to measure accurately. Chap 6-3/62 Cumulative Distribution Function § The cumulative distribution function, F(x), for a continuous random variable X expresses the probability that X does not exceed the value of x F(x) = P(X £ x) § Let a and b be two possible values of X, with a < b. The probability that X lies between a and b is P(a < X < b) = F(b) -F(a) Chap 6-4/62 Probability Density Function The probability density function, f(x), of random variable X has the following properties: 1. f(x) > 0 for all values of x 2. The area under the probability density function f(x) over all values of the random variable X is equal to 1.0 3. The probability that X lies between two values is the area under the density function graph between the two values 4. The cumulative density function F(x0) is the area under the probability density function f(x) from the minimum x value up to x0 x0 f(x ) = f(x)dx 0 ò xm where
    [Show full text]
  • Lecture.7 Poisson Distributions - Properties, Normal Distributions- Properties
    Lecture.7 Poisson Distributions - properties, Normal Distributions- properties Theoretical Distributions Theoretical distributions are 1. Binomial distribution Discrete distribution 2. Poisson distribution 3. Normal distribution Continuous distribution Discrete Probability distribution Bernoulli distribution A random variable x takes two values 0 and 1, with probabilities q and p ie., p(x=1) = p and p(x=0)=q, q-1-p is called a Bernoulli variate and is said to be Bernoulli distribution where p and q are probability of success and failure. It was given by Swiss mathematician James Bernoulli (1654-1705) Example • Tossing a coin(head or tail) • Germination of seed(germinate or not) Binomial distribution Binomial distribution was discovered by James Bernoulli (1654-1705). Let a random experiment be performed repeatedly and the occurrence of an event in a trial be called as success and its non-occurrence is failure. Consider a set of n independent trails (n being finite), in which the probability p of success in any trail is constant for each trial. Then q=1-p is the probability of failure in any trail. 1 The probability of x success and consequently n-x failures in n independent trails. But x successes in n trails can occur in ncx ways. Probability for each of these ways is pxqn-x. P(sss…ff…fsf…f)=p(s)p(s)….p(f)p(f)…. = p,p…q,q… = (p,p…p)(q,q…q) (x times) (n-x times) Hence the probability of x success in n trials is given by x n-x ncx p q Definition A random variable x is said to follow binomial distribution if it assumes non- negative values and its probability mass function is given by P(X=x) =p(x) = x n-x ncx p q , x=0,1,2…n q=1-p 0, otherwise The two independent constants n and p in the distribution are known as the parameters of the distribution.
    [Show full text]
  • 3.2.3 Binomial Distribution
    3.2.3 Binomial Distribution The binomial distribution is based on the idea of a Bernoulli trial. A Bernoulli trail is an experiment with two, and only two, possible outcomes. A random variable X has a Bernoulli(p) distribution if 8 > <1 with probability p X = > :0 with probability 1 − p, where 0 ≤ p ≤ 1. The value X = 1 is often termed a “success” and X = 0 is termed a “failure”. The mean and variance of a Bernoulli(p) random variable are easily seen to be EX = (1)(p) + (0)(1 − p) = p and VarX = (1 − p)2p + (0 − p)2(1 − p) = p(1 − p). In a sequence of n identical, independent Bernoulli trials, each with success probability p, define the random variables X1,...,Xn by 8 > <1 with probability p X = i > :0 with probability 1 − p. The random variable Xn Y = Xi i=1 has the binomial distribution and it the number of sucesses among n independent trials. The probability mass function of Y is µ ¶ ¡ ¢ n ¡ ¢ P Y = y = py 1 − p n−y. y For this distribution, t n EX = np, Var(X) = np(1 − p),MX (t) = [pe + (1 − p)] . 1 Theorem 3.2.2 (Binomial theorem) For any real numbers x and y and integer n ≥ 0, µ ¶ Xn n (x + y)n = xiyn−i. i i=0 If we take x = p and y = 1 − p, we get µ ¶ Xn n 1 = (p + (1 − p))n = pi(1 − p)n−i. i i=0 Example 3.2.2 (Dice probabilities) Suppose we are interested in finding the probability of obtaining at least one 6 in four rolls of a fair die.
    [Show full text]
  • Estimation of Common Location and Scale Parameters in Nonregular Cases Ahmad Razmpour Iowa State University
    Iowa State University Capstones, Theses and Retrospective Theses and Dissertations Dissertations 1982 Estimation of common location and scale parameters in nonregular cases Ahmad Razmpour Iowa State University Follow this and additional works at: https://lib.dr.iastate.edu/rtd Part of the Statistics and Probability Commons Recommended Citation Razmpour, Ahmad, "Estimation of common location and scale parameters in nonregular cases " (1982). Retrospective Theses and Dissertations. 7528. https://lib.dr.iastate.edu/rtd/7528 This Dissertation is brought to you for free and open access by the Iowa State University Capstones, Theses and Dissertations at Iowa State University Digital Repository. It has been accepted for inclusion in Retrospective Theses and Dissertations by an authorized administrator of Iowa State University Digital Repository. For more information, please contact [email protected]. INFORMATION TO USERS This reproduction was made from a copy of a document sent to us for microfilming. While the most advanced technology has been used to photograph and reproduce this document, the quality of the reproduction is heavily dependent upon the quality of the material submitted. The following explanation of techniques is provided to help clarify markings or notations which may appear on this reproduction. 1. The sign or "target" for pages apparently lacking from the document photographed is "Missing Page(s)". If it was possible to obtain the missing page(s) or section, they are spliced into the film along with adjacent pages. This may have necessitated cutting through an image and duplicating adjacent pages to assure complete continuity. 2. When an image on the film is obliterated with a round black mark, it is an indication of either blurred copy because of movement during exposure, duplicate copy, or copyrighted materials that should not have been filmed.
    [Show full text]
  • A Comparison of Unbiased and Plottingposition Estimators of L
    WATER RESOURCES RESEARCH, VOL. 31, NO. 8, PAGES 2019-2025, AUGUST 1995 A comparison of unbiased and plotting-position estimators of L moments J. R. M. Hosking and J. R. Wallis IBM ResearchDivision, T. J. Watson ResearchCenter, Yorktown Heights, New York Abstract. Plotting-positionestimators of L momentsand L moment ratios have several disadvantagescompared with the "unbiased"estimators. For generaluse, the "unbiased'? estimatorsshould be preferred. Plotting-positionestimators may still be usefulfor estimatingextreme upper tail quantilesin regional frequencyanalysis. Probability-Weighted Moments and L Moments •r+l-" (--1)r • P*r,k Olk '- E p *r,!•[J!•. Probability-weightedmoments of a randomvariable X with k=0 k=0 cumulativedistribution function F( ) and quantile function It is convenient to define dimensionless versions of L mo- x( ) were definedby Greenwoodet al. [1979]to be the quan- tities ments;this is achievedby dividingthe higher-orderL moments by the scale measure h2. The L moment ratios •'r, r = 3, Mp,ra= E[XP{F(X)}r{1- F(X)} s] 4, '", are definedby ßr-" •r/•2 ß {X(u)}PUr(1 -- U)s du. L momentratios measure the shapeof a distributionindepen- dently of its scaleof measurement.The ratios *3 ("L skew- ness")and *4 ("L kurtosis")are nowwidely used as measures Particularlyuseful specialcases are the probability-weighted of skewnessand kurtosis,respectively [e.g., Schaefer,1990; moments Pilon and Adamowski,1992; Royston,1992; Stedingeret al., 1992; Vogeland Fennessey,1993]. 12•r= M1,0, r = •01 (1 - u)rx(u) du, Estimators Given an ordered sample of size n, Xl: n • X2:n • ''' • urx(u) du. X.... there are two establishedways of estimatingthe proba- /3r--- Ml,r, 0 =f01 bility-weightedmoments and L moments of the distribution from whichthe samplewas drawn.
    [Show full text]
  • A Study of Non-Central Skew T Distributions and Their Applications in Data Analysis and Change Point Detection
    A STUDY OF NON-CENTRAL SKEW T DISTRIBUTIONS AND THEIR APPLICATIONS IN DATA ANALYSIS AND CHANGE POINT DETECTION Abeer M. Hasan A Dissertation Submitted to the Graduate College of Bowling Green State University in partial fulfillment of the requirements for the degree of DOCTOR OF PHILOSOPHY August 2013 Committee: Arjun K. Gupta, Co-advisor Wei Ning, Advisor Mark Earley, Graduate Faculty Representative Junfeng Shang. Copyright c August 2013 Abeer M. Hasan All rights reserved iii ABSTRACT Arjun K. Gupta, Co-advisor Wei Ning, Advisor Over the past three decades there has been a growing interest in searching for distribution families that are suitable to analyze skewed data with excess kurtosis. The search started by numerous papers on the skew normal distribution. Multivariate t distributions started to catch attention shortly after the development of the multivariate skew normal distribution. Many researchers proposed alternative methods to generalize the univariate t distribution to the multivariate case. Recently, skew t distribution started to become popular in research. Skew t distributions provide more flexibility and better ability to accommodate long-tailed data than skew normal distributions. In this dissertation, a new non-central skew t distribution is studied and its theoretical properties are explored. Applications of the proposed non-central skew t distribution in data analysis and model comparisons are studied. An extension of our distribution to the multivariate case is presented and properties of the multivariate non-central skew t distri- bution are discussed. We also discuss the distribution of quadratic forms of the non-central skew t distribution. In the last chapter, the change point problem of the non-central skew t distribution is discussed under different settings.
    [Show full text]
  • 1 One Parameter Exponential Families
    1 One parameter exponential families The world of exponential families bridges the gap between the Gaussian family and general dis- tributions. Many properties of Gaussians carry through to exponential families in a fairly precise sense. • In the Gaussian world, there exact small sample distributional results (i.e. t, F , χ2). • In the exponential family world, there are approximate distributional results (i.e. deviance tests). • In the general setting, we can only appeal to asymptotics. A one-parameter exponential family, F is a one-parameter family of distributions of the form Pη(dx) = exp (η · t(x) − Λ(η)) P0(dx) for some probability measure P0. The parameter η is called the natural or canonical parameter and the function Λ is called the cumulant generating function, and is simply the normalization needed to make dPη fη(x) = (x) = exp (η · t(x) − Λ(η)) dP0 a proper probability density. The random variable t(X) is the sufficient statistic of the exponential family. Note that P0 does not have to be a distribution on R, but these are of course the simplest examples. 1.0.1 A first example: Gaussian with linear sufficient statistic Consider the standard normal distribution Z e−z2=2 P0(A) = p dz A 2π and let t(x) = x. Then, the exponential family is eη·x−x2=2 Pη(dx) / p 2π and we see that Λ(η) = η2=2: eta= np.linspace(-2,2,101) CGF= eta**2/2. plt.plot(eta, CGF) A= plt.gca() A.set_xlabel(r'$\eta$', size=20) A.set_ylabel(r'$\Lambda(\eta)$', size=20) f= plt.gcf() 1 Thus, the exponential family in this setting is the collection F = fN(η; 1) : η 2 Rg : d 1.0.2 Normal with quadratic sufficient statistic on R d As a second example, take P0 = N(0;Id×d), i.e.
    [Show full text]
  • Solutions to the Exercises
    Solutions to the Exercises Chapter 1 Solution 1.1 (a) Your computer may be programmed to allocate borderline cases to the next group down, or the next group up; and it may or may not manage to follow this rule consistently, depending on its handling of the numbers involved. Following a rule which says 'move borderline cases to the next group up', these are the five classifications. (i) 1.0-1.2 1.2-1.4 1.4-1.6 1.6-1.8 1.8-2.0 2.0-2.2 2.2-2.4 6 6 4 8 4 3 4 2.4-2.6 2.6-2.8 2.8-3.0 3.0-3.2 3.2-3.4 3.4-3.6 3.6-3.8 6 3 2 2 0 1 1 (ii) 1.0-1.3 1.3-1.6 1.6-1.9 1.9-2.2 2.2-2.5 10 6 10 5 6 2.5-2.8 2.8-3.1 3.1-3.4 3.4-3.7 7 3 1 2 (iii) 0.8-1.1 1.1-1.4 1.4-1.7 1.7-2.0 2.0-2.3 2 10 6 10 7 2.3-2.6 2.6-2.9 2.9-3.2 3.2-3.5 3.5-3.8 6 4 3 1 1 (iv) 0.85-1.15 1.15-1.45 1.45-1.75 1.75-2.05 2.05-2.35 4 9 8 9 5 2.35-2.65 2.65-2.95 2.95-3.25 3.25-3.55 3.55-3.85 7 3 3 1 1 (V) 0.9-1.2 1.2-1.5 1.5-1.8 1.8-2.1 2.1-2.4 6 7 11 7 4 2.4-2.7 2.7-3.0 3.0-3.3 3.3-3.6 3.6-3.9 7 4 2 1 1 (b) Computer graphics: the diagrams are shown in Figures 1.9 to 1.11.
    [Show full text]
  • ONE SAMPLE TESTS the Following Data Represent the Change
    1 WORKED EXAMPLES 6 INTRODUCTION TO STATISTICAL METHODS EXAMPLE 1: ONE SAMPLE TESTS The following data represent the change (in ml) in the amount of Carbon monoxide transfer (an indicator of improved lung function) in smokers with chickenpox over a one week period: 33, 2, 24, 17, 4, 1, -6 Is there evidence of significant improvement in lung function (a) if the data are normally distributed with σ = 10, (b) if the data are normally distributed with σ unknown? Use a significance level of α = 0.05. SOLUTION: (a) Here we have a sample of size 7 with sample mean x = 10.71. We want to test H0 : μ = 0.0, H1 : μ = 0.0, 6 under the assumption that the data follow a Normal distribution with σ = 10.0 known. Then, we have, in the Z-test, 10.71 0.0 z = − = 2.83, 10.0/√7 which lies in the critical region, as the critical values for this test are 1.96, for significance ± level α = 0.05. Therefore we have evidence to reject H0. The p-value is given by p = 2Φ( 2.83) = 0.004 < α. − (b) The sample variance is s2 = 14.192. In the T-test, we have test statistic t given by x 0.0 10.71 0.0 t = − = − = 2.00. s/√n 14.19/√7 The upper critical value CR is obtained by solving FSt(n 1)(CR) = 0.975, − where FSt(n 1) is the cdf of a Student-t distribution with n 1 degrees of freedom; here n = 7, so − − we can use statistical tables or a computer to find that CR = 2.447, and note that, as Student-t distributions are symmetric the lower critical value is CR.
    [Show full text]