R Functions to Symbolically Compute the Central and Non-Central Moments of the Multivariate Normal Distribution

Total Page:16

File Type:pdf, Size:1020Kb

R Functions to Symbolically Compute the Central and Non-Central Moments of the Multivariate Normal Distribution JSS Journal of Statistical Software MMMMMM YYYY, Volume VV, Code Snippet II.doi: 10.18637/jss.v000.i00 R Functions to Symbolically Compute the Central and Non-central Moments of the Multivariate Normal Distribution Kem Phillips Abstract The central moments of the multivariate normal distribution are functions of its n × n variance-covariance matrix Σ. These moments can be expressed symbolically as linear combinations of products of powers of the elements of Σ. A formula for these moments derived by differentiating the characteristic function is developed. The formula requires searching integer matrices for matrices whose n successive row and column sums equal the n exponents of the moment. This formula is implemented in R, with R functions to display moments in LATEX and to evaluate moments at specified variance-covariance matrices are included. This vignette also describes how the symmoments package has been augmented to cal- culate symbolic representations of non-central multivariate moments, and to calculate the expected value of such moments as well as the expected value of multivariate polynomials. In addition, it is shown that first non-central multivariate moments are in a one-to-one relationship to phylogenetic trees. Functions making this correspondence, as well as a function for the correspondence with matchings are provided. Keywords: non-central moments, multivariate Normal distribution, symbolic computation, multivariate polynomials, phylogenetic trees, R,LATEX. 1. Introduction The central moments of an n-dimensional random vector X are defined as k1 k2 kn mk1;:::;kn = E[(X1 − µ1) (X2 − µ2) ··· (Xn − µn) ]; (1) where E[··· ] denotes expected value. Suppose that X is distributed according to the multi- variate normal distribution with mean µ and variance-covariance matrix Σ = (σij) (2) 2 Expected Values of Functions of a Multivariate Normal Random Variable where the variance terms are σii; i = 1; : : : ; n, the covariance terms are σij; i 6= j, and by symmetry σij = σji. For the multivariate normal distribution the central moments are not functions of the mean vector µ, and depend only on the variance-covariance terms σij. Simple cases are familiar. Setting µ to 0, 2 m2 = E[X1 ] = σ1;1 (3) m1;2 = E[X1X2] = σ1;2 (4) Slightly more complicated cases can be computed directly or by manipulating simple expres- sions obtained for moments of the form E[X1 ··· Xn](Wikipedia contributors 2009). For example, 2 2 2 m2;2 = E[X1 X2 ] = 2σ1;2 + σ1;1σ2;2 (5) 2 1 1 m2;1;1 = E[X1 X2 X3 ] = 2σ1;2σ1;3 + σ1;1σ2;3 (6) 3 1 m1;1 = E[X1 X2 ] = 3σ1;1σ1;2 (7) Although some higher order moments follow known patterns, most are much harder to deter- mine by simple calculations. We wish to compute the symbolic expression of any moment mk1;:::;kn in terms of the n(n+1) Pn symbols σij. Note that if i=1 ki is an odd integer, the moment is 0, so that only moments where this sum is even need be considered. The multivariate normal distribution is fundamental to mathematical statistics, and its mo- ments play a central role in statistical methodology. Various methods have been developed to numerically compute them (Muirhead 1982, p. 46) and (Anderson 1971, p. 49). Kan(2008) developed a formula (his Proposition 1) for the central moments as a repeated sum. He gives an excellent review of other formulas that have been developed, and cites Isserlis(1918) as deriving the first expression for the central moments. Muirhead (page 49) used the matrix derivatives of the multivariate normal distribution's characteristic function to derive a for- mula for multivariate cumulants. Tracy and Sultan(1993) also used matrix derivatives to derive an expression for the distribution's moments (their Theorem 2) based on a recurrence relationship of the derivatives. This article develops a new explicit formula for the moments starting with the derivatives of the characteristic function. The expression for the moments is based on a search algorithm over certain integer matrices. This formula has been translated into R functions that produce symbolic representations of moments in terms of the variance- covariance terms σij. In addition, a formula for non-central moments is developed based on central moments. This formula is programmed, and is extended to the expected values of multivariate polynomials. Finally, it is shown that first central moments are equivalent to phylogenetic trees, and R functions making this correspondence are described. The functions described here are based on Phillips(2010), and are available in the package symmoments implemented in the R system for statistical computing (R Development Core Team 2008). Both R itself and the symmoments package (as well as all other packages used in this paper) are available under the terms of the General Public License (GPL) from the Comprehensive R Archive Network (CRAN, http://CRAN.R-project.org/). Journal of Statistical Software { Code Snippets 3 2. Development of the formula for central moments The moments of any distribution can be represented by the derivatives of the distribution's characteristic function. The characteristic function of the multivariate normal distribution is (Muirhead 1982, p. 5, 49) it>X it>µ− 1 t>Σt E[e ] = e 2 (8) where t = (t1; t2; :::; tn). Within a constant, the moment is the k1; :::; kn-order derivative of the characteristic function evaluated at t = 0: Pn k Pn d i=1 i > − i=1 ki it X mk1;:::;kn = i k k k E[e ] jt=0 (9) d 1 t1d 2 t2 ··· d n tn where i is the imaginary unit. Expanding the exponential into an infinite sum, this is Pn k 1 Pn d i=1 i 1 − i=1 ki X > > ` mk1;:::;kn = i k k k (it µ − (t Σt)) =`! jt=0 (10) d 1 t1d 2 t2 ··· d n tn 2 `=0 Since we are to compute the central moment, we will set µ = 0, so that the term it>µ will not appear: Pn k 1 Pn d i=1 i 1 − i=1 ki X > ` mk1;:::;kn = i k k k (− (t Σt)) =`! jt=0 (11) d 1 t1d 2 t2 ··· d n tn 2 `=0 n Pn Pn Pn P − i=1 ki − i=1 ki=2 − i=1 ki Since i=1 ki is even, the term i = (−1) . We note that the term i will ultimately cancel with the negative in the infinite sum, and will be omitted for convenience in notation. The expression in t is > X t Σt = σijtitj (12) ij `1 `2 `n We need to find the coefficient of t1 t2 ··· tn in > ` X ` X X (t Σt) = ( σijtitj) = σijtitj ··· σijtitj (13) ij ij ij All products in the elements of the sum will occur. Any product will be obtained by choosing σijtitj a certain number of times, say `ij. Since one term is chosen from each of ` terms, P ij `ij = `. Further, for any such matrix (`ij), there will be a term, since it can be constructed by choosing σijtitj from the first `ij terms, and so forth for each (ij) until ` is exhausted. For any (`ij) there are ` (14) `11 : : : `nn ways to choose the terms, where this is the multinomial coefficient. So, X ` X ` Y `ij ( σijtitj) = (σijtitj) (15) P `11 : : : `nn ij f(`ij )j ij `ij =`g ij 4 Expected Values of Functions of a Multivariate Normal Random Variable X ` Y `ij Y `ij = σij (titj) (16) P `11 : : : `nn f(`ij )j ij `ij =`g ij ij Q `ij For the moment, we distinguish between σij and σji as symbols. As a result, each ij σij is unique as determined by unique (`ij). However, titj = tjti, so since each σij is combined with `1 `2 `n Pn two t's, the total exponent in t is 2`. That is, a term t1 t2 ··· tn must have i=1 `i = 2`. We need to determine the terms for which, for any (`1; : : : ; `n), Y `ij `1 `n (titj) = t1 ··· tn (17) ij We will get tk in the product in the following mutually exclusive cases: Condition Exponent of tk i = k; j 6= k 1 (18) i 6= k; j = k 1 i = j = k 2 So the exponent of tk will be X X `ij + `ij + 2`kk (19) i=k;j6=k i6=k;j=k This sum is obtained by adding the sum of `ij across row k to the sum across column k, since the diagonal element k occurs in both sums. That is, we get X X X `ik + `kj = (`ik + `ki) = `k (20) i j i We can now partition the set of (`ij) in Equation 16 according to these sums, that is, f`k; k = 1 : : : ng. As stated before, the sum of the exponents, `k, must be 2`. X ` ( σijtitj) = ij 0 1 n X X ` Y `ij Y `i @ σij A ti (21) P P `11 : : : `nn f(`1;:::`n)j k `k=2`g f(`11;:::;`nn)j i(`ik+`ki)=`k;k=1:::ng ij i=1 Since differentiation is distributive with respect to addition and multiplication by constants, the derivative of the product of ts can be determined from the derivatives of the individual terms: Pn k n n k d i=1 i Y Y d i t`i = t`i (22) dk1 t dk2 t ··· dkn t i ki i 1 2 n i=1 i=1 dti n Y `i! = Ifk ≤ ` g t`i−ki (23) i i (` − k )! i i=1 i i Journal of Statistical Software { Code Snippets 5 Thus, Pn ki d i=1 X ` ( σijtitj) = dk1 dk2 ··· dkn ij 0 1 X X ` Y `ij @ σij A P P `11 : : : `nn f(`1;:::;`n)j i `i=2`g f(`11;`12;:::;`nn)j i(`ih+`hi)=`h;h=1;:::;ng ij n Y `i! Ifk ≤ ` g t`i−ki (24) i i (` − k )! i i=1 i i Incorporating the constants from Equation 11, noting again that the negative signs will cancel, the full sum is 0 1 1 X 1 ` X X ` Y `ij ( ) =`! @ σij A 2 P P `11 : : : `nn `=0 f(`1;:::;`n)j i li=2`g f(`11;`12;:::;`nn)j j (`hj +`jh)=`i;h=1;:::;ng ij n Y `i! Ifk ≤ ` g t`i−ki (25) i i (` − k )! i i=1 i i Setting t = 0, only terms with `i = ki for all i will remain.
Recommended publications
  • Use of Proc Iml to Calculate L-Moments for the Univariate Distributional Shape Parameters Skewness and Kurtosis
    Statistics 573 USE OF PROC IML TO CALCULATE L-MOMENTS FOR THE UNIVARIATE DISTRIBUTIONAL SHAPE PARAMETERS SKEWNESS AND KURTOSIS Michael A. Walega Berlex Laboratories, Wayne, New Jersey Introduction Exploratory data analysis statistics, such as those Gaussian. Bickel (1988) and Van Oer Laan and generated by the sp,ge procedure PROC Verdooren (1987) discuss the concept of robustness UNIVARIATE (1990), are useful tools to characterize and how it pertains to the assumption of normality. the underlying distribution of data prior to more rigorous statistical analyses. Assessment of the As discussed by Glass et al. (1972), incorrect distributional shape of data is usually accomplished conclusions may be reached when the normality by careful examination of the values of the third and assumption is not valid, especially when one-tail tests fourth central moments, skewness and kurtosis. are employed or the sample size or significance level However, when the sample size is small or the are very small. Hopkins and Weeks (1990) also underlying distribution is non-normal, the information discuss the effects of highly non-normal data on obtained from the sample skewness and kurtosis can hypothesis testing of variances. Thus, it is apparent be misleading. that examination of the skewness (departure from symmetry) and kurtosis (deviation from a normal One alternative to the central moment shape statistics curve) is an important component of exploratory data is the use of linear combinations of order statistics (L­ analyses. moments) to examine the distributional shape characteristics of data. L-moments have several Various methods to estimate skewness and kurtosis theoretical advantages over the central moment have been proposed (MacGillivray and Salanela, shape statistics: Characterization of a wider range of 1988).
    [Show full text]
  • Probability and Statistics
    APPENDIX A Probability and Statistics The basics of probability and statistics are presented in this appendix, provid- ing the reader with a reference source for the main body of this book without interrupting the flow of argumentation there with unnecessary excursions. Only the basic definitions and results have been included in this appendix. More complicated concepts (such as stochastic processes, Ito Calculus, Girsanov Theorem, etc.) are discussed directly in the main body of the book as the need arises. A.1 PROBABILITY, EXPECTATION, AND VARIANCE A variable whose value is dependent on random events, is referred to as a random variable or a random number. An example of a random variable is the number observed when throwing a dice. The distribution of this ran- dom number is an example of a discrete distribution. A random number is discretely distributed if it can take on only discrete values such as whole numbers, for instance. A random variable has a continuous distribution if it can assume arbitrary values in some interval (which can by all means be the infinite interval from −∞ to ∞). Intuitively, the values which can be assumed by the random variable lie together densely, i.e., arbitrarily close to one another. An example of a discrete distribution on the interval from 0 to ∞ (infinity) might be a random variable which could take on the values 0.00, 0.01, 0.02, 0.03, ..., 99.98, 99.99, 100.00, 100.01, 100.02, ..., etc., like a Bund future with a tick-size of 0.01% corresponding to a value change of 10 euros per tick on a nominal of 100,000 euros.
    [Show full text]
  • Stochastic Time-Series Spectroscopy John Scoville1 San Jose State University, Dept
    Stochastic Time-Series Spectroscopy John Scoville1 San Jose State University, Dept. of Physics, San Jose, CA 95192-0106, USA Spectroscopically measuring low levels of non-equilibrium phenomena (e.g. emission in the presence of a large thermal background) can be problematic due to an unfavorable signal-to-noise ratio. An approach is presented to use time-series spectroscopy to separate non-equilibrium quantities from slowly varying equilibria. A stochastic process associated with the non-equilibrium part of the spectrum is characterized in terms of its central moments or cumulants, which may vary over time. This parameterization encodes information about the non-equilibrium behavior of the system. Stochastic time-series spectroscopy (STSS) can be implemented at very little expense in many settings since a series of scans are typically recorded in order to generate a low-noise averaged spectrum. Higher moments or cumulants may be readily calculated from this series, enabling the observation of quantities that would be difficult or impossible to determine from an average spectrum or from prinicipal components analysis (PCA). This method is more scalable than PCA, having linear time complexity, yet it can produce comparable or superior results, as shown in example applications. One example compares an STSS-derived CO2 bending mode to a standard reference spectrum and the result of PCA. A second example shows that STSS can reveal conditions of stress in rocks, a scenario where traditional methods such as PCA are inadequate. This allows spectral lines and non-equilibrium behavior to be precisely resolved. A relationship between 2nd order STSS and a time-varying form of PCA is considered.
    [Show full text]
  • Unbiased Central Moment Estimates
    Unbiased Central Moment Estimates Inna Gerlovina1;2, Alan Hubbard1 February 11, 2020 1University of California, Berkeley 2University of California, San Francisco [email protected] Contents 1 Introduction 2 2 Estimators: na¨ıve biased to unbiased 2 3 Estimates obtained from a sample 4 4 Higher order estimates (symbolic expressions for expectations) 5 1 1 Introduction Umoments package calculates unbiased central moment estimates and their two-sample analogs, also known as pooled estimates, up to sixth order (sample variance and pooled variance are commonly used second order examples of corresponding estimators). Orders four and higher include powers and products of moments - for example, fifth moment and a product of second and third moments are both of fifth order; those estimators are also included in the package. The estimates can be obtained directly from samples or from na¨ıve biased moment estimates. If the estimates of orders beyond sixth are needed, they can be generated using the set of functions also provided in the package. Those functions generate symbolic expressions for expectations of moment combinations of an arbitrary order and, in addition to moment estimation, can be used for other derivations that require such expectations (e.g. Edgeworth expansions). 2 Estimators: na¨ıve biased to unbiased For a sample X1;:::;Xn, a na¨ıve biased estimate of a k'th central moment µk is n 1 X m = (X − X¯)k; k = 2; 3;:::: (1) k n i i=1 These biased estimates together with the sample size n can be used to calculate unbiased n estimates of a given order, e.g.
    [Show full text]
  • Expectation and Functions of Random Variables
    POL 571: Expectation and Functions of Random Variables Kosuke Imai Department of Politics, Princeton University March 10, 2006 1 Expectation and Independence To gain further insights about the behavior of random variables, we first consider their expectation, which is also called mean value or expected value. The definition of expectation follows our intuition. Definition 1 Let X be a random variable and g be any function. 1. If X is discrete, then the expectation of g(X) is defined as, then X E[g(X)] = g(x)f(x), x∈X where f is the probability mass function of X and X is the support of X. 2. If X is continuous, then the expectation of g(X) is defined as, Z ∞ E[g(X)] = g(x)f(x) dx, −∞ where f is the probability density function of X. If E(X) = −∞ or E(X) = ∞ (i.e., E(|X|) = ∞), then we say the expectation E(X) does not exist. One sometimes write EX to emphasize that the expectation is taken with respect to a particular random variable X. For a continuous random variable, the expectation is sometimes written as, Z x E[g(X)] = g(x) d F (x). −∞ where F (x) is the distribution function of X. The expectation operator has inherits its properties from those of summation and integral. In particular, the following theorem shows that expectation preserves the inequality and is a linear operator. Theorem 1 (Expectation) Let X and Y be random variables with finite expectations. 1. If g(x) ≥ h(x) for all x ∈ R, then E[g(X)] ≥ E[h(X)].
    [Show full text]
  • Discrete Random Variables Randomness
    Discrete Random Variables Randomness • The word random effectively means unpredictable • In engineering practice we may treat some signals as random to simplify the analysis even though they may not actually be random Random Variable Defined X A random variable () is the assignment of numerical values to the outcomes of experiments Random Variables Examples of assignments of numbers to the outcomes of experiments. Discrete-Value vs Continuous- Value Random Variables •A discrete-value (DV) random variable has a set of distinct values separated by values that cannot occur • A random variable associated with the outcomes of coin flips, card draws, dice tosses, etc... would be DV random variable •A continuous-value (CV) random variable may take on any value in a continuum of values which may be finite or infinite in size Probability Mass Functions The probability mass function (pmf ) for a discrete random variable X is P x = P X = x . X () Probability Mass Functions A DV random variable X is a Bernoulli random variable if it takes on only two values 0 and 1 and its pmf is 1 p , x = 0 P x p , x 1 X ()= = 0 , otherwise and 0 < p < 1. Probability Mass Functions Example of a Bernoulli pmf Probability Mass Functions If we perform n trials of an experiment whose outcome is Bernoulli distributed and if X represents the total number of 1’s that occur in those n trials, then X is said to be a Binomial random variable and its pmf is n x nx p 1 p , x 0,1,2,,n P x () {} X ()= x 0 , otherwise Probability Mass Functions Binomial pmf Probability Mass Functions If we perform Bernoulli trials until a 1 (success) occurs and the probability of a 1 on any single trial is p, the probability that the k1 first success will occur on the kth trial is p()1 p .
    [Show full text]
  • A Semi-Parametric Approach to the Detection of Non-Gaussian Gravitational Wave Stochastic Backgrounds
    A Semi-Parametric Approach to the Detection of Non-Gaussian Gravitational Wave Stochastic Backgrounds Lionel Martellini1, 2, ∗ and Tania Regimbau2 1EDHEC-Risk Institute, 400 Promenade des Anglais, BP 3116, 06202 Nice Cedex 3, France 2UMR ARTEMIS, CNRS, University of Nice Sophia-Antipolis, Observatoire de la C^oted'Azur, CS 34229 F-06304 NICE, France (Dated: November 7, 2018) Abstract Using a semi-parametric approach based on the fourth-order Edgeworth expansion for the un- known signal distribution, we derive an explicit expression for the likelihood detection statistic in the presence of non-normally distributed gravitational wave stochastic backgrounds. Numerical likelihood maximization exercises based on Monte-Carlo simulations for a set of large tail sym- metric non-Gaussian distributions suggest that the fourth cumulant of the signal distribution can be estimated with reasonable precision when the ratio between the signal and the noise variances is larger than 0.01. The estimation of higher-order cumulants of the observed gravitational wave signal distribution is expected to provide additional constraints on astrophysical and cosmological models. arXiv:1405.5775v1 [astro-ph.CO] 22 May 2014 ∗Electronic address: [email protected] 1 I. INTRODUCTION According to various cosmological scenarios, we are bathed in a stochastic background of gravitational waves. Proposed theoretical models include the amplification of vacuum fluctuations during inflation[1–3], pre Big Bang models [4{6], cosmic (super)strings [7{10] or phase transitions [11{13]. In addition to the cosmological background (CGB) [14, 15], an astrophysical contribution (AGB) [16] is expected to result from the superposition of a large number of unresolved sources, such as core collapses to neutron stars or black holes [17{20], rotating neutron stars [21, 22] including magnetars [23{26], phase transition [27] or initial instabilities in young neutron stars [28, 29, 29, 30] or compact binary mergers [31{35].
    [Show full text]
  • Moment Generating Function
    Moment Generating Function Statistics 110 Summer 2006 Copyright °c 2006 by Mark E. Irwin Moments Revisited So far I've really only talked about the ¯rst two moments. Lets de¯ne what is meant by moments more precisely. De¯nition. The rth moment of a random variable X is E[Xr], assuming that the expectation exists. So the mean of a distribution is its ¯rst moment. De¯nition. The r central moment of a random variable X is E[(X ¡ E[X])r], assuming that the expectation exists. Thus the variance is the 2nd central moment of distribution. The 1st central moment usually isn't discussed as its always 0. The 3rd central moment is known as the skewness of a distribution and is used as a measure of asymmetry. Moments Revisited 1 If a distribution is symmetric about its mean (f(¹ ¡ x) = f(¹ + x)), the skewness will be 0. Similarly if the skewness is non-zero, the distribution is asymmetric. However it is possible to have asymmetric distribution with skewness = 0. Examples of symmetric distribution are normals, Beta(a; a), Bin(n; p = 0:5). Example of asymmetric distributions are Distribution Skewness Bin(n; p) np(1 ¡ p)(1 ¡ 2p) P ois(¸) ¸ 2 Exp(¸) ¸ Beta(a; b) Ugly formula The 4th central moment is known as the kurtosis. It can be used as a measure of how heavy the tails are for a distribution. The kurtosis for a normal is 3σ4. Moments Revisited 2 Note that these measures are often standardized as in their raw form they depend on the standard deviation.
    [Show full text]
  • Moments and Generating Functions
    Moments and Generating Functions September 24 and 29, 2009 Some choices of g yield a specific name for the value of Eg(X). 1 Moments, Factorial Moments, and Central Moments • For g(x) = x, we call EX the mean of X and often write µX or simply µ if only the random variable X is under consideration. { S, the number of successes in n Bernoulli trials with success parameter p, has mean np. { The mean of a geometric random variable with parameter p is 1=p. { The mean of a exponential random variable with parameter β is β. { A standard normal random variable has mean 0. • For g(x) = xm, EXm is called the m-th moment of X. { If X is a Bernoulli random variable, then X = Xm. Thus, EXm = EX = p. R 1 m { For a uniform random variable on [0; 1], the m-th moment is 0 x dx = 1=(m + 1). { The third moment for Z, a standard normal random, is 0. The fourth moment, 1 Z 1 z2 1 z2 1 Z 1 z2 4 4 3 2 EZ = p z exp − dz = −p z exp − + 3z exp − dz 2π −∞ 2 2π 2 −∞ −∞ 2 = 3EZ2 = 3 3 z2 u(z) = z v(t) = − exp − 2 0 2 0 z2 u (z) = 3z v (t) = z exp − 2 { For T , an exponential random variable, we integrate by parts Z 1 1 1 Z 1 m m m m−1 ET = t exp −(t/β) dt = t exp −(t/β) + mt exp −(t/β) dt 0 β 0 0 Z 1 1 = βm tm−1 exp −(t/β) dt = mβET m−1 0 β u(t) = tm v(t) = exp −(t/β) 0 m−1 0 1 u (t) = mt v (t) = β exp −(t/β) Thus, by induction, we have that ET m = βmm!: 1 • If g(x) = (x)k, where (x)k = x(x−1) ··· (x−k +1), then E(X)k is called the k-th factorial moment.
    [Show full text]
  • Moments Skew and Kurtosis
    APPENDIX 3.3: CALCULATING SKEW AND KURTOSIS 1 We mentioned in Chapter 1 that parameters are simply yy= numbers that characterize the scores of populations. 2 yy= * y The mean (µ) and variance (σ2) are two of the most 3 important parameters associated with statistical yy= **yy 4 analysis in psychology. In subsequent chapters, we will yy= ***yyy. use statistics to estimate these important parameters. Although µ and σ2 will be our primary concern, skew So, the exponents simply tell us how many times to and kurtosis are parameters in the same way as µ multiply a number by itself. and σ2. Skew and kurtosis can be computed from the scores in populations in a manner very similar to the Skew and Kurtosis computation of the mean and variance. Therefore, we will take a moment to comment on how skew and The second, third, and fourth central moments are related kurtosis are computed in populations and samples. to skew and kurtosis. In a population, skew is defined as Moments skew θ3 = 3 (3.A3.1) σ At the beginning of Chapter 3, we noted that the mean and variance had very similar definitions. The mean where σ is the population standard deviation (or the square of a population is the sum of all scores divided by N. root of the second central moment, θ2). Skew, as defined Statisticians sometimes call this the first raw moment in equation 3.A3.1, can take on positive and negative of the distribution. The variance in a population is the values. Symmetrical distributions (such as in Figures 3.3a sum of squared deviations from the mean, divided by and 3.4a) have zero skew.
    [Show full text]
  • To Study of the Appropriate Probability Distributions for Annual Maximum Series Using L-Moment Method in Arid and Semi-Arid Regions of Iran
    To Study of the Appropriate Probability Distributions for Annual Maximum Series Using L-Moment Method in Arid and Semi-arid Regions of Iran A. Salajegheh, A.R. Keshtkar & S.Dalfardi Faculty of Natural Resources, University of Tehran, Iran, P.O.Box:31585-3314 [email protected] ABSTRACT In Probability distributions within hydrology different methods are used for their application. The most current of they have been central moment and with the using of computers, maximum likelihood method is used, too. This research was carried out in order to recognition of suitable probability with pervious common methods. In order to investigation of suitable probability distribution for annul maximum series (AMS), by using of L-moment method through hydrometric stations which exiting in region, were be selected 20 hydrometric stations for maximum discharges. According to results of this research for annual maximum discharge, P3 distributions and L-moment method, LP3 distribution and ordinary moment method, LN2 distribution and ordinary moment method, LP3 distribution and L-moment method, P3 distribution and ordinary moment method have been suitable distinguished for %40, %30, %20, %10 and %5 of stations, respectively. For three distributions (P3, LN2, and LN3) and ordinary moment method with %23 of stations with P3 and ordinary moment were selected. Keywords: Linear moment, Maximum likelihood, Ordinary moment, Discharge, Frequency distribution function. INTRODUCTION Estimation of AMS is often required for watersheds with insufficient or nonexistent hydrometric information particularly in arid and semi-arid regions. Because parametric methods require a number of assumptions, nonparametric methods have been investigated as alternative methods. L-moment diagrams and associated goodness-of-fit procedures (Wallis, 1988; Hosking, 1990; Chowdhury et al., 1991; Pearson et al., 1991; Pearson, 1992: Vogel et al., 1993) have been advocated for evaluating the suitability of selecting various distributional alternatives for modeling flows in a region.
    [Show full text]
  • 2. Variance and Higher Moments
    Virtual Laboratories > 3. Expected Value > 1 2 3 4 5 6 2. Variance and Higher Moments Recall that by taking the expected value of various transformations of a random variable, we can measure many interesting characteristics of the distribution of the variable. In this section, we will study expected values that measure spread, skewness and other properties. Variance Definitions As usual, we start with a random experiment with probability measure ℙ on an underlying sample space. Suppose that X is a random variable for the experiment, taking values in S ⊆ ℝ. Recall that the expected value or mean of X gives the center of the distribution of X. The variance of X is a measure of the sprea d of the distribution about the mean and is defined by 2 var(X)=피 ((X − 피(X)) ) Recall that the second moment of X about a is 피((X−a) 2). Thus, the variance is the second moment of X about μ = 피(X), or equivalently, the second central moment of X. Second moments have a nice interpretation in physics, if we think of the distribution of X as a mass distribution in ℝ. Then the second moment of X about a is the moment of inertia of the mass distribution about a. This is a measure of the resistance of the mass distribution to any change in its rotational motion about a. In particular, the variance of X is the moment of inertia of the mass distribution about the center of mass μ. 1. Suppose that Xhas a discrete distribution with probability density function f .
    [Show full text]