Random Variables STATISTICS – Lecture No

Total Page:16

File Type:pdf, Size:1020Kb

Random Variables STATISTICS – Lecture No Random Variable Random Variables STATISTICS – Lecture no. 3 Jiˇr´ıNeubauer Department of Econometrics FEM UO Brno office 69a, tel. 973 442029 email:[email protected] 13. 10. 2009 Jiˇr´ıNeubauer Random Variables Discrete Random Variable Continuous Random Variable Random Variable Measures of Location Measures of Dispersion Measures of Concentration Random Variable Many random experiments have numerical outcomes. Definition A random variable is a real-valued function X (ω) defined on the sample space Ω. The set of possible values of the random variable X is called the range of X . M = {x; X (ω) = x}. Jiˇr´ıNeubauer Random Variables the number of dots when a die is rolled, the range is M = {1, 2,... 6} the number of rolls of a die until the first 6 appears, the range is M = {1, 2, }˙ the lifetime of the lightbulb, the range is M = {x; x ≥ 0}, Examples of random variables: Discrete Random Variable Continuous Random Variable Random Variable Measures of Location Measures of Dispersion Measures of Concentration Random Variable We denote random variables by capital letters X , Y ,... (eventually X1, X2,... ) and their particular values by small letters x, y,... Using random variables we can describe random events, for example X = x, X ≤ x, x1 < X < x2 etc. Jiˇr´ıNeubauer Random Variables Discrete Random Variable Continuous Random Variable Random Variable Measures of Location Measures of Dispersion Measures of Concentration Random Variable We denote random variables by capital letters X , Y ,... (eventually X1, X2,... ) and their particular values by small letters x, y,... Using random variables we can describe random events, for example X = x, X ≤ x, x1 < X < x2 etc. Examples of random variables: the number of dots when a die is rolled, the range is M = {1, 2,... 6} the number of rolls of a die until the first 6 appears, the range is M = {1, 2, }˙ the lifetime of the lightbulb, the range is M = {x; x ≥ 0}, Jiˇr´ıNeubauer Random Variables continuous ... M is a closed or open interval. Discrete Random Variable Continuous Random Variable Random Variable Measures of Location Measures of Dispersion Measures of Concentration Random Variable According to the range M we separate random variables to discrete ... M is finite or countable, Jiˇr´ıNeubauer Random Variables Discrete Random Variable Continuous Random Variable Random Variable Measures of Location Measures of Dispersion Measures of Concentration Random Variable According to the range M we separate random variables to discrete ... M is finite or countable, continuous ... M is a closed or open interval. Jiˇr´ıNeubauer Random Variables Examples of continuous random variables: the height of a person, M = (0, ∞) the time taken to complete an examination, M = (0, ∞) the amount of milk in a bottle, M = (0, ∞) Discrete Random Variable Continuous Random Variable Random Variable Measures of Location Measures of Dispersion Measures of Concentration Random Variable Examples of discrete random variables: the number of cars sold at a dealership during a given month, M = {0, 1, 2,... } the number of houses in a certain block, M = {1, 2,... } the number of fish caught on a fishing trip, M = {0, 1, 2,... } the number of heads obtained in three tosses of a coin, M = {0, 1, 2, 3} Jiˇr´ıNeubauer Random Variables Discrete Random Variable Continuous Random Variable Random Variable Measures of Location Measures of Dispersion Measures of Concentration Random Variable Examples of discrete random variables: the number of cars sold at a dealership during a given month, M = {0, 1, 2,... } the number of houses in a certain block, M = {1, 2,... } the number of fish caught on a fishing trip, M = {0, 1, 2,... } the number of heads obtained in three tosses of a coin, M = {0, 1, 2, 3} Examples of continuous random variables: the height of a person, M = (0, ∞) the time taken to complete an examination, M = (0, ∞) the amount of milk in a bottle, M = (0, ∞) Jiˇr´ıNeubauer Random Variables and some measures: measures of location, measures of dispersion, measures of concentration. Discrete Random Variable Continuous Random Variable Random Variable Measures of Location Measures of Dispersion Measures of Concentration Random Variable For the description of random variables we will use some functions: a (cumulative) distribution function F (x), a probability function p(x) – only for discrete random variables, a probability density function f (x) – only for continuous random variables. Jiˇr´ıNeubauer Random Variables Discrete Random Variable Continuous Random Variable Random Variable Measures of Location Measures of Dispersion Measures of Concentration Random Variable For the description of random variables we will use some functions: a (cumulative) distribution function F (x), a probability function p(x) – only for discrete random variables, a probability density function f (x) – only for continuous random variables. and some measures: measures of location, measures of dispersion, measures of concentration. Jiˇr´ıNeubauer Random Variables Discrete Random Variable Continuous Random Variable Random Variable Measures of Location Measures of Dispersion Measures of Concentration Distribution Function Definition Let X be any random variable. The distribution function F (x) of the random variable X is defined as F (x) = P(X ≤ x), x ∈ R. Note: distribution function = cumulative distribution function Jiˇr´ıNeubauer Random Variables F (x) is a non-decreasing, right-continuous function, it has limits lim F (x) = 0, lim F (x) = 1, x→−∞ x→∞ if range of X is M = {x; x ∈ (a, bi} then F (a) = 0 a F (b) = 1, for every real numbers x1 and x2: P(x1 < X ≤ x2) = F (x2) − F (x1). Discrete Random Variable Continuous Random Variable Random Variable Measures of Location Measures of Dispersion Measures of Concentration Distribution Function We mention some important properties of F (x): for every real x: 0 ≤ F (x) ≤ 1, Jiˇr´ıNeubauer Random Variables it has limits lim F (x) = 0, lim F (x) = 1, x→−∞ x→∞ if range of X is M = {x; x ∈ (a, bi} then F (a) = 0 a F (b) = 1, for every real numbers x1 and x2: P(x1 < X ≤ x2) = F (x2) − F (x1). Discrete Random Variable Continuous Random Variable Random Variable Measures of Location Measures of Dispersion Measures of Concentration Distribution Function We mention some important properties of F (x): for every real x: 0 ≤ F (x) ≤ 1, F (x) is a non-decreasing, right-continuous function, Jiˇr´ıNeubauer Random Variables for every real numbers x1 and x2: P(x1 < X ≤ x2) = F (x2) − F (x1). Discrete Random Variable Continuous Random Variable Random Variable Measures of Location Measures of Dispersion Measures of Concentration Distribution Function We mention some important properties of F (x): for every real x: 0 ≤ F (x) ≤ 1, F (x) is a non-decreasing, right-continuous function, it has limits lim F (x) = 0, lim F (x) = 1, x→−∞ x→∞ if range of X is M = {x; x ∈ (a, bi} then F (a) = 0 a F (b) = 1, Jiˇr´ıNeubauer Random Variables Discrete Random Variable Continuous Random Variable Random Variable Measures of Location Measures of Dispersion Measures of Concentration Distribution Function We mention some important properties of F (x): for every real x: 0 ≤ F (x) ≤ 1, F (x) is a non-decreasing, right-continuous function, it has limits lim F (x) = 0, lim F (x) = 1, x→−∞ x→∞ if range of X is M = {x; x ∈ (a, bi} then F (a) = 0 a F (b) = 1, for every real numbers x1 and x2: P(x1 < X ≤ x2) = F (x2) − F (x1). Jiˇr´ıNeubauer Random Variables Discrete Random Variable Continuous Random Variable Random Variable Measures of Location Measures of Dispersion Measures of Concentration Discrete Random Variable For a discrete random variable X , we are interested in computing probabilities of the type P(X = xk ) for various values xk in range of X . Definition Let X be a discrete random variable with range {x1, x2,... } (finite or countably infinite). The function p(x) = P(X = x) is called the probability function of X . Note: probability function = probability mass function Jiˇr´ıNeubauer Random Variables X p(x) = 1 x∈M for every two real numbers xk and xl (xk ≤ xl ): x Xl P(xk ≤ X ≤ xl ) = p(xi ). xi =xk Discrete Random Variable Continuous Random Variable Random Variable Measures of Location Measures of Dispersion Measures of Concentration Discrete Random Variable We mention some important properties of p(x) for every real number x, 0 ≤ p(x) ≤ 1, Jiˇr´ıNeubauer Random Variables for every two real numbers xk and xl (xk ≤ xl ): x Xl P(xk ≤ X ≤ xl ) = p(xi ). xi =xk Discrete Random Variable Continuous Random Variable Random Variable Measures of Location Measures of Dispersion Measures of Concentration Discrete Random Variable We mention some important properties of p(x) for every real number x, 0 ≤ p(x) ≤ 1, X p(x) = 1 x∈M Jiˇr´ıNeubauer Random Variables Discrete Random Variable Continuous Random Variable Random Variable Measures of Location Measures of Dispersion Measures of Concentration Discrete Random Variable We mention some important properties of p(x) for every real number x, 0 ≤ p(x) ≤ 1, X p(x) = 1 x∈M for every two real numbers xk and xl (xk ≤ xl ): x Xl P(xk ≤ X ≤ xl ) = p(xi ). xi =xk Jiˇr´ıNeubauer Random Variables Discrete Random Variable Continuous Random Variable Random Variable Measures of Location Measures of Dispersion Measures of Concentration Probability Function The probability function p(x) can be described by the table, P X x1 x2 ... xi ... p(x) p(x1) p(x2) ... p(xi ) ... 1 Jiˇr´ıNeubauer Random Variables Discrete Random Variable Continuous Random Variable Random Variable Measures of Location Measures of Dispersion Measures of Concentration Probability Function the graph [x, p(x)], Jiˇr´ıNeubauer Random Variables Discrete Random Variable Continuous Random Variable Random Variable Measures of Location Measures of Dispersion Measures of Concentration Probability Function the formula, for example π(1 − π)x x = 0, 1, 2,..., p(x) = 0 otherwise, where π is a given probability.
Recommended publications
  • Use of Proc Iml to Calculate L-Moments for the Univariate Distributional Shape Parameters Skewness and Kurtosis
    Statistics 573 USE OF PROC IML TO CALCULATE L-MOMENTS FOR THE UNIVARIATE DISTRIBUTIONAL SHAPE PARAMETERS SKEWNESS AND KURTOSIS Michael A. Walega Berlex Laboratories, Wayne, New Jersey Introduction Exploratory data analysis statistics, such as those Gaussian. Bickel (1988) and Van Oer Laan and generated by the sp,ge procedure PROC Verdooren (1987) discuss the concept of robustness UNIVARIATE (1990), are useful tools to characterize and how it pertains to the assumption of normality. the underlying distribution of data prior to more rigorous statistical analyses. Assessment of the As discussed by Glass et al. (1972), incorrect distributional shape of data is usually accomplished conclusions may be reached when the normality by careful examination of the values of the third and assumption is not valid, especially when one-tail tests fourth central moments, skewness and kurtosis. are employed or the sample size or significance level However, when the sample size is small or the are very small. Hopkins and Weeks (1990) also underlying distribution is non-normal, the information discuss the effects of highly non-normal data on obtained from the sample skewness and kurtosis can hypothesis testing of variances. Thus, it is apparent be misleading. that examination of the skewness (departure from symmetry) and kurtosis (deviation from a normal One alternative to the central moment shape statistics curve) is an important component of exploratory data is the use of linear combinations of order statistics (L­ analyses. moments) to examine the distributional shape characteristics of data. L-moments have several Various methods to estimate skewness and kurtosis theoretical advantages over the central moment have been proposed (MacGillivray and Salanela, shape statistics: Characterization of a wider range of 1988).
    [Show full text]
  • A Skew Extension of the T-Distribution, with Applications
    J. R. Statist. Soc. B (2003) 65, Part 1, pp. 159–174 A skew extension of the t-distribution, with applications M. C. Jones The Open University, Milton Keynes, UK and M. J. Faddy University of Birmingham, UK [Received March 2000. Final revision July 2002] Summary. A tractable skew t-distribution on the real line is proposed.This includes as a special case the symmetric t-distribution, and otherwise provides skew extensions thereof.The distribu- tion is potentially useful both for modelling data and in robustness studies. Properties of the new distribution are presented. Likelihood inference for the parameters of this skew t-distribution is developed. Application is made to two data modelling examples. Keywords: Beta distribution; Likelihood inference; Robustness; Skewness; Student’s t-distribution 1. Introduction Student’s t-distribution occurs frequently in statistics. Its usual derivation and use is as the sam- pling distribution of certain test statistics under normality, but increasingly the t-distribution is being used in both frequentist and Bayesian statistics as a heavy-tailed alternative to the nor- mal distribution when robustness to possible outliers is a concern. See Lange et al. (1989) and Gelman et al. (1995) and references therein. It will often be useful to consider a further alternative to the normal or t-distribution which is both heavy tailed and skew. To this end, we propose a family of distributions which includes the symmetric t-distributions as special cases, and also includes extensions of the t-distribution, still taking values on the whole real line, with non-zero skewness. Let a>0 and b>0be parameters.
    [Show full text]
  • 5. the Student T Distribution
    Virtual Laboratories > 4. Special Distributions > 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 5. The Student t Distribution In this section we will study a distribution that has special importance in statistics. In particular, this distribution will arise in the study of a standardized version of the sample mean when the underlying distribution is normal. The Probability Density Function Suppose that Z has the standard normal distribution, V has the chi-squared distribution with n degrees of freedom, and that Z and V are independent. Let Z T= √V/n In the following exercise, you will show that T has probability density function given by −(n +1) /2 Γ((n + 1) / 2) t2 f(t)= 1 + , t∈ℝ ( n ) √n π Γ(n / 2) 1. Show that T has the given probability density function by using the following steps. n a. Show first that the conditional distribution of T given V=v is normal with mean 0 a nd variance v . b. Use (a) to find the joint probability density function of (T,V). c. Integrate the joint probability density function in (b) with respect to v to find the probability density function of T. The distribution of T is known as the Student t distribution with n degree of freedom. The distribution is well defined for any n > 0, but in practice, only positive integer values of n are of interest. This distribution was first studied by William Gosset, who published under the pseudonym Student. In addition to supplying the proof, Exercise 1 provides a good way of thinking of the t distribution: the t distribution arises when the variance of a mean 0 normal distribution is randomized in a certain way.
    [Show full text]
  • 1 One Parameter Exponential Families
    1 One parameter exponential families The world of exponential families bridges the gap between the Gaussian family and general dis- tributions. Many properties of Gaussians carry through to exponential families in a fairly precise sense. • In the Gaussian world, there exact small sample distributional results (i.e. t, F , χ2). • In the exponential family world, there are approximate distributional results (i.e. deviance tests). • In the general setting, we can only appeal to asymptotics. A one-parameter exponential family, F is a one-parameter family of distributions of the form Pη(dx) = exp (η · t(x) − Λ(η)) P0(dx) for some probability measure P0. The parameter η is called the natural or canonical parameter and the function Λ is called the cumulant generating function, and is simply the normalization needed to make dPη fη(x) = (x) = exp (η · t(x) − Λ(η)) dP0 a proper probability density. The random variable t(X) is the sufficient statistic of the exponential family. Note that P0 does not have to be a distribution on R, but these are of course the simplest examples. 1.0.1 A first example: Gaussian with linear sufficient statistic Consider the standard normal distribution Z e−z2=2 P0(A) = p dz A 2π and let t(x) = x. Then, the exponential family is eη·x−x2=2 Pη(dx) / p 2π and we see that Λ(η) = η2=2: eta= np.linspace(-2,2,101) CGF= eta**2/2. plt.plot(eta, CGF) A= plt.gca() A.set_xlabel(r'$\eta$', size=20) A.set_ylabel(r'$\Lambda(\eta)$', size=20) f= plt.gcf() 1 Thus, the exponential family in this setting is the collection F = fN(η; 1) : η 2 Rg : d 1.0.2 Normal with quadratic sufficient statistic on R d As a second example, take P0 = N(0;Id×d), i.e.
    [Show full text]
  • Random Variables and Probability Distributions 1.1
    RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS 1. DISCRETE RANDOM VARIABLES 1.1. Definition of a Discrete Random Variable. A random variable X is said to be discrete if it can assume only a finite or countable infinite number of distinct values. A discrete random variable can be defined on both a countable or uncountable sample space. 1.2. Probability for a discrete random variable. The probability that X takes on the value x, P(X=x), is defined as the sum of the probabilities of all sample points in Ω that are assigned the value x. We may denote P(X=x) by p(x) or pX (x). The expression pX (x) is a function that assigns probabilities to each possible value x; thus it is often called the probability function for the random variable X. 1.3. Probability distribution for a discrete random variable. The probability distribution for a discrete random variable X can be represented by a formula, a table, or a graph, which provides pX (x) = P(X=x) for all x. The probability distribution for a discrete random variable assigns nonzero probabilities to only a countable number of distinct x values. Any value x not explicitly assigned a positive probability is understood to be such that P(X=x) = 0. The function pX (x)= P(X=x) for each x within the range of X is called the probability distribution of X. It is often called the probability mass function for the discrete random variable X. 1.4. Properties of the probability distribution for a discrete random variable.
    [Show full text]
  • On the Meaning and Use of Kurtosis
    Psychological Methods Copyright 1997 by the American Psychological Association, Inc. 1997, Vol. 2, No. 3,292-307 1082-989X/97/$3.00 On the Meaning and Use of Kurtosis Lawrence T. DeCarlo Fordham University For symmetric unimodal distributions, positive kurtosis indicates heavy tails and peakedness relative to the normal distribution, whereas negative kurtosis indicates light tails and flatness. Many textbooks, however, describe or illustrate kurtosis incompletely or incorrectly. In this article, kurtosis is illustrated with well-known distributions, and aspects of its interpretation and misinterpretation are discussed. The role of kurtosis in testing univariate and multivariate normality; as a measure of departures from normality; in issues of robustness, outliers, and bimodality; in generalized tests and estimators, as well as limitations of and alternatives to the kurtosis measure [32, are discussed. It is typically noted in introductory statistics standard deviation. The normal distribution has a kur- courses that distributions can be characterized in tosis of 3, and 132 - 3 is often used so that the refer- terms of central tendency, variability, and shape. With ence normal distribution has a kurtosis of zero (132 - respect to shape, virtually every textbook defines and 3 is sometimes denoted as Y2)- A sample counterpart illustrates skewness. On the other hand, another as- to 132 can be obtained by replacing the population pect of shape, which is kurtosis, is either not discussed moments with the sample moments, which gives or, worse yet, is often described or illustrated incor- rectly. Kurtosis is also frequently not reported in re- ~(X i -- S)4/n search articles, in spite of the fact that virtually every b2 (•(X i - ~')2/n)2' statistical package provides a measure of kurtosis.
    [Show full text]
  • Univariate and Multivariate Skewness and Kurtosis 1
    Running head: UNIVARIATE AND MULTIVARIATE SKEWNESS AND KURTOSIS 1 Univariate and Multivariate Skewness and Kurtosis for Measuring Nonnormality: Prevalence, Influence and Estimation Meghan K. Cain, Zhiyong Zhang, and Ke-Hai Yuan University of Notre Dame Author Note This research is supported by a grant from the U.S. Department of Education (R305D140037). However, the contents of the paper do not necessarily represent the policy of the Department of Education, and you should not assume endorsement by the Federal Government. Correspondence concerning this article can be addressed to Meghan Cain ([email protected]), Ke-Hai Yuan ([email protected]), or Zhiyong Zhang ([email protected]), Department of Psychology, University of Notre Dame, 118 Haggar Hall, Notre Dame, IN 46556. UNIVARIATE AND MULTIVARIATE SKEWNESS AND KURTOSIS 2 Abstract Nonnormality of univariate data has been extensively examined previously (Blanca et al., 2013; Micceri, 1989). However, less is known of the potential nonnormality of multivariate data although multivariate analysis is commonly used in psychological and educational research. Using univariate and multivariate skewness and kurtosis as measures of nonnormality, this study examined 1,567 univariate distriubtions and 254 multivariate distributions collected from authors of articles published in Psychological Science and the American Education Research Journal. We found that 74% of univariate distributions and 68% multivariate distributions deviated from normal distributions. In a simulation study using typical values of skewness and kurtosis that we collected, we found that the resulting type I error rates were 17% in a t-test and 30% in a factor analysis under some conditions. Hence, we argue that it is time to routinely report skewness and kurtosis along with other summary statistics such as means and variances.
    [Show full text]
  • A Family of Skew-Normal Distributions for Modeling Proportions and Rates with Zeros/Ones Excess
    S S symmetry Article A Family of Skew-Normal Distributions for Modeling Proportions and Rates with Zeros/Ones Excess Guillermo Martínez-Flórez 1, Víctor Leiva 2,* , Emilio Gómez-Déniz 3 and Carolina Marchant 4 1 Departamento de Matemáticas y Estadística, Facultad de Ciencias Básicas, Universidad de Córdoba, Montería 14014, Colombia; [email protected] 2 Escuela de Ingeniería Industrial, Pontificia Universidad Católica de Valparaíso, 2362807 Valparaíso, Chile 3 Facultad de Economía, Empresa y Turismo, Universidad de Las Palmas de Gran Canaria and TIDES Institute, 35001 Canarias, Spain; [email protected] 4 Facultad de Ciencias Básicas, Universidad Católica del Maule, 3466706 Talca, Chile; [email protected] * Correspondence: [email protected] or [email protected] Received: 30 June 2020; Accepted: 19 August 2020; Published: 1 September 2020 Abstract: In this paper, we consider skew-normal distributions for constructing new a distribution which allows us to model proportions and rates with zero/one inflation as an alternative to the inflated beta distributions. The new distribution is a mixture between a Bernoulli distribution for explaining the zero/one excess and a censored skew-normal distribution for the continuous variable. The maximum likelihood method is used for parameter estimation. Observed and expected Fisher information matrices are derived to conduct likelihood-based inference in this new type skew-normal distribution. Given the flexibility of the new distributions, we are able to show, in real data scenarios, the good performance of our proposal. Keywords: beta distribution; centered skew-normal distribution; maximum-likelihood methods; Monte Carlo simulations; proportions; R software; rates; zero/one inflated data 1.
    [Show full text]
  • Random Processes
    Chapter 6 Random Processes Random Process • A random process is a time-varying function that assigns the outcome of a random experiment to each time instant: X(t). • For a fixed (sample path): a random process is a time varying function, e.g., a signal. – For fixed t: a random process is a random variable. • If one scans all possible outcomes of the underlying random experiment, we shall get an ensemble of signals. • Random Process can be continuous or discrete • Real random process also called stochastic process – Example: Noise source (Noise can often be modeled as a Gaussian random process. An Ensemble of Signals Remember: RV maps Events à Constants RP maps Events à f(t) RP: Discrete and Continuous The set of all possible sample functions {v(t, E i)} is called the ensemble and defines the random process v(t) that describes the noise source. Sample functions of a binary random process. RP Characterization • Random variables x 1 , x 2 , . , x n represent amplitudes of sample functions at t 5 t 1 , t 2 , . , t n . – A random process can, therefore, be viewed as a collection of an infinite number of random variables: RP Characterization – First Order • CDF • PDF • Mean • Mean-Square Statistics of a Random Process RP Characterization – Second Order • The first order does not provide sufficient information as to how rapidly the RP is changing as a function of timeà We use second order estimation RP Characterization – Second Order • The first order does not provide sufficient information as to how rapidly the RP is changing as a function
    [Show full text]
  • Tests Based on Skewness and Kurtosis for Multivariate Normality
    Communications for Statistical Applications and Methods DOI: http://dx.doi.org/10.5351/CSAM.2015.22.4.361 2015, Vol. 22, No. 4, 361–375 Print ISSN 2287-7843 / Online ISSN 2383-4757 Tests Based on Skewness and Kurtosis for Multivariate Normality Namhyun Kim1;a aDepartment of Science, Hongik University, Korea Abstract A measure of skewness and kurtosis is proposed to test multivariate normality. It is based on an empirical standardization using the scaled residuals of the observations. First, we consider the statistics that take the skewness or the kurtosis for each coordinate of the scaled residuals. The null distributions of the statistics converge very slowly to the asymptotic distributions; therefore, we apply a transformation of the skewness or the kurtosis to univariate normality for each coordinate. Size and power are investigated through simulation; consequently, the null distributions of the statistics from the transformed ones are quite well approximated to asymptotic distributions. A simulation study also shows that the combined statistics of skewness and kurtosis have moderate sensitivity of all alternatives under study, and they might be candidates for an omnibus test. Keywords: Goodness of fit tests, multivariate normality, skewness, kurtosis, scaled residuals, empirical standardization, power comparison 1. Introduction Classical multivariate analysis techniques require the assumption of multivariate normality; conse- quently, there are numerous test procedures to assess the assumption in the literature. For a gen- eral review, some references are Henze (2002), Henze and Zirkler (1990), Srivastava and Mudholkar (2003), and Thode (2002, Ch.9). For comparative studies in power, we refer to Farrell et al. (2007), Horswell and Looney (1992), Mecklin and Mundfrom (2005), and Romeu and Ozturk (1993).
    [Show full text]
  • 11. Parameter Estimation
    11. Parameter Estimation Chris Piech and Mehran Sahami May 2017 We have learned many different distributions for random variables and all of those distributions had parame- ters: the numbers that you provide as input when you define a random variable. So far when we were working with random variables, we either were explicitly told the values of the parameters, or, we could divine the values by understanding the process that was generating the random variables. What if we don’t know the values of the parameters and we can’t estimate them from our own expert knowl- edge? What if instead of knowing the random variables, we have a lot of examples of data generated with the same underlying distribution? In this chapter we are going to learn formal ways of estimating parameters from data. These ideas are critical for artificial intelligence. Almost all modern machine learning algorithms work like this: (1) specify a probabilistic model that has parameters. (2) Learn the value of those parameters from data. Parameters Before we dive into parameter estimation, first let’s revisit the concept of parameters. Given a model, the parameters are the numbers that yield the actual distribution. In the case of a Bernoulli random variable, the single parameter was the value p. In the case of a Uniform random variable, the parameters are the a and b values that define the min and max value. Here is a list of random variables and the corresponding parameters. From now on, we are going to use the notation q to be a vector of all the parameters: Distribution Parameters Bernoulli(p) q = p Poisson(l) q = l Uniform(a,b) q = (a;b) Normal(m;s 2) q = (m;s 2) Y = mX + b q = (m;b) In the real world often you don’t know the “true” parameters, but you get to observe data.
    [Show full text]
  • Approximating the Distribution of the Product of Two Normally Distributed Random Variables
    S S symmetry Article Approximating the Distribution of the Product of Two Normally Distributed Random Variables Antonio Seijas-Macías 1,2 , Amílcar Oliveira 2,3 , Teresa A. Oliveira 2,3 and Víctor Leiva 4,* 1 Departamento de Economía, Universidade da Coruña, 15071 A Coruña, Spain; [email protected] 2 CEAUL, Faculdade de Ciências, Universidade de Lisboa, 1649-014 Lisboa, Portugal; [email protected] (A.O.); [email protected] (T.A.O.) 3 Departamento de Ciências e Tecnologia, Universidade Aberta, 1269-001 Lisboa, Portugal 4 Escuela de Ingeniería Industrial, Pontificia Universidad Católica de Valparaíso, Valparaíso 2362807, Chile * Correspondence: [email protected] or [email protected] Received: 21 June 2020; Accepted: 18 July 2020; Published: 22 July 2020 Abstract: The distribution of the product of two normally distributed random variables has been an open problem from the early years in the XXth century. First approaches tried to determinate the mathematical and statistical properties of the distribution of such a product using different types of functions. Recently, an improvement in computational techniques has performed new approaches for calculating related integrals by using numerical integration. Another approach is to adopt any other distribution to approximate the probability density function of this product. The skew-normal distribution is a generalization of the normal distribution which considers skewness making it flexible. In this work, we approximate the distribution of the product of two normally distributed random variables using a type of skew-normal distribution. The influence of the parameters of the two normal distributions on the approximation is explored. When one of the normally distributed variables has an inverse coefficient of variation greater than one, our approximation performs better than when both normally distributed variables have inverse coefficients of variation less than one.
    [Show full text]