Random Variables Generation

Total Page:16

File Type:pdf, Size:1020Kb

Random Variables Generation Random Variables Generation Revised version of the slides based on the book Discrete-Event Simulation: a first course L.L. Leemis & S.K. Park Section(s) 6.1, 6.2, 7.1, 7.2 c 2006 Pearson Ed., Inc. 0-13-142917-5 Discrete-Event Simulation Random Variables Generation 1/80 Introduction Monte Carlo Simulators differ from Trace Driven simulators because of the use of Random Number Generators to represent the variability that affects the behaviors of real systems. Uniformly distributed random variables are the most elementary representations that we can use in Monte Carlo simulation, but are not enough to capture the complexity of real systems. We must thus devise methods for generating instances (variates) of arbitrary random variables Properly using uniform random numbers, it is possible to obtain this result. In the sequel we will first recall some basic properties of Discrete and Continuous random variables and then we will discuss several methods to obtain their variates Discrete-Event Simulation Random Variables Generation 2/80 Basic Probability Concepts Empirical Probability, derives from performing an experiment many times n and counting the number of occurrences na of an event A The relative frequency of occurrence of event is na/n The frequency theory of probability asserts thatA the relative frequency converges as n → ∞ n Pr( )= lim a A n→∞ n Axiomatic Probability is a formal, set-theoretic approach Mathematically construct the sample space and calculate the number of events A The two are complementary! Discrete-Event Simulation Random Variables Generation 3/80 Example Roll two dice and observe the up faces (1, 1) (1, 2) (1, 3) (1, 4) (1, 5) (1, 6) (2, 1) (2, 2) (2, 3) (2, 4) (2, 5) (2, 6) (3, 1) (3, 2) (3, 3) (3, 4) (3, 5) (3, 6) (4, 1) (4, 2) (4, 3) (4, 4) (4, 5) (4, 6) (5, 1) (5, 2) (5, 3) (5, 4) (5, 5) (5, 6) (6, 1) (6, 2) (6, 3) (6, 4) (6, 5) (6, 6) If the two up faces are summed, an integer-valued random variable, say X, is defined with possible values 2 through 12 sum, x : 2 3 4 5 6 7 8 9101112 1 2 3 4 5 6 5 4 3 2 1 Pr(X = x) : 36 36 36 36 36 36 36 36 36 36 36 Pr(X = 7) could be estimated by replicating the experiment many times and calculating the relative frequency of occurrence of 7’s Discrete-Event Simulation Random Variables Generation 4/80 Discrete Random Variables A random variable X is discrete if and only if its set of possible values is finite or, at most, countably infinite X A discrete random variable X is uniquely determined by Its set of possible values Its probability distributionX function (pdf): A real-valued function f ( ) defined for each x as the probability that X has the· value x ∈ X f (x)= Pr(X = x) By definition, f (x)=1 x X Discrete-Event Simulation Random Variables Generation 5/80 Examples Example 1 X is Equilikely (Discrete Uniform dist.) (a, b) = b a + 1 and each possible value is equally likely |X | − 1 f (x)= x = a, a + 1,..., b b a + 1 − Example 2 Roll two fair dice If X is the sum of the two up faces, = x x = 2, 3,..., 12 X { | } From example 2.3.1, 6 7 x f (x)= − | − | x = 2, 3,..., 12 36 or alternatively x 1 − 2 x = 7 36 ≤ ≤ f (x)= 13 x − 8 x = 12 36 ≤ ≤ Discrete-Event Simulation Random Variables Generation 6/80 Examples (cont) Example 3 Toss a coin until the first tail occurs Assume the coin has p as its probability of a head If X is the number of heads, = x x = 0, 1, 2, ... and the X { | } pdf is f (x)= px (1 p) x = 0, 1, 2, ... − X is Geometric(p) and the set of possible values is infinite Verify that x f (x) = 1: P ∞ f (x)= px (1 p) = (1 p)(1+p+p2+p3+p4+ ) = 1 x x − − · · · X X=0 Discrete-Event Simulation Random Variables Generation 7/80 Cumulative Distribution Function The cumulative distribution function(cdf) of X is the real-valued function F ( ) for each x as · ∈X F (x)= Pr(X x)= f (t) ≤ ≤ tXx If X is Equilikely(a, b) then the cdf is x F (x)= 1/(b − a +1)=(x −a+1)/(b −a+1) x = a, a+1,..., b t=a X If X is Geometric(p) then the cdf is x F (x)= pt (1−p)=(1−p)(1+p+· · ·+px ) = 1−px+1 x = 0, 1, 2, ... t=0 X Discrete-Event Simulation Random Variables Generation 8/80 Example No simple equation for F ( ) for sum of two dice · is small enough to tabulate the cdf |X | 0.2 1.0 f(x) 0.1 F (x) 0.5 0.0 0.0 2 4 6 8 1012 2 4 6 8 1012 x x Discrete-Event Simulation Random Variables Generation 9/80 Relationship Between cdfs and pdfs A cdf can be generated from its corresponding pdf by recursion For example, = x x = a, a +1, ..., b X { | } F (a) = f (a) F (x) = F (x 1) + f (x) x = a +1, a +2, ..., b − A pdf can be generated from its corresponding cdf by subtraction f (a) = F (a) f (x) = F (x) F (x 1) x = a + 1, a + 2, ..., b − − A discrete random variable can be defined by specifying either its pdf or its cdf Discrete-Event Simulation Random Variables Generation 10/80 Other cdf Properties A cdf is strictly monotone increasing: if x1 < x2, then F (x1) < F (x2) The cdf values are bounded between 0.0 and 1.0 Monotonicity of F ( ) is the basis to generate discrete random · variates as we will see soon Discrete-Event Simulation Random Variables Generation 11/80 Mean and Standard Deviation The mean µ of the discrete random variable X is µ = xf (x) x X The variance of X is σ2 = (x µ)2f (x) or σ2 = x 2f (x) µ2 x − x ! − X X The corresponding standard deviation σ is σ = (x µ)2f (x) or σ = x 2f (x) µ2 v s x − u x ! − X u X t Discrete-Event Simulation Random Variables Generation 12/80 Examples If X is Equilikely(a, b) then the mean and standard deviation are b b a−1 x 1 • µ = = x − x b − a + 1 b − a + 1 x=a " x=1 x=1 # X X X 1 b(b + 1) (a − 1)a b2 − b − a2 + a a + b = − = = b − a + 1 2 2 2(b − a + 1) 2 b b b a−1 (x − µ)2 x 2 1 • σ2 = = − µ2 = x 2 − x 2 − µ2 b − a + 1 b − a + 1 b − a + 1 x=a x=a " x=1 x=1 # X X X X 1 b(b + 1)(2b + 1) (a − 1)a(2a − 1) = − − µ2 b − a + 1 6 6 2 (b − a + 1) − 1 ∗ = ( ) 12 35 When X is Equilikely(1, 6), µ = 3.5 and σ = 12 ∼= 1.708 q (*)recall that the sum of the squares of the first n natural numbers is n(n + 1)(2n + 1) S2 = n 6 Discrete-Event Simulation Random Variables Generation 13/80 Roll two fair dice: average If X is the sum of two dice then 12 12 6 7 x µ = xf (x)= x − | − | x x 36 X=2 X=2 6 6 7+ x 6 7 + 7 = [x + (14 x)] − + 7 − | | x − 36 36 X=2 6 x 1 7 7 6 7 = 14 − + = (x 1) + x 36 6 18 x − 6 X=2 X=2 7 4 7 7 4 7 = (2 + k 1) + = (1 + k)+ 18 − 6 18 6 kX=0 kX=0 7 4 (4 + 1) 7 7 21 = 5+ ∗ + = 10 + = 18 2 6 18 18 70 21 7 18 = + = ∗ = 7 18 18 18 Discrete-Event Simulation Random Variables Generation 14/80 Roll two fair dice: alternative derivations 12 7 x 1 12 13 x µ = xf (x)= x − + x − x x 36 x 36 X=2 X=2 X=8 6 x 1 6 2 y 1 = x − + 7 + (14 y) − x 36 36 y − 36 X=2 X=6 where y = 14 x. With this substitution it is easy to see that − 6 x 1 7 6 x 1 7 µ = [x + (14 x)] − + = 14 − + x − 36 6 x 36 6 X=2 X=2 7 6 7 7 4 7 = (x 1) + = (2 + k 1) + 18 x − 6 18 − 6 X=2 kX=0 7 4 7 7 4 (4 + 1) 7 = (1 + k)+ = 5+ ∗ + 18 6 18 2 6 kX=0 7 21 70 21 7 18 = 10 + = + = ∗ = 7 18 18 18 18 18 Discrete-Event Simulation Random Variables Generation 15/80 Roll two fair dice: variance If X is the sum of two dice then 12 σ2 = (x µ)2f (x)=35/6=5.813 − x=2 X Standard Deviation σ = 35/6 ∼= 2.415 p Discrete-Event Simulation Random Variables Generation 16/80 Another Example If X is Geometric(p) then the mean and standard deviation are ∞ ∞ p µ = xf (x)= xpx (1 p)= = x x − · · · 1 p X=0 X=1 − ∞ ∞ 2 2 2 2 2 x p σ = x f (x) µ = x p (1 p) 2 x ! − x − ! − (1 p) X=0 X=1 − . p σ2 = (1 p)2 − √p σ = (1 p) − Discrete-Event Simulation Random Variables Generation 17/80 Expected Value The mean of a random variable is also known as the expected value The expected value of the discrete random variable X is E[X]= xf (x)= µ x X Expected value refers to the expected average of a large sample x , x ,..., xn corresponding to X:x ¯ E[X]= µ as 1 2 → n .
Recommended publications
  • 4.2 Variance and Covariance
    4.2 Variance and Covariance The most important measure of variability of a random variable X is obtained by letting g(X) = (X − µ)2 then E[g(X)] gives a measure of the variability of the distribution of X. 1 Definition 4.3 Let X be a random variable with probability distribution f(x) and mean µ. The variance of X is σ 2 = E[(X − µ)2 ] = ∑(x − µ)2 f (x) x if X is discrete, and ∞ σ 2 = E[(X − µ)2 ] = ∫ (x − µ)2 f (x)dx if X is continuous. −∞ ¾ The positive square root of the variance, σ, is called the standard deviation of X. ¾ The quantity x - µ is called the deviation of an observation, x, from its mean. 2 Note σ2≥0 When the standard deviation of a random variable is small, we expect most of the values of X to be grouped around mean. We often use standard deviations to compare two or more distributions that have the same unit measurements. 3 Example 4.8 Page97 Let the random variable X represent the number of automobiles that are used for official business purposes on any given workday. The probability distributions of X for two companies are given below: x 12 3 Company A: f(x) 0.3 0.4 0.3 x 01234 Company B: f(x) 0.2 0.1 0.3 0.3 0.1 Find the variances of X for the two companies. Solution: µ = 2 µ A = (1)(0.3) + (2)(0.4) + (3)(0.3) = 2 B 2 2 2 2 σ A = (1− 2) (0.3) + (2 − 2) (0.4) + (3− 2) (0.3) = 0.6 2 2 4 σ B > σ A 2 2 4 σ B = ∑(x − 2) f (x) =1.6 x=0 Theorem 4.2 The variance of a random variable X is σ 2 = E(X 2 ) − µ 2.
    [Show full text]
  • A Tail Quantile Approximation Formula for the Student T and the Symmetric Generalized Hyperbolic Distribution
    A Service of Leibniz-Informationszentrum econstor Wirtschaft Leibniz Information Centre Make Your Publications Visible. zbw for Economics Schlüter, Stephan; Fischer, Matthias J. Working Paper A tail quantile approximation formula for the student t and the symmetric generalized hyperbolic distribution IWQW Discussion Papers, No. 05/2009 Provided in Cooperation with: Friedrich-Alexander University Erlangen-Nuremberg, Institute for Economics Suggested Citation: Schlüter, Stephan; Fischer, Matthias J. (2009) : A tail quantile approximation formula for the student t and the symmetric generalized hyperbolic distribution, IWQW Discussion Papers, No. 05/2009, Friedrich-Alexander-Universität Erlangen-Nürnberg, Institut für Wirtschaftspolitik und Quantitative Wirtschaftsforschung (IWQW), Nürnberg This Version is available at: http://hdl.handle.net/10419/29554 Standard-Nutzungsbedingungen: Terms of use: Die Dokumente auf EconStor dürfen zu eigenen wissenschaftlichen Documents in EconStor may be saved and copied for your Zwecken und zum Privatgebrauch gespeichert und kopiert werden. personal and scholarly purposes. Sie dürfen die Dokumente nicht für öffentliche oder kommerzielle You are not to copy documents for public or commercial Zwecke vervielfältigen, öffentlich ausstellen, öffentlich zugänglich purposes, to exhibit the documents publicly, to make them machen, vertreiben oder anderweitig nutzen. publicly available on the internet, or to distribute or otherwise use the documents in public. Sofern die Verfasser die Dokumente unter Open-Content-Lizenzen (insbesondere CC-Lizenzen) zur Verfügung gestellt haben sollten, If the documents have been made available under an Open gelten abweichend von diesen Nutzungsbedingungen die in der dort Content Licence (especially Creative Commons Licences), you genannten Lizenz gewährten Nutzungsrechte. may exercise further usage rights as specified in the indicated licence. www.econstor.eu IWQW Institut für Wirtschaftspolitik und Quantitative Wirtschaftsforschung Diskussionspapier Discussion Papers No.
    [Show full text]
  • On the Efficiency and Consistency of Likelihood Estimation in Multivariate
    On the efficiency and consistency of likelihood estimation in multivariate conditionally heteroskedastic dynamic regression models∗ Gabriele Fiorentini Università di Firenze and RCEA, Viale Morgagni 59, I-50134 Firenze, Italy <fi[email protected]fi.it> Enrique Sentana CEMFI, Casado del Alisal 5, E-28014 Madrid, Spain <sentana@cemfi.es> Revised: October 2010 Abstract We rank the efficiency of several likelihood-based parametric and semiparametric estima- tors of conditional mean and variance parameters in multivariate dynamic models with po- tentially asymmetric and leptokurtic strong white noise innovations. We detailedly study the elliptical case, and show that Gaussian pseudo maximum likelihood estimators are inefficient except under normality. We provide consistency conditions for distributionally misspecified maximum likelihood estimators, and show that they coincide with the partial adaptivity con- ditions of semiparametric procedures. We propose Hausman tests that compare Gaussian pseudo maximum likelihood estimators with more efficient but less robust competitors. Fi- nally, we provide finite sample results through Monte Carlo simulations. Keywords: Adaptivity, Elliptical Distributions, Hausman tests, Semiparametric Esti- mators. JEL: C13, C14, C12, C51, C52 ∗We would like to thank Dante Amengual, Manuel Arellano, Nour Meddahi, Javier Mencía, Olivier Scaillet, Paolo Zaffaroni, participants at the European Meeting of the Econometric Society (Stockholm, 2003), the Sympo- sium on Economic Analysis (Seville, 2003), the CAF Conference on Multivariate Modelling in Finance and Risk Management (Sandbjerg, 2006), the Second Italian Congress on Econometrics and Empirical Economics (Rimini, 2007), as well as audiences at AUEB, Bocconi, Cass Business School, CEMFI, CREST, EUI, Florence, NYU, RCEA, Roma La Sapienza and Queen Mary for useful comments and suggestions. Of course, the usual caveat applies.
    [Show full text]
  • Sampling Student's T Distribution – Use of the Inverse Cumulative
    Sampling Student’s T distribution – use of the inverse cumulative distribution function William T. Shaw Department of Mathematics, King’s College, The Strand, London WC2R 2LS, UK With the current interest in copula methods, and fat-tailed or other non-normal distributions, it is appropriate to investigate technologies for managing marginal distributions of interest. We explore “Student’s” T distribution, survey its simulation, and present some new techniques for simulation. In particular, for a given real (not necessarily integer) value n of the number of degrees of freedom, −1 we give a pair of power series approximations for the inverse, Fn ,ofthe cumulative distribution function (CDF), Fn.Wealsogivesomesimpleandvery fast exact and iterative techniques for defining this function when n is an even −1 integer, based on the observation that for such cases the calculation of Fn amounts to the solution of a reduced-form polynomial equation of degree n − 1. We also explain the use of Cornish–Fisher expansions to define the inverse CDF as the composition of the inverse CDF for the normal case with a simple polynomial map. The methods presented are well adapted for use with copula and quasi-Monte-Carlo techniques. 1 Introduction There is much interest in many areas of financial modeling on the use of copulas to glue together marginal univariate distributions where there is no easy canonical multivariate distribution, or one wishes to have flexibility in the mechanism for combination. One of the more interesting marginal distributions is the “Student’s” T distribution. This statistical distribution was published by W. Gosset in 1908.
    [Show full text]
  • The Smoothed Median and the Bootstrap
    Biometrika (2001), 88, 2, pp. 519–534 © 2001 Biometrika Trust Printed in Great Britain The smoothed median and the bootstrap B B. M. BROWN School of Mathematics, University of South Australia, Adelaide, South Australia 5005, Australia [email protected] PETER HALL Centre for Mathematics & its Applications, Australian National University, Canberra, A.C.T . 0200, Australia [email protected] G. A. YOUNG Statistical L aboratory, University of Cambridge, Cambridge CB3 0WB, U.K. [email protected] S Even in one dimension the sample median exhibits very poor performance when used in conjunction with the bootstrap. For example, both the percentile-t bootstrap and the calibrated percentile method fail to give second-order accuracy when applied to the median. The situation is generally similar for other rank-based methods, particularly in more than one dimension. Some of these problems can be overcome by smoothing, but that usually requires explicit choice of the smoothing parameter. In the present paper we suggest a new, implicitly smoothed version of the k-variate sample median, based on a particularly smooth objective function. Our procedure preserves many features of the conventional median, such as robustness and high efficiency, in fact higher than for the conventional median, in the case of normal data. It is however substantially more amenable to application of the bootstrap. Focusing on the univariate case, we demonstrate these properties both theoretically and numerically. Some key words: Bootstrap; Calibrated percentile method; Median; Percentile-t; Rank methods; Smoothed median. 1. I Rank methods are based on simple combinatorial ideas of permutations and sign changes, which are attractive in applications far removed from the assumptions of normal linear model theory.
    [Show full text]
  • A Note on Inference in a Bivariate Normal Distribution Model Jaya
    A Note on Inference in a Bivariate Normal Distribution Model Jaya Bishwal and Edsel A. Peña Technical Report #2009-3 December 22, 2008 This material was based upon work partially supported by the National Science Foundation under Grant DMS-0635449 to the Statistical and Applied Mathematical Sciences Institute. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the National Science Foundation. Statistical and Applied Mathematical Sciences Institute PO Box 14006 Research Triangle Park, NC 27709-4006 www.samsi.info A Note on Inference in a Bivariate Normal Distribution Model Jaya Bishwal¤ Edsel A. Pena~ y December 22, 2008 Abstract Suppose observations are available on two variables Y and X and interest is on a parameter that is present in the marginal distribution of Y but not in the marginal distribution of X, and with X and Y dependent and possibly in the presence of other parameters which are nuisance. Could one gain more e±ciency in the point estimation (also, in hypothesis testing and interval estimation) about the parameter of interest by using the full data (both Y and X values) instead of just the Y values? Also, how should one measure the information provided by random observables or their distributions about the parameter of interest? We illustrate these issues using a simple bivariate normal distribution model. The ideas could have important implications in the context of multiple hypothesis testing or simultaneous estimation arising in the analysis of microarray data, or in the analysis of event time data especially those dealing with recurrent event data.
    [Show full text]
  • Algorithms for Operations on Probability Distributions in a Computer Algebra System
    W&M ScholarWorks Dissertations, Theses, and Masters Projects Theses, Dissertations, & Master Projects 2001 Algorithms for operations on probability distributions in a computer algebra system Diane Lynn Evans College of William & Mary - Arts & Sciences Follow this and additional works at: https://scholarworks.wm.edu/etd Part of the Mathematics Commons, and the Statistics and Probability Commons Recommended Citation Evans, Diane Lynn, "Algorithms for operations on probability distributions in a computer algebra system" (2001). Dissertations, Theses, and Masters Projects. Paper 1539623382. https://dx.doi.org/doi:10.21220/s2-bath-8582 This Dissertation is brought to you for free and open access by the Theses, Dissertations, & Master Projects at W&M ScholarWorks. It has been accepted for inclusion in Dissertations, Theses, and Masters Projects by an authorized administrator of W&M ScholarWorks. For more information, please contact [email protected]. Reproduced with with permission permission of the of copyright the copyright owner. owner.Further reproductionFurther reproduction prohibited without prohibited permission. without permission. ALGORITHMS FOR OPERATIONS ON PROBABILITY DISTRIBUTIONS IN A COMPUTER ALGEBRA SYSTEM A Dissertation Presented to The Faculty of the Department of Applied Science The College of William & Mary in Virginia In Partial Fulfillment Of the Requirements for the Degree of Doctor of Philosophy by Diane Lynn Evans July 2001 Reproduced with permission of the copyright owner. Further reproduction prohibited without permission. UMI Number: 3026405 Copyright 2001 by Evans, Diane Lynn All rights reserved. ___ ® UMI UMI Microform 3026405 Copyright 2001 by Bell & Howell Information and Learning Company. All rights reserved. This microform edition is protected against unauthorized copying under Title 17, United States Code.
    [Show full text]
  • Chapter 23 Flexible Budgets and Standard Cost Systems
    Chapter 23 Flexible Budgets and Standard Cost Systems Review Questions 1. What is a variance? A variance is the difference between an actual amount and the budgeted amount. 2. Explain the difference between a favorable and an unfavorable variance. A variance is favorable if it increases operating income. For example, if actual revenue is greater than budgeted revenue or if actual expense is less than budgeted expense, then the variance is favorable. If the variance decreases operating income, the variance is unfavorable. For example, if actual revenue is less than budgeted revenue or if actual expense is greater than budgeted expense, the variance is unfavorable. 3. What is a static budget performance report? A static budget is a budget prepared for only one level of sales volume. A static budget performance report compares actual results to the expected results in the static budget and reports the differences (static budget variances). 4. How do flexible budgets differ from static budgets? A flexible budget is a budget prepared for various levels of sales volume within a relevant range. A static budget is prepared for only one level of sales volume—the expected number of units sold— and it doesn’t change after it is developed. 5. How is a flexible budget used? Because a flexible budget is prepared for various levels of sales volume within a relevant range, it provides the basis for preparing the flexible budget performance report and understanding the two components of the overall static budget variance (a static budget is prepared for only one level of sales volume, and a static budget performance report shows only the overall static budget variance).
    [Show full text]
  • Generating Random Samples from User-Defined Distributions
    The Stata Journal (2011) 11, Number 2, pp. 299–304 Generating random samples from user-defined distributions Katar´ına Luk´acsy Central European University Budapest, Hungary lukacsy [email protected] Abstract. Generating random samples in Stata is very straightforward if the distribution drawn from is uniform or normal. With any other distribution, an inverse method can be used; but even in this case, the user is limited to the built- in functions. For any other distribution functions, their inverse must be derived analytically or numerical methods must be used if analytical derivation of the inverse function is tedious or impossible. In this article, I introduce a command that generates a random sample from any user-specified distribution function using numeric methods that make this command very generic. Keywords: st0229, rsample, random sample, user-defined distribution function, inverse method, Monte Carlo exercise 1 Introduction In statistics, a probability distribution identifies the probability of a random variable. If the random variable is discrete, it identifies the probability of each value of this vari- able. If the random variable is continuous, it defines the probability that this variable’s value falls within a particular interval. The probability distribution describes the range of possible values that a random variable can attain, further referred to as the sup- port interval, and the probability that the value of the random variable is within any (measurable) subset of that range. Random sampling refers to taking a number of independent observations from a probability distribution. Typically, the parameters of the probability distribution (of- ten referred to as true parameters) are unknown, and the aim is to retrieve them using various estimation methods on the random sample generated from this probability dis- tribution.
    [Show full text]
  • Comparison of the Estimation Efficiency of Regression
    J. Stat. Appl. Pro. Lett. 6, No. 1, 11-20 (2019) 11 Journal of Statistics Applications & Probability Letters An International Journal http://dx.doi.org/10.18576/jsapl/060102 Comparison of the Estimation Efficiency of Regression Parameters Using the Bayesian Method and the Quantile Function Ismail Hassan Abdullatif Al-Sabri1,2 1 Department of Mathematics, College of Science and Arts (Muhayil Asir), King Khalid University, KSA 2 Department of Mathematics and Computer, Faculty of Science, Ibb University, Yemen Received: 3 Jul. 2018, Revised: 2 Sep. 2018, Accepted: 12 Sep. 2018 Published online: 1 Jan. 2019 Abstract: There are several classical as well as modern methods to predict model parameters. The modern methods include the Bayesian method and the quantile function method, which are both used in this study to estimate Multiple Linear Regression (MLR) parameters. Unlike the classical methods, the Bayesian method is based on the assumption that the model parameters are variable, rather than fixed, estimating the model parameters by the integration of prior information (prior distribution) with the sample information (posterior distribution) of the phenomenon under study. By contrast, the quantile function method aims to construct the error probability function using least squares and the inverse distribution function. The study investigated the efficiency of each of them, and found that the quantile function is more efficient than the Bayesian method, as the values of Theil’s Coefficient U and least squares of both methods 2 2 came to be U = 0.052 and ∑ei = 0.011, compared to U = 0.556 and ∑ei = 1.162, respectively. Keywords: Bayesian Method, Prior Distribution, Posterior Distribution, Quantile Function, Prediction Model 1 Introduction Multiple linear regression (MLR) analysis is an important statistical method for predicting the values of a dependent variable using the values of a set of independent variables.
    [Show full text]
  • On Half-Cauchy Distribution and Process
    International Journal of Statistika and Mathematika, ISSN: 2277- 2790 E-ISSN: 2249-8605, Volume 3, Issue 2, 2012 pp 77-81 On Half-Cauchy Distribution and Process Elsamma Jacob1 , K Jayakumar2 1Malabar Christian College, Calicut, Kerala, INDIA. 2Department of Statistics, University of Calicut, Kerala, INDIA. Corresponding addresses: [email protected], [email protected] Research Article Abstract: A new form of half- Cauchy distribution using sin ξ cos ξ Marshall-Olkin transformation is introduced. The properties of () = − ξ, () = ξ, ≥ 0 the new distribution such as density, cumulative distribution ξ ξ function, quantiles, measure of skewness and distribution of the extremes are obtained. Time series models with half-Cauchy Remark 1.1. For the HC distribution the moments do distribution as stationary marginal distributions are not not exist. developed so far. We develop first order autoregressive process Remark 1.2. The HC distribution is infinitely divisible with the new distribution as stationary marginal distribution and (Bondesson (1987)) and self decomposable (Diedhiou the properties of the process are studied. Application of the (1998)). distribution in various fields is also discussed. Keywords: Autoregressive Process, Geometric Extreme Stable, Relationship with other distributions: Half-Cauchy Distribution, Skewness, Stationarity, Quantiles. 1. Let Y be a folded t variable with pdf given by 1. Introduction: ( ) > 0, (1.5) The half Cauchy (HC) distribution is derived from 2Γ( ) () = 1 + , ν ∈ , the standard Cauchy distribution by folding the ν ()√νπ curve on the origin so that only positive values can be observed. A continuous random variable X is said When ν = 1 , (1.5) reduces to 2 1 > 0 to have the half Cauchy distribution if its survival () = , function is given by 1 + (x)=1 − tan , > 0 (1.1) Thus, HC distribution coincides with the folded t distribution with = 1 degree of freedom.
    [Show full text]
  • The Third Moment in Equity Portfolio Theory – Skew
    May 2020 Michael Cheung, Head of Quantitative Research & Portfolio Manager Michael Messinger, Member & Portfolio Manager Summary Richard Duff, President & Numbers are nothing more than an abstract way to describe and model natural phenomena. Portfolio Manager Applied to capital markets, historical numerical data is often utilized as a proxy to describe the Michael Sasaki, CFA, Analyst behavior of security price movement. Harry Markowitz proposed Modern Portfolio Theory (MPT) in 1952, and since then, it has been widely adopted as the standard for portfolio optimization based on expected returns and market risk. While the principles in MPT are important to portfolio Skew Portfolio Theory construction, they fail to address real-time investment challenges. In this primer, we explore the (SPT) utilizes statistical efficacy of statistical skew and its application to real-time portfolio management. What we find is skew to build a that statistical skew can be utilized to build a more dynamic portfolio with potential to capture dynamic weighted additional risk premia – Skew Portfolio Theory (SPT). We then apply SPT to propose a new portfolio with potential rebalancing process based on statistical skewness (Skew). Skew not only captures risk in terms to capture additional of standard deviation, but analyzing how skew changes over time, we can measure how risk premia. statistical trends are changing, thus facilitating active management decisions to accommodate the trends. The positive results led to the creation and management of Redwood’s Equity Skew Strategy that is implemented in real-time. In short: • Skew is a less commonly viewed statistic that captures the mean, standard deviation, and tilt of historical returns in one number.
    [Show full text]