Introduction to Probability and Statistics Using R

Total Page:16

File Type:pdf, Size:1020Kb

Introduction to Probability and Statistics Using R Introduction to Probability and Statistics Using R G. Jay Kerns First Edition ii IPSUR: Introduction to Probability and Statistics Using R Copyright © 2010 G. Jay Kerns ISBN: 978-0-557-24979-4 Permission is granted to copy, distribute and/or modify this document under the terms of the GNU Free Documentation License, Version 1.3 or any later version published by the Free Software Foun- dation; with no Invariant Sections, no Front-Cover Texts, and no Back-Cover Texts. A copy of the license is included in the section entitled “GNU Free Documentation License”. Date: March 24, 2011 Contents Preface vii List of Figures xiii List of Tables xv 1 An Introduction to Probability and Statistics1 1.1 Probability....................................... 1 1.2 Statistics........................................ 1 Chapter Exercises...................................... 2 2 An Introduction to R 3 2.1 Downloading and Installing R ............................ 3 2.2 Communicating with R ................................ 4 2.3 Basic R Operations and Concepts .......................... 6 2.4 Getting Help...................................... 12 2.5 External Resources.................................. 14 2.6 Other Tips....................................... 14 Chapter Exercises...................................... 16 3 Data Description 17 3.1 Types of Data..................................... 17 3.2 Features of Data Distributions ............................ 33 3.3 Descriptive Statistics ................................. 35 3.4 Exploratory Data Analysis .............................. 41 3.5 Multivariate Data and Data Frames.......................... 45 3.6 Comparing Populations................................ 47 Chapter Exercises...................................... 55 4 Probability 67 4.1 Sample Spaces .................................... 67 4.2 Events......................................... 73 4.3 Model Assignment .................................. 78 4.4 Properties of Probability ............................... 83 4.5 Counting Methods .................................. 87 4.6 Conditional Probability................................ 93 iii iv CONTENTS 4.7 Independent Events.................................. 99 4.8 Bayes’ Rule......................................102 4.9 Random Variables...................................106 Chapter Exercises......................................109 5 Discrete Distributions 111 5.1 Discrete Random Variables..............................112 5.2 The Discrete Uniform Distribution..........................114 5.3 The Binomial Distribution ..............................116 5.4 Expectation and Moment Generating Functions...................122 5.5 The Empirical Distribution..............................125 5.6 Other Discrete Distributions .............................128 5.7 Functions of Discrete Random Variables.......................136 Chapter Exercises......................................138 6 Continuous Distributions 143 6.1 Continuous Random Variables ............................143 6.2 The Continuous Uniform Distribution ........................148 6.3 The Normal Distribution ...............................149 6.4 Functions of Continuous Random Variables.....................153 6.5 Other Continuous Distributions............................157 Chapter Exercises......................................164 7 Multivariate Distributions 165 7.1 Joint and Marginal Probability Distributions.....................166 7.2 Joint and Marginal Expectation............................172 7.3 Conditional Distributions...............................174 7.4 Independent Random Variables............................176 7.5 Exchangeable Random Variables...........................178 7.6 The Bivariate Normal Distribution..........................179 7.7 Bivariate Transformations of Random Variables...................181 7.8 Remarks for the Multivariate Case..........................184 7.9 The Multinomial Distribution.............................186 Chapter Exercises......................................190 8 Sampling Distributions 191 8.1 Simple Random Samples...............................192 8.2 Sampling from a Normal Distribution ........................193 8.3 The Central Limit Theorem..............................196 8.4 Sampling Distributions of Two-Sample Statistics ..................197 8.5 Simulated Sampling Distributions ..........................200 Chapter Exercises......................................203 9 Estimation 205 9.1 Point Estimation....................................205 9.2 Confidence Intervals for Means............................214 9.3 Confidence Intervals for Differences of Means....................221 CONTENTS v 9.4 Confidence Intervals for Proportions.........................223 9.5 Confidence Intervals for Variances..........................225 9.6 Fitting Distributions..................................225 9.7 Sample Size and Margin of Error...........................225 9.8 Other Topics......................................227 Chapter Exercises......................................228 10 Hypothesis Testing 229 10.1 Introduction......................................229 10.2 Tests for Proportions .................................230 10.3 One Sample Tests for Means and Variances .....................235 10.4 Two-Sample Tests for Means and Variances.....................239 10.5 Other Hypothesis Tests................................241 10.6 Analysis of Variance .................................241 10.7 Sample Size and Power................................243 Chapter Exercises......................................248 11 Simple Linear Regression 249 11.1 Basic Philosophy ...................................249 11.2 Estimation.......................................253 11.3 Model Utility and Inference..............................262 11.4 Residual Analysis...................................267 11.5 Other Diagnostic Tools................................275 Chapter Exercises......................................283 12 Multiple Linear Regression 285 12.1 The Multiple Linear Regression Model........................285 12.2 Estimation and Prediction...............................288 12.3 Model Utility and Inference..............................296 12.4 Polynomial Regression................................299 12.5 Interaction.......................................304 12.6 Qualitative Explanatory Variables ..........................307 12.7 Partial F Statistic ...................................310 12.8 Residual Analysis and Diagnostic Tools .......................312 12.9 Additional Topics...................................313 Chapter Exercises......................................317 13 Resampling Methods 319 13.1 Introduction......................................319 13.2 Bootstrap Standard Errors...............................321 13.3 Bootstrap Confidence Intervals............................326 13.4 Resampling in Hypothesis Tests ...........................328 Chapter Exercises......................................332 14 Categorical Data Analysis 333 15 Nonparametric Statistics 335 vi CONTENTS 16 Time Series 337 A R Session Information 339 B GNU Free Documentation License 341 C History 349 D Data 351 D.1 Data Structures ....................................351 D.2 Importing Data ....................................356 D.3 Creating New Data Sets................................357 D.4 Editing Data......................................357 D.5 Exporting Data ....................................359 D.6 Reshaping Data....................................359 E Mathematical Machinery 361 E.1 Set Algebra......................................361 E.2 Differential and Integral Calculus...........................362 E.3 Sequences and Series.................................365 E.4 The Gamma Function.................................368 E.5 Linear Algebra ....................................368 E.6 Multivariable Calculus ................................369 F Writing Reports with R 373 F.1 What to Write.....................................373 F.2 How to Write It with R ................................374 F.3 Formatting Tables...................................377 F.4 Other Formats.....................................377 G Instructions for Instructors 379 G.1 Generating This Document..............................380 G.2 How to Use This Document..............................381 G.3 Ancillary Materials..................................381 G.4 Modifying This Document ..............................382 H RcmdrTestDrive Story 383 Bibliography 389 Index 395 Preface This book was expanded from lecture materials I use in a one semester upper-division undergradu- ate course entitled Probability and Statistics at Youngstown State University. Those lecture mate- rials, in turn, were based on notes that I transcribed as a graduate student at Bowling Green State University. The course for which the materials were written is 50-50 Probability and Statistics, and the attendees include mathematics, engineering, and computer science majors (among others). The catalog prerequisites for the course are a full year of calculus. The book can be subdivided into three basic parts. The first part includes the introductions and elementary descriptive statistics; I want the students to be knee-deep in data right out of the gate. The second part is the study of probability, which begins at the basics of sets and the equally likely model, journeys past discrete/continuous random variables, and continues through to multivariate distributions. The chapter on sampling distributions
Recommended publications
  • A Test for Singularity 1
    STATISTIII & PROBABIUTY Lk'TTERS ELSEVIER Statistics & Probability Letters 40 (1998) 215-226 A test for singularity 1 Attila Frigyesi *, Ola H6ssjer Institute for Matematisk Statistik, Lunds University, Box 118, 22100 Lund, Sweden Received March 1997; received in revised form February 1998 Abstract The LP-norm and other functionals of the kernel estimate of density functions (as functions of the bandwidth) are studied as means to test singularity. Given an absolutely continuous distribution function, the LP-norm, and the other studied functionals, will converge under some mild side-constraints as the bandwidth decreases. In the singular case, the functionals will diverge. These properties may be used to test whether data comes from an arbitrarily large class of absolutely continuous distributions or not. (~) 1998 Elsevier Science B.V. All rights reserved AMS classification." primary 60E05; secondary 28A99; 62G07 Keywords: Singularity; Absolute continuity; Kernel density estimation 1. Introduction In this paper we aim to test whether a given set of data (= {XI ..... X,}), with a marginal distribution #, stems from an absolutely continuous distribution function or from a singular one. A reference to this problem is by Donoho (1988). Consider a probability measure # on (g~P,~(~P)), where ~(~P) is the Borel a-algebra on R p. The measure tt is said to be singular with respect to the Lebesgue measure 2 if there exists a measurable set A such that /~(AC)=0 and 2(A)=0. If we only require #(A)>0 and 2(A)=0, then /~ is said to have a singular part. If tt does not have a singular part, then it is absolutely continuous with respect to 2.
    [Show full text]
  • Vector Spaces
    ApPENDIX A Vector Spaces This appendix reviews some of the basic definitions and properties of vectm spaces. It is presumed that, with the possible exception of Theorem A.14, all of the material presented here is familiar to the reader. Definition A.I. A set.A c Rn is a vector space if, for any x, y, Z E .A and scalars a, 13, operations of vector addition and scalar multiplication are defined such that: (1) (x + y) + Z = x + (y + z). (2) x + y = y + x. (3) There exists a vector 0 E.A such that x + 0 = x = 0 + x. (4) For any XE.A there exists y = -x such that x + y = 0 = y + x. (5) a(x + y) = ax + ay. (6) (a + f3)x = ax + f3x. (7) (af3)x = a(f3x). (8) 1· x = x. Definition A.2. Let .A be a vector space and let JV be a set with JV c .A. JV is a subspace of .A if and only if JV is a vector space. Theorem A.3. Let .A be a vector space and let JV be a nonempty subset of .A. If JV is closed under addition and scalar multiplication, then JV is a subspace of .A. Theorem A.4. Let .A be a vector space and let Xl' ••• , Xr be in .A. The set of all linear combinations of Xl' .•• ,xr is a subspace of .A. Appendix A. Vector Spaces 325 Definition A.S. The set of all linear combinations of Xl'OO',X, is called the space spanned by Xl' ..
    [Show full text]
  • Quantization of Self-Similar Probability Measures and Optimal Quantizers by Do˘Ganc¸¨Omez Department of Mathematics, North Dakota State University
    Quantization of self-similar probability measures and optimal quantizers by Do˘ganC¸¨omez Department of Mathematics, North Dakota State University Outline: 1. Five-minute Crash Course in Probability 2. Probability Measures and Distributions 3. Voronoi Partitions 4. Centers and Moments of Probability Distributions 5. Quantization 6. Some Recent Results n Throughout we will deal with R with a norm k k: In particular k kr will be the r-norm for some 1 ≤ r < 1: Some of the information below are adopted from Foundations of Quantization for Probability Distributions, by S. Graf and H. Luschgy, Lecture Notes in Mathematics, # 1730, Springer, 2000. 1. Five-minute Crash Course in Probability. (Ω; F;P ) : a probability space. A measurable function X :Ω ! R is called a random variable. R The expectation (or mean) of X is defined as E(X) := Ω XdP: (i) E(Xk): k-th moment of X; and E(jXjk): k-th absolute moment of X: (ii) E[(X − E(X))k]: k-th central moment of X; and E[jX − E(X)jk]: k-th absolute central moment of X: Note: If k = 1;E[(X − E(X))k] = 0: When k = 2;E[(X − E(X))2] is called the variance of X: If X is a random variable, then it induces a probability measure on Borel subsets of R by −1 PX := P ◦ X : (Sometimes PX is called a law.) The distribution (function) of a r.v. X is a function F : R ! R such that F (x) = PX (−∞; x] = P (f! : X(!) ≤ xg): Observe that F (b) − F (a) = PX (a; b]: Fact.
    [Show full text]
  • THE DERIVATION of the CHI-Square TEST of GOODNESS of FIT
    THE DERIVATION OF THE CHI-sQUARE TEST OF GOODNESS OF FIT Jacqueline Prillaman Submitted to McGill University in partial fulfilment of the requirements for the de6ree of Master of Science, August, 1956 I wish to express my gratitude to Prof. W. Kozakiewicz for his generous assistance and valuable guidance in the preparation of this thesis. TABLE OF CONTENTS Introduction .................................................. 1 Part One: An Intuitive Approach to the Chi-Square Test of Goodness of Fit ............................ 5 I: Presentation of the Test II: The Number of Degrees of Fre?dom III: Greenhood's Derivation of the Chi-Square Test of Goodness of Fit Part Two: A Rigorous De velopment of the Chi- Square Test of Goodness of Fit .......................... 22 I: The Normal Distribution in the Space of n-Dimensions II: The Limiting Distribution of the Chi-Square Variable by the Method of Characteristic Functions Introduction Let ~, ~, . ., x be s independent sample values drawn from 5 a population of normally distributed values with zero mean and unit variance. Then the variable 2 2 2 u = x1 + x 2 + . + xs is said to follow the ~t distribution with parameter s. This para­ meter is called the number of degrees of freedom for reasons which will be explained later. Using the moment-generating function we shall prove that s s-2 u 2 2- (1) u e 2 u > o. f(u) = r(~) (!) Since the basic variable, x, in the population has the frequency function f(x) = m.- e , we have, for G < l2 , M 2(&) = Mx2(&) = xi +• 2 ~ (1-2&) = 1 e -2 dx rh 1-· 2 1 .
    [Show full text]
  • On Lin's Condition for Products of Random Variables with Singular Joint
    PROBABILITY AND MATHEMATICAL STATISTICS Vol. 40, Fasc. 1 (2020), pp. 97–104 Published online 2.4.2020 doi:10.37190/0208-4147.40.1.6 ON LIN’S CONDITION FOR PRODUCTS OF RANDOM VARIABLES WITH SINGULAR JOINT DISTRIBUTION BY ALEXANDER I L’ I N S K I I (KHARKOV) AND SOFIYA O S T ROV S K A (ANKARA) Abstract. Lin’s condition is used to establish the moment determi- nacy/indeterminacy of absolutely continuous probability distributions. Re- cently, a number of papers related to Lin’s condition for functions of random variables have appeared. In the present paper, this condition is studied for products of random variables with given densities in the case when their joint distribution is singular. It is proved, assuming that the densities of both random variables satisfy Lin’s condition, that the density of their product may or may not satisfy this condition. 2020 Mathematics Subject Classification: Primary 60E05; Secondary 44A60. Key words and phrases: random variable, singular distribution, Lin’s con- dition. 1. INTRODUCTION Lin’s condition plays a significant role in the investigation of the moment deter- minacy of absolutely continuous probability distributions. Generally speaking, it is used along with Krein’s logarithmic integral to establish the so-called ‘converse criteria’. See, for example, [4, Section 5, Theorems 5–10] and [7, Section 2]. This condition was introduced and applied by G. D. Lin [3], while the name ‘Lin’s con- dition’ was put forth in [10]. Let us recall the pertinent notions. DEFINITION 1.1. Let f be a probability density continuously differentiable on (0; +∞).
    [Show full text]
  • Multivariate Stable Distributions
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Elsevier - Publisher Connector JOURNAL OF MULTIVARIATE ANALYSIS 2, 444-462 (1972) Multivariate Stable Distributions S. J. PRESS* Graduate School of Business, University of Chicago, Chicago, Illinois 60637 Communicated by A. P. Dempster This paper is devoted to the theory and application of multidimensional stable distributions. Properties of these laws are developed and explicit algebraic representations are given in terms of characteristic functions. Symmetric and asymmetric laws and their properties are studied separately. A measure of association for variables following a symmetric bivariate stable distribution is provided and its properties are shown to be analogous to those of the ordinary correlation coefficient. Examples are given, and the class of symmetric multi- variate stable distributions is applied to a problem in portfolio analysis. 1. INTRODUCTION This work is concerned with probabilistic aspects of multivariate stable distributions. In Section 2, this class of distributions is defined, algebraic characteristic function representations are given, and some properties of the classare presented. Section 3 defines the symmetric multivariate stable distribu- tions and presents some of their properties. Since the variables under study generally do not possesssecond moments, correlation coefficients do not exist. However, an extended notion of correlation coefficient applicable to symmetric multivariate stable laws is presented in Section
    [Show full text]
  • Statistical Theory Distribution Theory Reading in Casella and Berger
    STAT 450: Statistical Theory Distribution Theory Reading in Casella and Berger: Ch 2 Sec 1, Ch 4 Sec 1, Ch 4 Sec 6. Basic Problem: Start with assumptions about f or CDF of random vector X = (X1;:::;Xp). Define Y = g(X1;:::;Xp) to be some function of X (usually some statistic of interest like a sample mean or correlation coefficient or a test statistic). How can we compute the distribution or CDF or density of Y ? Univariate Techniques In a univariate problem we have X a real-valued random variable and Y , a real valued transformation of X. Often we start with X and are told either the density or CDF of X and then asked to find the distribution of something like X2 or sin(X) { a function of X which we call a transformation. I am going to show you the steps you should follow in an example problem. Method 1: compute the CDF by integration and differentiate to find fY . Example: Suppose U ∼ Uniform[0; 1]. What is the distribution of − log(X)?. Step 1: Make sure you give the transformed variable a name. Let Y = − log U. Step 2: Write down the definition of the CDF of Y : FY (y) = P (Y ≤ y): Step 3: Substitute in the definition of Y in terms of the original variable { in this case, U: FY (y) = P (Y ≤ y) = P (− log(U) ≤ y): Step 4: Solve the inequality to get, hopefully, one or more intervals for U. In this case − log(U) ≤ y is the same as log(U) ≥ −y 1 which is the same as U ≥ e−y: So P (− log(U) ≤ Y ) = P U ≥ e−y : Step 5: now use the assumptions to write the resulting probability in terms of FU .
    [Show full text]
  • Splitting Models for Multivariate Count Data Jean Peyhardi, Pierre Fernique, Jean-Baptiste Durand
    Splitting models for multivariate count data Jean Peyhardi, Pierre Fernique, Jean-Baptiste Durand To cite this version: Jean Peyhardi, Pierre Fernique, Jean-Baptiste Durand. Splitting models for multivariate count data. Journal of Multivariate Analysis, Elsevier, 2021, 181, pp.104677. 10.1016/j.jmva.2020.104677. hal- 02962877 HAL Id: hal-02962877 https://hal.inria.fr/hal-02962877 Submitted on 9 Oct 2020 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Splitting models for multivariate count data Jean Peyhardia,∗, Pierre Ferniqueb, Jean-Baptiste Durandc aIMAG, Universit´ede Montpellier, place Eug`eneBataillon, 34090 Montpellier, France bCIRAD, UMR AGAP, 34098 Montpellier cLaboratoire Jean Kuntzmann, 38058 Grenoble, France Abstract We introduce the class of splitting distributions as the composition of a singular multivariate distribution and an uni- variate distribution. This class encompasses most common multivariate discrete distributions (multinomial, negative multinomial, multivariate hypergeometric, multivariate negative hypergeometric, . ) and contains several new ones. We highlight many probabilistic properties deriving from the compound aspect of splitting distributions and their un- derlying algebraic properties. These simplify the study of their characteristics, inference methods, interpretation and extensions to regression models. Parameter inference and model selection are thus reduced to two separate problems, preserving time and space complexity of the base models.
    [Show full text]
  • the MULTIVARIATE T-DISTRIBUTION ASSOCIATED
    THE MULTIVARIATE t-DISTRIBUTION ASSOCIATED WITH A SET OF NORMAL SAMPLE DEVIATES By E. A. OORNIsn* [Manu8cript received July 29, 1954] Sumnwry This paper gives a short account of the more important properties of the multi­ variate t-distribution, which arises in association with a set of normal sample deviates. I. INTRODUCTION The -multivariate t-distribution described herein was encountered in the course of a recent investigation of the frequency distribution of the spectro­ graphic error which occurs in the D.O. arc excitation of samples of soil and plant ash (Oertel and Oornish 1953). The original observations consisted of triplicate and sextuplicate determinations of copper, manganese, molybdenum, and tin in samples containing varying amounts of each of these elements, and it was found that the variance of measured line intensity was proportional to the square of the mean intensity. A logarithmic transformation stabilized the variance, which, for any individual element, could then be estimated from all the samples involving 1hat element, and by subsequent standardization all the values of the new metri' could be placed on a comparable basis. The standardized variates appeared to be normally distributed, and it became desirable, for various reasons, to test this point. In providing tests of normality, two general lines of approach have been / followed: (1) a normal distribution is fitted to the sample data, and the X2 test of goodness of fit is applied; (2) certain functions of the sample moments are calculated, and the significance of their departure from expectations based on the assumption of normality is examined, but neither of these procedures, as ordinarily used, was suitable owing to the particular nature of the data.
    [Show full text]
  • Some Methods of Constructing Multivariate Distributions Abderrahmane Chakak Iowa State University
    Iowa State University Capstones, Theses and Retrospective Theses and Dissertations Dissertations 1993 Some methods of constructing multivariate distributions Abderrahmane Chakak Iowa State University Follow this and additional works at: https://lib.dr.iastate.edu/rtd Part of the Statistics and Probability Commons Recommended Citation Chakak, Abderrahmane, "Some methods of constructing multivariate distributions " (1993). Retrospective Theses and Dissertations. 10411. https://lib.dr.iastate.edu/rtd/10411 This Dissertation is brought to you for free and open access by the Iowa State University Capstones, Theses and Dissertations at Iowa State University Digital Repository. It has been accepted for inclusion in Retrospective Theses and Dissertations by an authorized administrator of Iowa State University Digital Repository. For more information, please contact [email protected]. INFORMATION TO USERS This manuscript has been reproduced from the microfilm master. UMI films the text directly from the original or copy submitted. Thus, some thesis and dissertation copies are in typewriter face, while others may be from any type of computer printer. The quality of this reproduction is dependent upon the quality of the copy submitted. Broken or indistinct print, colored or poor quality illustrations and photographs, print bleedthrough, substandard margins, and improper alignment can adversely affect reproduction. In the unlikely event that the author did not send UMI a complete manuscript and there are missing pages, these will be noted. Also, if unauthorized copyright material had to be removed, a note will indicate the deletion. Oversize materials (e.g., maps, drawings, charts) are reproduced by sectioning the original, beginning at the upper left-hand corner and continuing from left to right in equal sections with small overlaps.
    [Show full text]
  • Econstor Wirtschaft Leibniz Information Centre Make Your Publications Visible
    A Service of Leibniz-Informationszentrum econstor Wirtschaft Leibniz Information Centre Make Your Publications Visible. zbw for Economics Erdely, Arturo Article Value at risk and the diversification dogma Revista de Métodos Cuantitativos para la Economía y la Empresa Provided in Cooperation with: Universidad Pablo de Olavide, Sevilla Suggested Citation: Erdely, Arturo (2017) : Value at risk and the diversification dogma, Revista de Métodos Cuantitativos para la Economía y la Empresa, ISSN 1886-516X, Universidad Pablo de Olavide, Sevilla, Vol. 24, pp. 209-219 This Version is available at: http://hdl.handle.net/10419/195388 Standard-Nutzungsbedingungen: Terms of use: Die Dokumente auf EconStor dürfen zu eigenen wissenschaftlichen Documents in EconStor may be saved and copied for your Zwecken und zum Privatgebrauch gespeichert und kopiert werden. personal and scholarly purposes. Sie dürfen die Dokumente nicht für öffentliche oder kommerzielle You are not to copy documents for public or commercial Zwecke vervielfältigen, öffentlich ausstellen, öffentlich zugänglich purposes, to exhibit the documents publicly, to make them machen, vertreiben oder anderweitig nutzen. publicly available on the internet, or to distribute or otherwise use the documents in public. Sofern die Verfasser die Dokumente unter Open-Content-Lizenzen (insbesondere CC-Lizenzen) zur Verfügung gestellt haben sollten, If the documents have been made available under an Open gelten abweichend von diesen Nutzungsbedingungen die in der dort Content Licence (especially Creative Commons Licences), you genannten Lizenz gewährten Nutzungsrechte. may exercise further usage rights as specified in the indicated licence. https://creativecommons.org/licenses/by-sa/4.0/ www.econstor.eu REVISTA DE METODOS´ CUANTITATIVOS PARA LA ECONOM´IA Y LA EMPRESA (24).
    [Show full text]
  • Doubly Singular Matrix Variate Beta Type I and II and Singular Inverted Matricvariate $ T $ Distributions
    Doubly singular matrix variate beta type I and II and singular inverted matricvariate t distributions Jos´eA. D´ıaz-Garc´ıa ∗ Department of Statistics and Computation 25350 Buenavista, Saltillo, Coahuila, Mexico E-mail: [email protected] Ram´on Guti´errez J´aimez Department of Statistics and O.R. University of Granada Granada 18071, Spain E-mail: [email protected] Abstract In this paper, the densities of the doubly singular beta type I and II distributions are found, and the joint densities of their corresponding nonzero eigenvalues are provided. As a consequence, the density function of a singular inverted matricvariate t distribution is obtained. 1 Introduction Let A and B be independent Wishart matrices or, alternatively, let B = YY′ where Y has a matrix variate normal distribution. Then the matrix variate beta type I distribution is arXiv:0904.2147v1 [math.ST] 14 Apr 2009 defined in the literature as (A + B)−1/2B(A + B)−1/2, U = B1/2(A + B)−1B1/2, (1) ′ −1 Y (A + B) Y. Analogously, the matrix variate beta type II distribution is defined as A−1/2BA−1/2, F = B1/2A−1B1/2, (2) ′ −1 Y A Y. These can be classified as central, noncentral or doubly noncentral, depending on whether A and B are central, B is noncentral or A and B are noncentral, respectively; see D´ıaz-Garc´ıaand Guti´errez-J´aimez ∗Corresponding author Key words. Singular distribution, matricvariate t distribution, matrix variate beta type I distribution, matrix variate beta type II distribution. 2000 Mathematical Subject Classification. 62E15, 15A52 1 (2007, 2008b).
    [Show full text]