4. MULTIVARIATE NORMAL DISTRIBUTION (Part I) 1 Lecture 3 Review

4. MULTIVARIATE NORMAL DISTRIBUTION (Part I) 1 Lecture 3 Review

4. MULTIVARIATE NORMAL DISTRIBUTION (Part I) 1 Lecture 3 Review: Random vectors: vectors of random variables. The expectation of a random vector is just the vector of expectations. • cov(X, Y) is a matrix with i, j entry cov(X , Y ) • i j cov(AX, BY)= Acov(X, Y)B0 • We introduced quadratic forms – X0AX, where X is a random vector and A is • a matrix. More to come ... 4.1 Definition of the Multivariate Normal Distribution The following are equivalent definitions of the multivariate normal distribution (MVN). Given a vector µ and p.s.d. matrix Σ, Y N (µ, Σ) if: ∼ n Definition 1: For p.d. Σ, the density function of Y is −n/2 −1/2 1 0 −1 fY(y)=(2π) Σ exp (y µ) Σ (y µ) . | | {−2 − − } Definition 2: The moment generating function (m.g.f.) of Y is t0Y 0 1 0 MY(t) E[e ] = exp µ t + t Σt . ≡ { 2 } Definition 3: Y has the same distribution as AZ + µ, where Z =(Z1,...,Zk) are independent N(0, 1) random variables and An×k satisfies AA0 = Σ. COMMENT: You may be inclined to focus on definition 1, but the others are more useful. 2 4. MULTIVARIATE NORMAL DISTRIBUTION (Part I) Theorem: Definitions 1, 2, and 3 are equivalent for Σ > 0. Definitions 2 and 3 are equivalent for Σ 0 ≥ Proof of Def 3 Def 2: ⇒ For Z N(0, 1), i ∼ ∞ ∞ 2 2 − − i− i zi /2 2 (z t ) /2 2 tiZi ziti e ti /2 e ti /2 MZi (ti)= E[e ]= e dzi = e dzi = e . Z √2π Z √2π −∞ −∞ If Z =(Z1,...,Zk) is a random sample from N(0, 1), then k k k k Σiziti ziti ind ziti 2 0 MZ(t)= E[e ]= E[ e ] = E[e ]= M i (t ) = exp t /2 = exp t t/2 . Z i { i } { } Yi=1 Yi=1 Yi=1 Xi=1 If Y = AZ + µ, 0 MY(t) E[exp Y t ] ≡ { } = E[exp (AZ + µ)0t ] { } = exp µ0t E[exp (AZ)0t ] { 0 } {0 } = exp µ t MZ(A t) { } 1 = exp µ0t exp (A0t)0(A0t) { } {2 } 1 = exp µ0t exp t0(AA0)t { } {2 } 1 = exp µ0t + t0Σt . { 2 } 4. MULTIVARIATE NORMAL DISTRIBUTION (Part I) 3 Proof of Def 2 Def 3: ⇒ Since Σ 0 (and Σ = Σ0), there exists an orthogonal matrix, Tn×n, such that T0ΣT = Λ≥, where Λ is diagonal with non-negative elements. Therefore, Σ = TΛT0 = TΛ1/2Λ1/2T0 = (TΛ1/2)(TΛ1/2)0 = AA0. In other words, let A = TΛ1/2. Now, in the previous proof we showed the m.g.f. of AZ + µ is 1 exp µ0t + t0Σt , { 2 } the same as Y. Because the m.g.f. uniquely determines the distribution (when the m.g.f. exists in a neighbourhood of t = 0), Y has the same distribution as AZ + µ. 4 4. MULTIVARIATE NORMAL DISTRIBUTION (Part I) Proof of Def 3 Def 1: (for p.d. Σ). ⇒ 0 Because Σ is positive definite, there is a non-singular An×n such that AA = Σ (lecture notes # 2, page 10). Let Y = AZ + µ, where Z =(Z1,...,Zn) is a random sample from N(0, 1). The density of Z is n −1/2 1 2 −n/2 1 0 fZ(z)= (2π) exp Z = (2π) exp Z Z . {−2 i } {−2 } Yi=1 The density function of Y is fY(y)= fZ(z(y)) J , | | where J is the Jacobian ∂Z J = i = A−1 = A −1, ∂Y | | | | j because Z = A−1(Y µ). Therefore, − −1 −1 fY(y)= fZ(A (y µ)) A − | | 1 = (2π)−n/2 A −1 exp [A−1(y µ)]0[A−1(y µ)] | | {−2 − − } 1 = (2π)−n/2 AA0 −1/2 exp (y µ)0(AA0)−1(y µ) | | {−2 − − } 1 = (2π)−n/2 Σ −1/2 exp (y µ)0Σ−1(y µ) | | {−2 − − } − − 1 − 1 − 1 0 − 1 0 − 1 0 − 1 (Using: A 1 = A 2 A 2 = A 2 A 2 =( A A ) 2 = AA 2 ) | | | | | | | | | | | || | | | Proof of Def 1 Def 2 (for p.d. Σ): Exercise: Use pdf in Def 1 and solve directly for mgf. ⇒ 4. MULTIVARIATE NORMAL DISTRIBUTION (Part I) 5 4.2 Properties of the Multivariate Normal Distribution 1. E[Y]= µ, cov(Y)= Σ (verify using Definition 3 and properties of means and covariances of random vectors) 2. If Z =(Z1,...,Zn) is a random sample from N(0, 1) then Z has the Nn(0n, In×n) distribution (use Definition 3). 3. If Σ is not p.d. then Y has a singular MVN distribution and no density function exists. 0 Example: A singular MVN distribution. Let Z = (Z1, Z2) N2(0, I), and let A be 1 1 ∼ 2 2 the linear transformation matrix A = 1 − 1 . 2 2 0 − Let Y =(Y1, Y2) be the linear transformation (Z Z )/2 Y = AZ = 1 − 2 . (Z2 Z1)/2 − By Definition 3, Y N(0, Σ), where Σ = AA0. ∼ 1 1 1 1 1 1 0 Σ = AA = 2 − 2 2 − 2 = 2 − 2 1 1 1 1 1 1 − 2 2 − 2 2 − 2 2 1 1 corr= − . Makes sense! 1 1 − 6 4. MULTIVARIATE NORMAL DISTRIBUTION (Part I) 4.3 Linear Transformations of MVN Vectors 0 1. If Y N (µ, Σ) and C × is a matrix of rank p, then CY N (Cµ, CΣC ). ∼ n p n ∼ p Proof: By Def 3, Y = AZ + µ, where AA0 = Σ. Then CY = C(AZ + µ) = CAZ + Cµ N(Cµ, CA(CA)0) (by Def 3) ∼ = N(Cµ, C(AA0)C) = N(Cµ, CΣC0). 2. Y is MVN if and only if a0Y is normally distributed for all non-zero vectors a. Proof: If Y N (µ, Σ) then a0Y N(a0µ, a0Σa) by 4.3.1 (above). ∼ n ∼ Conversely, assume that X = a0Y is univariate normal for all non-zero a. In other words, X N(a0µ, a0Σa), where µ = E[Y] and Σ = cov(Y). Using the form of the m.g.f.∼ of a univariate normal random variable, the m.g.f. of X is 1 E[exp(Xt)] = M (t) = exp (a0µ)t + (a0Σa)t2 X { 2 } for all t. Setting t =1 in MX (t) gives MY: 0 0 1 0 E[exp(a Y)] = M (t = 1) = exp (a µ)+ (a Σa) = MY(a), X { 2 } which is the m.g.f. of N (µ, Σ). Therefore, Y N (µ, Σ) by Def 2. n ∼ n In words, a random vector is MVN iff every linear combination of its random variable components is a normal random variable..

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    6 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us