<<

MSc. Econ: MATHEMATICAL , 1995

THE MULTIVARIATE We say that the n 1 random vector x is normally distributed with a of E(x)=and a dispersion matrix of D(x) = if the density is n/2| |1/2 1 0 1 (17.22) N(x; , )=(2) exp 2 (x ) (x ) .

It is understood that x is non-degenerate with Rank() = n and ||6=0.To denote that x has this distribution, we can write x N(, ). We shall demonstrate two notable features of the normal distribution. The rst feature is that the conditional and marginal distributions associated with a normally distributed vector are also normal. The second is that any linear function of a normally distributed vector is itself normally distributed. We shall base our arguments on two fundamental facts. The rst is that

(17.23) If x N(, ) and if y = A(x b) where A is nonsingular, then y N{A( b),AA0}.

This may be illustrated by considering the case where b = 0. Then, according to the result in (17.8), y has the distribution

N(A1y; , )k∂x/∂yk =(2)n/2||1/2 exp 1 (A1y )01(A1y ) kA1k 2 n/2| 0|1/2 1 0 0 1 =(2) AA exp 2 (y A) (AA ) (y A) ; so, clearly, y N(A, AA0). The second of the fundamental facts is that

(17.25) If x N(, ) can be written in partitioned form as x 0 1 N 1 , 11 , x2 2 022

then x1 N(1, 11) and x2 N(2, 22) are independently distributed normal variates.

To see this, we need only consider the quadratic form

0 1 0 1 0 1 (x ) (x )=(x1 1)11 (x1 1)+(x2 2)22 (x2 2)

1 MSc. Econ: MATHEMATICAL STATISTICS, 1995 which arises in this particular case. On substituting in into the expression for N(x, , ) in (17.22) and using || = |11||22|,weget N(x;, )=(2)m/2| |1/2 exp 1 (x )0 1(x ) 11 2 1 1 11 1 1 (mn)/2| |1/2 1 0 1 (2) 22 exp 2 (x2 2) 22 (x2 2)

= N(x1; 1, 11)N(x2; 2, 22).

The latter can only be the product of the marginal distributions of x1 and x2, which proves that these vectors are independently distributed. The essential feature of the result is that

(17.26) If x1 and x2 are normally distributed with C(x1,x2) = 0, then they are mutually independent.

A zero covariances does not generally imply statistical independence. Even when x1, x2 are not independently distributed, their marginal dis- tributions are still formed in the same way from the appropriate components of and . This is entailed in the rst of our two main results which is that

(17.27) If x N(, ) is partitioned as x 1 N 1 , 11 12 , x2 2 21 22

then the marginal distribution of x1 is N(1, 11) and the condi- tional distribution of x2 given x1 is

| 1 1 N(x2 x1; 2 +2111 (x1 1), 22 2111 12).

Proof. Consider a non-singular transformation y I 0 x 1 = 1 y2 FI x2 such that C(y1,y2)=C(Fx1 +x2,x1)=FD(x1)+C(x2,x1) = 0. Writing this 1 condition as F 11 +21 = 0 gives F = 2111 . It follows that y1 1 E = 1 ; y2 2 2111 1

2 MSc. Econ: MATHEMATICAL STATISTICS, 1995 and, since D(y1)=11, C(y1,y2)=0and

D(y2)=D(Fx1 +x2) 0 0 = FD(x1)F +D(x2)+FC(x1,x2)+C(x2,x1)F 1 1 1 1 =2111 1111 12 +22 2111 12 2111 12 1 =22 2111 12, it also follows that y1 11 0 D = 1 . y2 022 2111 12

Therefore, according to (17.25), we can write the joint density function of y1, y2 as 1 1 N(y1; 1, 11)N(y2; 2 2111 1, 22 2111 12).

Integrating with respect to y2 gives the marginal distribution of x1 = y1 as N(x1; 1, 11). Now consider the inverse transformation x = x(y). The Jacobian of this transformation is unity. Therefore, to obtain an expression for N(x; , ), we 1 need only write y2 = x2 2111 x1 and y1 = x1 in the expression for the joint distribution of y1,, y2. This gives

N(x; , ) = N(x1; 1, 11) 1 1 1 N(x2 2111 x1; 2 2111 1, 22 2111 12), which is the product of the marginal distribution of x1 and the conditional | 1 1 distribution N(x2 x1; 2 2111 (x1 1), 22 2111 12)ofx2 given x1. | 1 The linear function E(x2 x1)=2 2111 (x1 1), which denes the expected value of x2 for given values of x1, is described as the regression of x2 1 on x1. The matrix 2111 is the matrix regression coecients.

3