View metadata, citation and similar papers at core.ac.uk brought to you by CORE

provided by Elsevier - Publisher Connector

JOURNAL OF MULTIVARIATE ANALYSIS 2, 339-344 (1972)

Some Characterizations of the Multivariate t Distribution*

PI-ERH LIN

Department of , Florida State University, Tallahassee, Florida 32306

Communicated by E. Lukacs

A multivariate t vector X is represented in two different forms, one associated with a normal vector and an independent chi-squared variable, and the other with a normal vector and an independent Wishart . We show that X is multivariate t with mean p, V(V - 2)-lC, v > 2 and degrees of freedom v if and only if for any a # 0, (a’Za)-‘1” a’(X - p,) has the Student’s t distribution with Y degrees of freedom under both representations. Some other characterizations are also obtained.

1. INTRODUCTION AND SUMMARY

The multivariate t distribution was first derived independently by Cornish [4] and Dunnett and Sobel [8]. It arises in multiple decision problems concerned with the selection and ranking of population means of several normal populations having a common unknown (see Bechhofer, Dunnett and Sobel [3]). It can be used in setting up simultaneous confidence bounds for the means of correlated normal variables, for parameters in a linear model and for future observations from a multivariate (see John [ll]). The multivariate t distribution also appears in the Bayesian multivariate analysis of variance and regression, treated by Tiao and Zellner [la], Geisser and Cornfield [9], Raiffa and Schlaifer [13], and Ando and Kaufmann [2], where the normal- Wishart distribution is considered to be the distribution (in the sense of Raiffa and Schlaifer) of the mean vector and covariance matrix

Received March 4, 1971; revised March 8, 1972. AMS 1970 subject classification: Primary 62H05; Secondary 62HlO. Key words: Symmetric square root of a positive ; conditionally inde- pendent; conditional marginal probability density; almost surely with respect to a o-finite measure. * Research supported in part by the Florida State University Grant No. 011204-043 and in part by the National Institute of General Medical Sciences through Training Grant 5Tl GM-913. 339 Copyright0 1972by AcademicPress, Inc. All rights of reproductionin any form reserved. 340 LIN of a multivariate normal distribution. However, the representation of the multi- variate t vector used in the Bayesian multivariate analysis is somewhat different from that employed by Cornish [4], and Dunnett and Sobel [S]. The multivariate t can be considered as a generalization of the univariate t. Siddiqui [14] generalizes the univariate t to a bivariate t in a different direction and Dickey [7] generalizes the multivariate t to a matricvariate t. The probability integrals and the percentage points of a multivariate t have been studied by Gupta [lo] and Krishnaiah and Armitage [12]. In Section 2 of this note we consider the distribution of a multivariate t vector which may be represented in two different forms, one associated with a normal vector and an independent chi-squared variable, and the other with a normal vector and an independent Wishart matrix. In Section 3, the characterization properties of a t vector are developed through multivariate normal theory. As usual, we denote by N( p, Z) a multivariate normal distribution with mean IL and covariance matrix 2, by W(Z, n) a Wishart distribution with parametersn and c by xy2 a chi-squared distribution with v degreesof freedom; and by Z$JT~) a noncentral F distribution with p and v degreesof freedom and non- centrality ?. We write F2),+ forFL,,(O).

2. THE MULTIVARIATE t DISTRIBUTION

A p-variate random vector X = (X1 ,..., X,) ’ is said to have a (nonsingular) multivariate t distribution with mean vector p = (pi ,..., pP)’ and covariance matrix V(V - 2)-l& v > 2, denoted by T,(p, Z, p), if it has the probability density function (pdf) given by

m& + PI1 (x - p)‘F(x - p) +(“+p) v > 2 (l) fix) = (7w)f’qQu) 1.z p 1l + v I ,

It is noted that T,(O, 1, 1) = t, is the Student’s t distribution with v degreesof freedom. Unlessotherwise mentioned, it is assumedthat X, TV,a, and 0 arep x 1 vectors, V and Z are p x p positive definite symmetric matrices, and I is the of order p. For the purpose of establishing the characterizations and the distribution theory in Section 3 it is convenient to represent a multivariate t vector in the following forms: Representation A. Let X N T,( TV,,Z, p). Then X may be written as

x = s-1Y + p, (2) CHARACTERIZATIONS OF MULTIVARIATE t 341 where Y N N(0, 2) and v2.P- xV2, independent of Y. This implies that X \ S2 = s2- N(p, s-~Z), where vS2 - ~,a. The Student’s t is then represented as X 1S2 = ss- N(0, sw2),where vS2 - xy2. Representation B. Let X - T,( Jo,2, p). Then X may be written as

x = (P’2)-1Y + p, (3) where Vi/2 is the symmetric squareroot of V, i.e.,

f?v2p/2 zzz I' - W(Fl, v + p - 1) and Y - N(0, VI), independent of V. This implies that X j Y - N(y, ~l”-l), where V - W(Z-l, v + p - 1). The Student’s t random variable is then represented as X 1 V N N(0, va’V-la/a’Za) for any a # 0, where

v w W(Z-1, v + p - 1).

It can be shown that the pdf of X, under both RepresentationsA and B, is given by (1).

3. CHARACTERIZATIONS OF THE MULTIVARIATE t

Cornish [6] gives a characterization of the multivariate t using Representa- tion A. His proof is lengthy and complicated. We present here the sameresult under RepresentationB and also give a simpler proof under RepresentationA, utilizing multivariate normal distribution theory. Two other characterizations are also obtained under RepresentationA.

THEOREM 1. X - T,(p, 2, p) if and only ifi for any a # 0, (a’Za)-1/2a’(X - p) - t, . Proof. (i) Assume RepresentationA. x 1 s2 = s2- N( P, s-~E), -a (a’L’a)-l’2a‘(X - p) ( S2 = s2- N(0, s-3, for any a # 0 0 (a’Za)-1/2a’(X - p) - t, . (ii) Assume RepresentationB. X 1 V - N(p, vV-l) t> (a’.Z’a)-1/2a’(X - Jo) 1 V N NIO, v(a’V-‘a)/(a’Ca)], for any a f 0 0 (a’2a)-1’2a’(X - p) - t, , 342 LIN since a%a/a’ V-la N xy2. (See, for example, Stein [15, pp. 32-331.) This completesthe proof of the theorem. In order to establish some further characterizations, we need the following lemmas.The proofs are straightforward and are omitted.

LEMMA 1. Let X N T,(p, Z, p), then for any nonsa’ngularp x p scalar matrix C and any a, CX + a - TV(Cp + a, CZC’, p).

LEMMA 2. Let X - T,( p, z: p), then X’zl--lx/p - F;&~‘Z-~lr/p). When p. = 0, the distribution is central F.

LEMMA 3. Let Y be a symmetrically distributed random variable. Then Y2 N xl2 if, and only if Y - N(0, 1).

LEMMA 4. Let YI ,..., Y, beindependent random variables with$nite . Assumethat Yl, , (k = I,..., p), has pdf hk(y) such that hk(y) > 0 and d@eren- tiable for ally, -cc < y < 00. Then the joint pdf of Yl ,..., Y, is a function on& ofy12 + .a* + ys2 if, and only if(YI ,..., Y,)’ - N(0, a21),for somea2 < co. For the remainder of this section, we will consider a multivariate t vector using RepresentationA.

THEOREM 2. Let vS2 - xy2 and let XI ,..., X, be continuousrandom variables. Assumethat XI ,..., X, are conditionally independent,and that X, is symmetrically distributedwith meanpk andJinite variance, given S2 = s2. Then

(4) k=l if and only sf (XI ,..., X,,)’ - T,(p, D, p), where p = (r-l1,..., pp)’ and D is a p x p diagonalmatrix with its k-th diagonalelement equal to uk2. Proof. The sufficiency is a special case of Lemma 2. We shall show the necessity. Let Yk = (xk - &)/ok , k = l,...,p. Then the Y’s also satisfy the conditions of the theorem with mean 0 and finite variance, given S2 = s2.It is well known that an F,,,-distributed random variable may be represented as the ratio of two independent chi-squared random variables with p and v degreesof freedom, repectively, multiplied by a suitable constant. Thus, letting S12- xs2 independent of S2, we have

k=l CHARACTERIZATIONS OF MULTIVARIATE t 343 or ?J c (SYfJ2 1 s2 - S12 1 s2 - x9 23 (6) k=l which is independent of Ss. This implies that

(syk)z 1 s2 -xl’, k = 1,...,p, (7) and by Lemma 3, we have SY j S2 = s2N N(0, I), where Y = (Y, ,..., Y,)‘. Thus, by RepresentationA, Y N T,(O, I, p) and by Lemma 1, X N Ty( p, D, p).

THEOREM 3. Let vS2 N xy2 and let X, , k = 1 ,..., p, be conditionaZZy inde- pendent random variables, given S2 = s2, with Var(X, 1s2) = u2/s2,u2 < 00. Assume that X, (k = l,..., p), given S2 = s2, has conditional pdff,(x I s2)which is positive and differentiable for all x, --00 < x < 03. Then the joint pdf of X 1 ,..., X, is a function on& of xl2 + .** + x,2 if, and only if (X, ,..., X,)’ N T&J ~“4 P). Proof. The sufficiency is easily seenfrom the joint pdf of XI ,.,., X, . We shall prove the necessity. Since the joint pdf of XI ,..., X, is a function of Xl2 + *.. + x 92, it follows that the conditional joint pdf of XI ,..., X, , given S2 = s2,is a function of xl2 + 0.. + xD2 and s2 almost surely with respect to a u-finite measurem, i.e.,

jjfkt’k I s2> = g0 (2 xk29 s2)* aS- (m>. (8) k-l

Similarly, the conditional marginal pdf of X, (k = l,...,p), given 5’s = s2, is

f7cbc I q = g!&.2, s2), a.s. (m), (9) where gk(t, s2) (k = O,..., p) is a positive and differentiable function of t 3 0. Now, the result of Lemma 4 and the fact that Var(Xk ) s2) = a2/s2,for all k = l,..., p, show that (X, ,..., X,)’ 1 S2 = s2N N(0, s-~u~I). Therefore, (Xl ,--*>X,)’ - Tv(O, u24 P).

[l] ANIXRSON, T. W. (1958). An Introduction to Multi~ariate Sta&tical Analysis. Wiley, New York. [2] ANDO, A. AND KAUFMANN, G. W. (1965). Bayesian Analysis of the independent multinormal process-neither mean nor known. J. Amer. Statist. Assoc. 60 347-358. [3] BECHHOFER, R. E., DUNNETT, C. W., AND SOBEL, M. (1954). A two-sample multiple decision procedure for ranking means of normal populations with a common unknown variance. Biometrika 41 170-176. [4] CORNISH, E. A. (1954a). The multivariate t-distribution associated with a set of normal sample deviates. Austral. /. Phys. 7 531-542. [5] CORNISH, E. A. (1954b). The sampling distributions of statistics derived from the multivariate t-distribution. Austral. J. Pkys. 8 193-I 99. [6] COFWISH, E. A. (1962). The multivariate t-distribution associated with the general multivariate normal distribution. CSIRO Tech. paper No. 13. CSIRO Div. Math. Stat., Adelaide. [7] DICKEY, J. M. (1967). Matricvariate generalizations of the multivariate t distribution and theinverted multivariate t distribution. Ann. Math. Stat. 38 511-518. [8] DUNNETT, C. W. AND SOBEL, M. (1954). A bivariate generalization of Student t-distribution with tables for certain special cases. Biometriku 41 153-169. [9] GEISSER, S. AND CORNFIELD, J. (1963). Posterior distributions for multivariate normal parameters. J. Roy. Statist. Sot. Ser. B 25 368-376. [lo] GUPTA, S. S. (1963). Probability integrals of multivariate normal and multivariate t. Ann. Math. Stat. 34 792-828. [ll] JOHN, S. (1961). On the evaluation of the probability integral of the multivariate t-distribution. Biometrika 48 409-417. [12] KRISHNAIAH, P. R. AND ARMITAGE, J.V. (1965). Percentage points of the multivariate t distribution, ARL Tech. Rep. No. 65-199. Aerospace Res. Labs. Wright-Patterson AFB, Ohio. [13] RAIFFA, H. AND SCHLAIFER, R. (1961). Applied Statistical Decision Theory. Division of Research, Harvard Business School. [14] SIDDIQUI, M. M. (1967). A bivariate t distribution. Ann. Math. Stat. 38 162-166. [15] STEIN, C. M. (1969). Mdtiwariate Analysis 1. (Notes prepared by M. L. Eaton). Tech. Rep. No. 42, Dept. of Statistics, Stanford University. [16] TIAO, G. C. AND ZELLNER, A. (1964). On the Bayesian estimation of multivariate regression. J. Roy. Statist. Sot. Ser. B 26 277-285.