VOL. 12, NO. 12, JUNE 2017 ISSN 1819-6608 ARPN Journal of Engineering and Applied Sciences ©2006-2017 Asian Research Publishing Network (ARPN). All rights reserved.
www.arpnjournals.com
ON GENERALIZED VARIANCE OF NORMAL-POISSON MODEL AND POISSON VARIANCE ESTIMATION UNDER GAUSSIANITY
Khoirin Nisa1, Célestin C. Kokonendji2, Asep Saefuddin3, AjiHamim Wigena3 and I WayanMangku4 1Department of Mathematics, University of Lampung, Bandar Lampung, Indonesia 2Laboratoire de Mathématiques de Besançon, Université Bourgogne Franche-Comté, France 3Department of Statistics, Bogor Agricultural University, Bogor, Indonesia 4Department of Mathematics, Bogor Agricultural University, Bogor, Indonesia E-Mail: [email protected]
ABSTRACT As an alternative to full Gaussianity, multivariate normal-Poisson model has been recently introduced. The model is composed by a univariate Poisson variable, and the remaining random variables given the Poisson one are real independent Gaussian variables with the same variance equal to the Poisson component. Under the statistical aspect of the generalized variance of normal-Poisson model, the parameter of the unobserved Poisson variable is estimated through a standardized generalized variance of the observations from the normal components. The proposed estimation is successfully evaluated through simulation study.
Keywords: covariance matrix, determinant, exponential family, generalized variance, infinitely divisible measure.
INTRODUCTION scatter. The uses of generalized variance have been Normal-Poisson model is a special case of normal discussed by several authors. In sampling theory, it can be stable Tweedie (NST) models which were introduced by used as a loss function on multiparametric sampling Boubacar Maïnassara and Kokonendji [1] as the extension allocation [9]. In the theory of statistical hypothesis of normal gamma [2] and normal inverse Gaussian [3] testing, generalized variance is used as a criterion for an models. The NST family is composed by distributions of unbiased critical region to have the maximum Gaussian T random vector X= (X1, …, Xk) where Xj is a univariate curvature [10]. In the descriptive statistics, Goodman [11] (non-negative) stable Tweedie variable and (X1, …, Xj-1, proposed a classification of some groups according to their T Xj+1,…,Xk) =: given Xj are k-1 real independent Gaussian generalized variances. In the last two decades the variables with 푐variance X j k}. generalized variance has been extended for non-normal i, for any fixed {1, 2, …, Several particular퐗 cases have already appeared in different distributions in particular for natural exponential families contexts; one can refer to [1] and references therein. (NEFs) [12] [13]. Normal-Poisson is the only NST model which Three generalize variance estimators of normal- has a discrete component and it is correlated to the Poisson models have been introduced (see [14]). Also, the continuous normal parts. Similar to all NST models, this characterization by variance function and by generalized model was introduced in [1] for a particular case of j that variance of normal-Poisson have been successfully proven is j=1. For a normal-Poisson random vector X as described (see [15]). In this paper, a new statistical aspect of normal Poisson model is presented, i.e. the Poisson variance above, Xjis a univariate Poisson variable. In literatures, there is a model called "Poisson Gaussian" [4] [5] which is estimation under only observations of normal components also composed by Poisson and normal distributions. leading to an extension of generalized variance term i.e. However, normal-Poisson and Poisson Gaussian are two the "standardized generalized variance". completely different models. Indeed, for any value of j NORMAL POISSON MODELS {1, 2, …, k}, a normal-Poissonjmodel has only one Poisson component and k-1 Gaussian components, while a The family of multivariate normal-Poissonj Poisson-Gaussianj model has j Poisson components and k-j models for allj {1, 2, …, k}and fixed positive integer k>1 Gaussian components which are all independent. Normal- is defined as follows: T Poisson is also different from the purely discrete Poisson- Definition 1. ForX = (X1, …,Xk) a k-dimensional normal model of Steyn [6] which can be defined as a normal-Poisson random vector, it must hold that multiple mixture of k independent Poisson distributions X with parameters m1, m2 mk and those parameters have 1. j is a univariate Poisson random variable, and , … , T a multivariate normal distribution. Hence, the multivariate 2. :=(X1, …, Xj-1, Xj+1,…,Xk) given Xjfollows the Poisson-normal distribution is a multivariate version of the 푐-variate normalNk-1(0,XjIk-1) distribution, whereIk-1= X 푘 − Hermite distribution [7]. diagk-1(1, …, 1) denotes the (k-1) (k-1)unit matrix. Generalized variance (i.e. the determinant of covariance matrix expressed in term of mean vector) has In order to satisfy the second× condition we need important roles in statistical analysis of multivariate data. Xj>0. But in practice it is possible to have Xj=0 in the It was introduced by Wilks [8] as a scalar measure of Poisson component. In this case, the corresponding normal multivariate dispersion and used for overall multivariate components are degenerated as the Dirac mass 0which
3836 VOL. 12, NO. 12, JUNE 2017 ISSN 1819-6608 ARPN Journal of Engineering and Applied Sciences ©2006-2017 Asian Research Publishing Network (ARPN). All rights reserved.
www.arpnjournals.com makes their values become 0s. We have shown that zero values in Xj do not affect the estimation of the generalized variance of normal-Poisson [16]. From Definition 1, for a fixed power of convolution t>0 and givenj {1, 2, …,k}, denote Ft;j=F(t;j) the multivariate NEF of normal-Poisson with t* t:= , the NEF of a k-dimensional normal-Poisson random vector X is generated by
Indeed, for the covariance matrix above one can use the Schur complement [18] of a matrix block to obtain the following representation of determinant
where 1A is the indicator function of the set A. Since t>0 then is known to be an infinitely divisible t;j measure; see, e.g., Sato [17]. T with the non-null scalar = 1, the vector a =(2, The cumulant function of normal-Poisson is -1 T obtained from the logarithm of the Laplace transform …,k) and the (k-1) (k-1) matrixA= aa + 1Ik-1, where Ij= =diag(1, …, 1) is the j j unit matrix. Consequently, the oft;j, i.e. =log ) and the determinant of the covariance × matrix for j = 1 is T probability distribution of normal-Poissonj which is a × K 푡; θ 푅 exp 휃 x 푡; 퐱 member of NEF is given by∫ det with 퐕퐹푡;1 흁 = 흁 ∈ 퐹푡;1 P ) = exp ) Then, it is trivial to show that for j {1, 2, …,k} T the generalized variance of normal-Poissonj model is θ; t; dx {θ x − K t;j θ } t; dx The mean vector and the covariance matrix of Ft;j given by can be calculated using the first and the second derivatives of the cumulant function, i.e.: (5) 퐹푡;1 퐹푡; det 퐕 Equation흁 = with (5) expresses흁 ∈ the generalized variance of normal-Poisson model depends only on the mean of the 푡; and = K′ 휃 Poisson component and the dimension space k>1.
CHARACTERIZATIONS AND GENERALIZED ′′ VARIANCE ESTIMATIONS 퐹푡; 푡; V =For K practical(휃 ). calculation we need to use the Among NST models, normal-Poisson and following mean parameterization: normal-gamma are the only models which are already characterized by generalized variance (see [19] for characterization of normal-gamma by generalized P variance). In this section we present the characterizations of normal-Poisson by variance function and by generalized (흁; 푡; ) ≔ 푃(휽 흁 ; 푡; ), where is the solution in of the equation variance, then we present three estimations of generalized Then for a fixed j {1, 2, …, k}, the variance by maximum likelihood (ML), uniformly ′ 휃 휽 variance푡; function (i.e the variance-covariance matrix in minimum variance unbiased (UMVU) and Bayesian term = Kof mean 휃 .parameterization) is given by methods.
(2) Characterization T The characterizations of normal-Poisson models 퐹푡; V = + diag , … , , , , … , are stated in the following theorems without proof. on its support
k Theorem 1 = {µR ; µj> 0 and µlR for l ≠j}. (3) Let k{2,3, …} and t>0. If an NEFFt;j satisfies 푡; (2) for a given j{1,2, …,k}, then up to affinity, Ft;jis a 퐹 For j = 1, the covariance matrix of X can be normal-Poissonj model. expressed as follows:
3837 VOL. 12, NO. 12, JUNE 2017 ISSN 1819-6608 ARPN Journal of Engineering and Applied Sciences ©2006-2017 Asian Research Publishing Network (ARPN). All rights reserved.
www.arpnjournals.com
Theorem 2 Generalized variance estimators Let Ft;j=F( ) be an infinitely divisible NEF on Let X1, …,Xn be random vectors i.i.d. with k R (k>1) such that distribution P(;Ft;j) in a normal-Poissonj model 푡; Ft;j=F(t;j) for fixed j{1,2, …,k}. Denoting 1) and the sample mean. ̅ 푋1+⋯+푋푛 퐗 = 2) T 푡; 횯( ′′) = 푅 T 푐 푛 = X̅ ,…,X̅ 푡; a) Maximum likelihood estimator det 휽 =T 푡 exp 푘 × 휽 휽̃ T for =(1, …. k) and =(1,…, j-1, 1, j+1, …, k) . The ML generalized variance estimator of normal Then F is of normal-Poisson푐 type. Poisson model is given by t,j 휽 All technical details휽̃ of proofs can be seen in [15]. 퐹푡; In fact, the proof of Theorem 1 is established by analytical det 퐕 흁 .= (6) calculations and using the well-known properties of NEFs described in Proposition 3 below. 푛,푡; 퐹푡; ̅ ̅ 푇 = detThe 퐕 ML 퐗 estimator = (6) is directly obtained from
(5) by substituting µj with its ML estimator . For all p≥1, Proposition 3 Tn,t;j is a biased estimator of with a given Let and be two -finite positive measures on ̅ k quadratic risk with tedious calculation of explicit R such that F=F()$, and µMF. det 퐕퐹푡; 흁 ̃ expression or infinite. (i) If there exists ̃ = (d,̃ c) such that = T b) Uniformly minimum variance unbiased estimator exp (d x+c)(dx), then and The UMVU generalized variance estimator of ; for , 푅 × 푅 ̃ x ̃ ̃ and = ̃:Θ = Θ − d K 휃 = normal Poisson models is given by 퐹 K 휃 + d + ̃= ϵM 퐹푡; 퐹̃ 퐹 퐹̃ 퐹 det 퐕 흁 = V ̃ =(ii)If V det V ̃ with= det V . , then: if (7) and ; for − + n,t; ̅ ̅ ̅ ̅ T ̃= 휑 ∗ T휑 x =TAx + b U = nThe UMVUX (nX −estimator ) … nX of − k + , 푛 is deduced≥ 푘 ̃ Θ ̃ = A Θ K 휃 = K A 휃 + b 휃 퐹 by using intrinsic moment formula of 퐹 univarite푡; Poisson and ̃= A + b ϵ 휑M− , T distribution as follows: det 퐕 흁 퐹̃ 퐹 V ̃ = AV (휑 ̃ )A . ퟐ . 퐹̃ 퐹 det 퐕 흁̃(iii) = If det 퐀 isdet the 퐕 t-th 흁 convolution power of for t>0, then, for ∗ [ ( −Indeed, ) … ( letting− 푘 + )] = gives the result that (7) is ̃= the UMVU estimator of (5). Because, by the completness 퐹 흁̃= 푡흁ϵand 푡 det, of NEF, the unbiased estimator = 푛 ̅is unique. − 퐹̃ 퐹 퐹̃ 퐹 퐕 흁̃ =The 푡퐕 proof푡 흁̃ of Theorem퐕 흁̃ 2 is = obtained 푡 det 퐕 by 흁 using the c) Bayesian estimator infinite divisibility property of normal-Poisson, also Under assumption of prior gamma distribution of applying two properties of determinant and affine µjwith parameter >0 and >0, the Bayesian estimator of polynomial. The infinite divisibility property used in is given by theproof is provided in Proposition 4below. det 퐕퐹푡; 흁 = Proposition 4 (8) k If is an infinitely divisible measure on R , then +푛푋̅ there exist a symmetric non-negative definite d d matrix 퐵푛,푡; , , = +푛 . k To show this, let Xj1,…,Xjn given µj be Poisson with rank r k and a positive measure on R such that × distribution with mean µj, then the probability mass ). function is given by See, e.g.[20, page 342].T T " 훉 = Σ + ∫푅 퐱퐱 exp 휽 퐱 퐱 The above expression of is an equivalent 푥 of the Lévy-Khinchine formula [17]; thus, comes from a Brownian part and " 휽 the rest 푝(풙 | ) = exp − ∀ 푥 휖 푁 Assuming푥 ! thatµ follows gamma(,), then the )corresponds to jumps j prior probability distribution function ofµjis written as T T part of the associated Lévy process through the Lévy 푅 measure " 휽 = .∫ 퐱퐱 exp 휽 퐱 퐱