Gaussian, Gaussian-Gamma, Gaussian-Wishart, Multinomial

Gaussian, Gaussian-Gamma, Gaussian-Wishart, Multinomial

Exponential Families: Gaussian, Gaussian-Gamma, Gaussian-Wishart, Multinomial Yee Whye Teh Gatsby Computational Neuroscience Unit, University College London, 17 Queen Square, London WC1N 3AR, United Kingdom [email protected] August 13, 2007 1 Gaussian-Wishart This section derives the formulas for a Gaussian distribution along with a Gaussian-Wishart conjugate prior. 1 Data: n The number of data vectors. x The data vectors x1; : : : ; xn. Pn X = i=1 xi. Pn > C = i=1 xixi . Parameters: µ Mean of data. R Precision of data. Hyperparameters: d Dimensionality of data. r Relative precision of µ versus data. The precision of µ is rR. ν Degrees of freedom of precision of R. m Mean of µ is m. S Mean of R is νS−1. Prior: d −1 Y ν + 1 − i 1 Wishart: p(R) = 2−νd=2π−d(d−1)=4jSjν=2 Γ jRj(ν−d−1)=2 exp − Tr [RS] (1) 2 2 i=1 1 Gaussian: p(µ j R) = (2π)−d=2jrRj1=2 exp − Tr rR (µ − m)(µ − m)> (2) 2 1 1 p(µ, R) = jRj(ν−d)=2 exp − Tr R r(µ − m)(µ − m)> + S (3) Z(d; r; ν; S) 2 1 1 = jRj(ν−d)=2 exp − Tr R rµµ> − 2µ(rm)> + rmm> + S (4) Z(d; r; ν; S) 2 d (ν+1)d d(d+1)=4 −d=2 −ν=2 Y ν + 1 − i where Z(d; r; ν; S) = 2 2 π r jSj Γ (5) 2 i=1 Likelihood of data: " n !#! n=2 1 X p(x j µ, R) = (2π)−nd=2 jRj exp − Tr R (µ − x )(µ − x )> (6) 2 i i i=1 1 = (2π)−nd=2 jRjn=2 exp − Tr R nµµ> − 2Xµ> + C (7) 2 Joint likelihood: (2π)−nd=2 1 p(x; µ, R) = jRj(ν+n−d)=2 exp − Tr R (r + n)µµ> − 2µ(rm + X)> + rmm> + C + S (8) Z(d; r; ν; S) 2 > > rm+X rm+X (rm+X)(rm+X) > = ··· R (r + n) µ − r+n µ − r+n − r+n + rmm + C + S (9) 2 Posterior hyperparameters: r0 = r + n ν0 = ν + n rm + X m0 = (10) r + n S0 = S + C + rmm> − r0m0m0> Marginal probability: ν0+1−i 0 0 0 d=2 ν d Γ Z(d; r ; ν ;S ) r jSj 2 Y 2 p(x) = (2π)−nd=2 = π−nd=2 (11) ν0 ν+1−i Z(d; r; ν; S) 0d=2 0 2 Γ r jS j i=1 2 Predictive probability: p(y; x) p(yjx) = (12) p(x) Posterior: 1 0 1 p(µ, R j x) = jRj(ν −d)=2 exp − Tr R r0(µ − m0)(µ − m0)> + S0 (13) Z(d; r0; ν0;S0) 2 Computations in mixture models: We are interested in mixture models, in which each component is a Gaussian and the prior for parameters is the Gaussian-Wishart distribution. All mixture components share the same hyperparameters, but each component has its own set of parameters. In the posterior of the mixture model, the distribution for which data vectors belong to which components and the distribution over parameters are coupled and cannot be solved exactly. We have two choices: variational Bayes, or Markov chain Monte Carlo (MCMC) sampling. We develop our technique for MCMC sampling here. In particular we Gibbs sample indicator variables which determine the components to which data vectors belong to, while integrating out the parameters. The only computa- tions needed are marginal likelihood computations, and posterior computations for the parameters (which amounts to computing the posterior hyperparameters r0; ν0; m0 and S0). First consider computing the marginal likelihood. Note that to compute conditional likelihoods p(x j y) we just evaluate p(x; y)=p(y). There are two complications, computing jSj and jS0j efficiently, and computing the ratio of Gamma terms in (11). We deal with computing the determinants jSj and jS0j by representing S and S0 using their Cholesky decomposi- tions. In particular, updates to S and S0 will be carried out by directly updating their Cholesky decompositions. Given the Cholesky decomposition the determinant is just the product of the diagonal terms. We will come to updates to the Cholesky decomposition later. To efficiently compute the ratio of Gamma terms, we note that ν is the same for every component, and the ratio of Gamma terms can only taken on at most N + 1 values, one for each n = 0; 1;:::;N, where N is the total number of data vectors. By expanding the Gamma terms, we obtain: 8 n 0 Q 2 Qd ν−1−i d Γ ν +1−i > + j if n is even. Y 2 < j=1 i=1 2 ν+2−i ν+1−i = n−1 d d Γ( ) (14) Γ Q 2 Q ν−i Q 2 i=1 2 > j=1 i=1 2 + j i=1 ν+1−i if n is odd. : Γ( 2 ) We see that we can compute all (14) with O(dN) simple arithmetic operations, and only need to evaluate the Gamma function 2d times. We compute these once and store the results to be used in the future. In fact, we do not need to precompute (14) for all n. We can compute and store them as we need. Now consider updates to the posterior hyperparameters. In our setting we only need to worry about when a data vector is added to or removed from a component. Without loss of generality we assume r; ν; m and S are the 3 posterior hyperparameters for the component before adding/removing this new data vector into the component. The only complication is in updates to S0, since we are representing S and S0 using their Cholesky decompositions, say S = M >M; S0 = M 0>M 0 where M; M 0 are upper triangular matrices. Note that S0 differs from S only via three symmetric rank one matrices. Hence we can compute M 0 from M using three rank one Cholesky update, which takes O(d2) operations each. 4.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    4 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us