Gaussian Process Structural Equation Models with Latent Variables

Gaussian Process Structural Equation Models with Latent Variables

Gaussian Process Structural Equation Models with Latent Variables Ricardo Silva Robert B. Gramacy Department of Statistical Science Statistical Laboratory University College London University of Cambridge [email protected] [email protected] Abstract tent space (with confidence regions) for ranking, clustering and visualization; density estimation; missing data imputa- In a variety of disciplines such as social sci- tion; and causal inference(Pearl, 2000;Spirtes et al., 2000). ences, psychology, medicine and economics, the This paper introduces a nonparametric formulation of recorded data are considered to be noisy mea- SEMs with hidden nodes, where functions connecting la- surements of latent variables connected by some tent variables are given a Gaussian process prior. An effi- causal structure. This corresponds to a fam- cient but flexible sparse formulation is adopted. To the best ily of graphical models known as the structural of our knowledge, our contribution is the first full Gaussian equation model with latent variables. While process treatment of SEMs with latent variables. linear non-Gaussian variants have been well- studied, inference in nonparametric structural We assume that the model graphical structure is given. equation models is still underdeveloped. We in- Structural model selection with latent variables is a com- troduce a sparse Gaussian process parameteriza- plex topic which we will not pursue here: a detailed dis- tion that defines a non-linear structure connect- cussion of model selection is left as future work. As- ing latent variables, unlike common formulations parouhov and Muth´en (2009) and Silva et al. (2006) discuss of Gaussian process latent variable models. The relevant issues. Our goal is to be able to generate poste- sparse parameterization is given a full Bayesian rior distributions over parameters and latent variables with treatment without compromising Markov chain scalable sampling procedures with good mixing properties, Monte Carlo efficiency. We compare the stabil- while being competitive against non-sparse Gaussian pro- ity of the sampling procedure and the predictive cess models. ability of the model against the current practice. In Section 2, we specify the likelihood function for our structural equation models and its implications. In Sec- tion 3, we elaborate on priors, Bayesian learning, and a 1 CONTRIBUTION sparse variation of the basic model which is able to handle larger datasets. Section 4 describes a Markov chain Monte A cornerstone principle of many disciplines is that obser- Carlo (MCMC) procedure. Section 5 evaluates the useful- vations are noisy measurements of hidden variables of in- ness of the model and the stability of the sampler in a set of terest. This is particularly prominent in fields such as so- real-world SEM applications with comparisons to modern cial sciences, psychology, marketing and medicine. For alternatives. Finally, in Section 6 we discuss related work. instance, data can come in the form of social and eco- nomical indicators, answers to questionnaires in a medical exam or marketing survey, and instrument readings such as 2 THE MODEL: LIKELIHOOD fMRI scans. Such indicators are treated as measures of la- tent factors such as the latent ability levels of a subject in Let G be a given directed acyclic graph (DAG). For sim- a psychological study, or the abstract level of democrati- plicity, in this paper we assume that no observed variable zation of a country. The literature on structural equation is a parent in G of any latent variable. Many SEM appli- models (SEMs) (Bartholomew et al., 2008; Bollen, 1989) cations are of this type (Bollen, 1989; Silva et al., 2006), approaches such problems with directed graphical models, and this will simplify our presentation. Likewise, we will where each node in the graph is a noisy function of its par- treat models for continuous variables only. Although cyclic ents. The goals of the analysis include typical applications SEMs are also well-defined for the linear case (Bollen, of latent variable models, such as projecting points in a la- 1989), non-linear cyclic models are not trivial to define and IL PDL ζ IL ζ Y1 YYYYYY2 3 4 5 6 7 Y1 YYYYYY2 3 4 5 6 7 ε ε 1 ε2 ε 3 ε 4 ε 5 ε 6 ε7 1 ε2 ε 3 ε 4 ε 5 ε 6 ε7 (a) (b) Figure 1: (a) An example adapted from Palomo et al. (2007): latent variable IL corresponds to a scalar labeled as the industrialization level of a country. PDL is the corresponding political democratization level. Variables Y1, Y2, Y3 are indicators of industrialization (e.g., gross national product) while Y4,...,Y7 are indicators of democratization (e.g., expert assessements of freedom of press). Each variable is a function of its parents with a corresponding additive error term: i for each Yi, and ζ for democratization levels. For instance, PDL = f(IL)+ζ for some function f(·). (b) Dependence among latent variables is essential to obtain sparsity in the measurement structure. Here we depict how the graphical dependence structure would look like if we regressed the observed variables on the independent latent variables of (a). as such we will exclude them from this paper. much has been done on exploring nonparametric models with dependent latent structure (a loosely related excep- Let X be our set of latent variables and X ∈ X be a partic- i tion being dynamic systems, where filtering is the typical ular latent variable. Let X be the set of parents of X in Pi i application). Figure 1(b) illustrates how modeling can be G. The latent structure in our SEM is given by the follow- affected by discarding the structure among latents2. ing generative model: if the parent set of Xi is not empty, X Xi = fi( Pi )+ ζi, where ζi ∼ N (0, vζi ) (1) 2.1 Identifiability Conditions Latent variable models might be unidentifiable. In the con- N (m, v) is the Gaussian distribution with mean m and text of Bayesian inference, this is less of a theoretical issue variance v. If X has no parents (i.e., it is an exogenous i than a computational one: unidentifiable models might lead latent variable, in SEM terminology), it is given a mixture to poor mixing in MCMC, as discussed in Section 5. More- of Gaussians marginal1. over, in many applications, the latent embedding of the data The measurement model, i.e., the model that describes the points is of interest itself, or the latent regression functions distribution of observations Y given latent variables X , is are relevant for causal inference purposes. In such appli- X as follows. For each Yj ∈ Y with parent set Pj , we have cations, an unidentifiable model is of limited interest. In this Section, we show how to derive sufficient conditions XT Yj = λj0 + Pj Λj + j , where j ∼ N (0, vj ) (2) for identifiability. Consider the case where a latent variable Xi has at least Error terms {j } are assumed to be mutually independent three unique indicators Yi ≡ {Yiα, Yiβ , Yiγ }, in the sense and independent of all latent variables in X . Moreover, Λj T that no element in Yi has any other parent in G but Xi. It is is a vector of linear coefficients Λj = [λj1 ... λ X ] . j| Pj | known that in this case (Bollen, 1989) the parameters of the Following SEM terminology, we say that Yj is an indicator structural equations for each element of Y are identifiable X i of the latent variables in Pj . (i.e., the linear coefficients and the error term variance) up An example is shown in Figure 1(a). Following the nota- to a scale and sign of the latent variable. This can be re- tion of Bollen (1989), squares represent observed variables solved by setting the linear structural equation of (say) Yiα and circles, latent variables. SEMs are graphical models to Yiα = Xi + iα. The distribution of the error terms with an emphasis on sparse models where: 1. latent vari- is then identifiable. The distribution of Xi follows from a ables are dependent according to a directed graph model; deconvolution between the observed distribution of an ele- 2. observed variables measure (i.e., are children of) very ment of Yi and the identified distribution of the error term. few latent variables. Although sparse latent variable mod- 2Another consequence of modeling latent dependencies is re- els have been the object of study in machine learning and ducing the number of parameters of the model: a SEM with a lin- statistics (e.g., Wood et al. (2006); Zou et al. (2006)), not ear measurement model can be seen as a type of module network (Segal et al., 2005) where the observed children of a particular 1 X For simplicity of presentation, in this paper we adopt a fi- latent Xi share the same nonlinearities propagated from Pi : in nite mixture of Gaussians marginal for the exogenous variables. the context of Figure 1, each indicator Yi ∈ {Y4,...,Y7} has a However, introducing a Dirichlet process mixture of Gaussians conditional expected value of λi0 + λi1f2(X1) for a given X1: marginal is conceptually straightforward. function f2(·) is shared among the indicators of X2. x Identifiability of the joint of X can be resolved by mul- For each Pi , the corresponding Gaussian process prior for tivariate deconvolution under extra assumptions. For in- f 1:N (1) (N) function values i ≡{fi ,...,fi } is stance, Masry (2003) describesone setup for the problemin f 1:N x1:N K the context of kernel density estimation (with known joint i | Pi ∼ N (0, i) distribution of error terms, but unknown joint of Y). where Ki is a N × N kernel matrix (Rasmussen and 1:N Assumptions for the identifiability of functions fi(·), given Williams, 2006), as determined by x .

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    9 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us