Bayesian Networks and the Search for Causality
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
The Practice of Causal Inference in Cancer Epidemiology
Vol. 5. 303-31 1, April 1996 Cancer Epidemiology, Biomarkers & Prevention 303 Review The Practice of Causal Inference in Cancer Epidemiology Douglas L. Weedt and Lester S. Gorelic causes lung cancer (3) and to guide causal inference for occu- Preventive Oncology Branch ID. L. W.l and Comprehensive Minority pational and environmental diseases (4). From 1965 to 1995, Biomedical Program IL. S. 0.1. National Cancer Institute, Bethesda, Maryland many associations have been examined in terms of the central 20892 questions of causal inference. Causal inference is often practiced in review articles and editorials. There, epidemiologists (and others) summarize evi- Abstract dence and consider the issues of causality and public health Causal inference is an important link between the recommendations for specific exposure-cancer associations. practice of cancer epidemiology and effective cancer The purpose of this paper is to take a first step toward system- prevention. Although many papers and epidemiology atically reviewing the practice of causal inference in cancer textbooks have vigorously debated theoretical issues in epidemiology. Techniques used to assess causation and to make causal inference, almost no attention has been paid to the public health recommendations are summarized for two asso- issue of how causal inference is practiced. In this paper, ciations: alcohol and breast cancer, and vasectomy and prostate we review two series of review papers published between cancer. The alcohol and breast cancer association is timely, 1985 and 1994 to find answers to the following questions: controversial, and in the public eye (5). It involves a common which studies and prior review papers were cited, which exposure and a common cancer and has a large body of em- causal criteria were used, and what causal conclusions pirical evidence; over 50 studies and over a dozen reviews have and public health recommendations ensued. -
Statistics and Causal Inference (With Discussion)
Applied Statistics Lecture Notes Kosuke Imai Department of Politics Princeton University February 2, 2008 Making statistical inferences means to learn about what you do not observe, which is called parameters, from what you do observe, which is called data. We learn the basic principles of statistical inference from a perspective of causal inference, which is a popular goal of political science research. Namely, we study statistics by learning how to make causal inferences with statistical methods. 1 Statistical Framework of Causal Inference What do we exactly mean when we say “An event A causes another event B”? Whether explicitly or implicitly, this question is asked and answered all the time in political science research. The most commonly used statistical framework of causality is based on the notion of counterfactuals (see Holland, 1986). That is, we ask the question “What would have happened if an event A were absent (or existent)?” The following example illustrates the fact that some causal questions are more difficult to answer than others. Example 1 (Counterfactual and Causality) Interpret each of the following statements as a causal statement. 1. A politician voted for the education bill because she is a democrat. 2. A politician voted for the education bill because she is liberal. 3. A politician voted for the education bill because she is a woman. In this framework, therefore, the fundamental problem of causal inference is that the coun- terfactual outcomes cannot be observed, and yet any causal inference requires both factual and counterfactual outcomes. This idea is formalized below using the potential outcomes notation. -
Common Quandaries and Their Practical Solutions in Bayesian Network Modeling
Ecological Modelling 358 (2017) 1–9 Contents lists available at ScienceDirect Ecological Modelling journa l homepage: www.elsevier.com/locate/ecolmodel Common quandaries and their practical solutions in Bayesian network modeling Bruce G. Marcot U.S. Forest Service, Pacific Northwest Research Station, 620 S.W. Main Street, Suite 400, Portland, OR, USA a r t i c l e i n f o a b s t r a c t Article history: Use and popularity of Bayesian network (BN) modeling has greatly expanded in recent years, but many Received 21 December 2016 common problems remain. Here, I summarize key problems in BN model construction and interpretation, Received in revised form 12 May 2017 along with suggested practical solutions. Problems in BN model construction include parameterizing Accepted 13 May 2017 probability values, variable definition, complex network structures, latent and confounding variables, Available online 24 May 2017 outlier expert judgments, variable correlation, model peer review, tests of calibration and validation, model overfitting, and modeling wicked problems. Problems in BN model interpretation include objec- Keywords: tive creep, misconstruing variable influence, conflating correlation with causation, conflating proportion Bayesian networks and expectation with probability, and using expert opinion. Solutions are offered for each problem and Modeling problems researchers are urged to innovate and share further solutions. Modeling solutions Bias Published by Elsevier B.V. Machine learning Expert knowledge 1. Introduction other fields. Although the number of generally accessible journal articles on BN modeling has continued to increase in recent years, Bayesian network (BN) models are essentially graphs of vari- achieving an exponential growth at least during the period from ables depicted and linked by probabilities (Koski and Noble, 2011). -
Bayesian Causal Inference
Bayesian Causal Inference Maximilian Kurthen Master’s Thesis Max Planck Institute for Astrophysics Within the Elite Master Program Theoretical and Mathematical Physics Ludwig Maximilian University of Munich Technical University of Munich Supervisor: PD Dr. Torsten Enßlin Munich, September 12, 2018 Abstract In this thesis we address the problem of two-variable causal inference. This task refers to inferring an existing causal relation between two random variables (i.e. X → Y or Y → X ) from purely observational data. We begin by outlining a few basic definitions in the context of causal discovery, following the widely used do-Calculus [Pea00]. We continue by briefly reviewing a number of state-of-the-art methods, including very recent ones such as CGNN [Gou+17] and KCDC [MST18]. The main contribution is the introduction of a novel inference model where we assume a Bayesian hierarchical model, pursuing the strategy of Bayesian model selection. In our model the distribution of the cause variable is given by a Poisson lognormal distribution, which allows to explicitly regard discretization effects. We assume Fourier diagonal covariance operators, where the values on the diagonal are given by power spectra. In the most shallow model these power spectra and the noise variance are fixed hyperparameters. In a deeper inference model we replace the noise variance as a given prior by expanding the inference over the noise variance itself, assuming only a smooth spatial structure of the noise variance. Finally, we make a similar expansion for the power spectra, replacing fixed power spectra as hyperparameters by an inference over those, where again smoothness enforcing priors are assumed. -
Bayesian Network Modelling with Examples
Bayesian Network Modelling with Examples Marco Scutari [email protected] Department of Statistics University of Oxford November 28, 2016 What Are Bayesian Networks? Marco Scutari University of Oxford What Are Bayesian Networks? A Graph and a Probability Distribution Bayesian networks (BNs) are defined by: a network structure, a directed acyclic graph G = (V;A), in which each node vi 2 V corresponds to a random variable Xi; a global probability distribution X with parameters Θ, which can be factorised into smaller local probability distributions according to the arcs aij 2 A present in the graph. The main role of the network structure is to express the conditional independence relationships among the variables in the model through graphical separation, thus specifying the factorisation of the global distribution: p Y P(X) = P(Xi j ΠXi ;ΘXi ) where ΠXi = fparents of Xig i=1 Marco Scutari University of Oxford What Are Bayesian Networks? How the DAG Maps to the Probability Distribution Graphical Probabilistic DAG separation independence A B C E D F Formally, the DAG is an independence map of the probability distribution of X, with graphical separation (??G) implying probabilistic independence (??P ). Marco Scutari University of Oxford What Are Bayesian Networks? Graphical Separation in DAGs (Fundamental Connections) separation (undirected graphs) A B C d-separation (directed acyclic graphs) A B C A B C A B C Marco Scutari University of Oxford What Are Bayesian Networks? Graphical Separation in DAGs (General Case) Now, in the general case we can extend the patterns from the fundamental connections and apply them to every possible path between A and B for a given C; this is how d-separation is defined. -
Articles Causal Inference in Civil Rights Litigation
VOLUME 122 DECEMBER 2008 NUMBER 2 © 2008 by The Harvard Law Review Association ARTICLES CAUSAL INFERENCE IN CIVIL RIGHTS LITIGATION D. James Greiner TABLE OF CONTENTS I. REGRESSION’S DIFFICULTIES AND THE ABSENCE OF A CAUSAL INFERENCE FRAMEWORK ...................................................540 A. What Is Regression? .........................................................................................................540 B. Problems with Regression: Bias, Ill-Posed Questions, and Specification Issues .....543 1. Bias of the Analyst......................................................................................................544 2. Ill-Posed Questions......................................................................................................544 3. Nature of the Model ...................................................................................................545 4. One Regression, or Two?............................................................................................555 C. The Fundamental Problem: Absence of a Causal Framework ....................................556 II. POTENTIAL OUTCOMES: DEFINING CAUSAL EFFECTS AND BEYOND .................557 A. Primitive Concepts: Treatment, Units, and the Fundamental Problem.....................558 B. Additional Units and the Non-Interference Assumption.............................................560 C. Donating Values: Filling in the Missing Counterfactuals ...........................................562 D. Randomization: Balance in Background Variables ......................................................563 -
Empirical Evaluation of Scoring Functions for Bayesian Network Model Selection
Liu et al. BMC Bioinformatics 2012, 13(Suppl 15):S14 http://www.biomedcentral.com/1471-2105/13/S15/S14 PROCEEDINGS Open Access Empirical evaluation of scoring functions for Bayesian network model selection Zhifa Liu1,2†, Brandon Malone1,3†, Changhe Yuan1,4* From Proceedings of the Ninth Annual MCBIOS Conference. Dealing with the Omics Data Deluge Oxford, MS, USA. 17-18 February 2012 Abstract In this work, we empirically evaluate the capability of various scoring functions of Bayesian networks for recovering true underlying structures. Similar investigations have been carried out before, but they typically relied on approximate learning algorithms to learn the network structures. The suboptimal structures found by the approximation methods have unknown quality and may affect the reliability of their conclusions. Our study uses an optimal algorithm to learn Bayesian network structures from datasets generated from a set of gold standard Bayesian networks. Because all optimal algorithms always learn equivalent networks, this ensures that only the choice of scoring function affects the learned networks. Another shortcoming of the previous studies stems from their use of random synthetic networks as test cases. There is no guarantee that these networks reflect real-world data. We use real-world data to generate our gold-standard structures, so our experimental design more closely approximates real-world situations. A major finding of our study suggests that, in contrast to results reported by several prior works, the Minimum Description Length (MDL) (or equivalently, Bayesian information criterion (BIC)) consistently outperforms other scoring functions such as Akaike’s information criterion (AIC), Bayesian Dirichlet equivalence score (BDeu), and factorized normalized maximum likelihood (fNML) in recovering the underlying Bayesian network structures. -
Bayesian Networks
Bayesian Networks Read R&N Ch. 14.1-14.2 Next lecture: Read R&N 18.1-18.4 You will be expected to know • Basic concepts and vocabulary of Bayesian networks. – Nodes represent random variables. – Directed arcs represent (informally) direct influences. – Conditional probability tables, P( Xi | Parents(Xi) ). • Given a Bayesian network: – Write down the full joint distribution it represents. • Given a full joint distribution in factored form: – Draw the Bayesian network that represents it. • Given a variable ordering and some background assertions of conditional independence among the variables: – Write down the factored form of the full joint distribution, as simplified by the conditional independence assertions. Computing with Probabilities: Law of Total Probability Law of Total Probability (aka “summing out” or marginalization) P(a) = Σb P(a, b) = Σb P(a | b) P(b) where B is any random variable Why is this useful? given a joint distribution (e.g., P(a,b,c,d)) we can obtain any “marginal” probability (e.g., P(b)) by summing out the other variables, e.g., P(b) = Σa Σc Σd P(a, b, c, d) Less obvious: we can also compute any conditional probability of interest given a joint distribution, e.g., P(c | b) = Σa Σd P(a, c, d | b) = (1 / P(b)) Σa Σd P(a, c, d, b) where (1 / P(b)) is just a normalization constant Thus, the joint distribution contains the information we need to compute any probability of interest. Computing with Probabilities: The Chain Rule or Factoring We can always write P(a, b, c, … z) = P(a | b, c, …. -
Deep Regression Bayesian Network and Its Applications Siqi Nie, Meng Zheng, Qiang Ji Rensselaer Polytechnic Institute
1 Deep Regression Bayesian Network and Its Applications Siqi Nie, Meng Zheng, Qiang Ji Rensselaer Polytechnic Institute Abstract Deep directed generative models have attracted much attention recently due to their generative modeling nature and powerful data representation ability. In this paper, we review different structures of deep directed generative models and the learning and inference algorithms associated with the structures. We focus on a specific structure that consists of layers of Bayesian Networks due to the property of capturing inherent and rich dependencies among latent variables. The major difficulty of learning and inference with deep directed models with many latent variables is the intractable inference due to the dependencies among the latent variables and the exponential number of latent variable configurations. Current solutions use variational methods often through an auxiliary network to approximate the posterior probability inference. In contrast, inference can also be performed directly without using any auxiliary network to maximally preserve the dependencies among the latent variables. Specifically, by exploiting the sparse representation with the latent space, max-max instead of max-sum operation can be used to overcome the exponential number of latent configurations. Furthermore, the max-max operation and augmented coordinate ascent are applied to both supervised and unsupervised learning as well as to various inference. Quantitative evaluations on benchmark datasets of different models are given for both data representation and feature learning tasks. I. INTRODUCTION Deep learning has become a major enabling technology for computer vision. By exploiting its multi-level representation and the availability of the big data, deep learning has led to dramatic performance improvements for certain tasks. -
A Widely Applicable Bayesian Information Criterion
JournalofMachineLearningResearch14(2013)867-897 Submitted 8/12; Revised 2/13; Published 3/13 A Widely Applicable Bayesian Information Criterion Sumio Watanabe [email protected] Department of Computational Intelligence and Systems Science Tokyo Institute of Technology Mailbox G5-19, 4259 Nagatsuta, Midori-ku Yokohama, Japan 226-8502 Editor: Manfred Opper Abstract A statistical model or a learning machine is called regular if the map taking a parameter to a prob- ability distribution is one-to-one and if its Fisher information matrix is always positive definite. If otherwise, it is called singular. In regular statistical models, the Bayes free energy, which is defined by the minus logarithm of Bayes marginal likelihood, can be asymptotically approximated by the Schwarz Bayes information criterion (BIC), whereas in singular models such approximation does not hold. Recently, it was proved that the Bayes free energy of a singular model is asymptotically given by a generalized formula using a birational invariant, the real log canonical threshold (RLCT), instead of half the number of parameters in BIC. Theoretical values of RLCTs in several statistical models are now being discovered based on algebraic geometrical methodology. However, it has been difficult to estimate the Bayes free energy using only training samples, because an RLCT depends on an unknown true distribution. In the present paper, we define a widely applicable Bayesian information criterion (WBIC) by the average log likelihood function over the posterior distribution with the inverse temperature 1/logn, where n is the number of training samples. We mathematically prove that WBIC has the same asymptotic expansion as the Bayes free energy, even if a statistical model is singular for or unrealizable by a statistical model. -
Causation and Causal Inference in Epidemiology
PUBLIC HEALTH MAnERS Causation and Causal Inference in Epidemiology Kenneth J. Rothman. DrPH. Sander Greenland. MA, MS, DrPH. C Stat fixed. In other words, a cause of a disease Concepts of cause and causal inference are largely self-taught from early learn- ing experiences. A model of causation that describes causes in terms of suffi- event is an event, condition, or characteristic cient causes and their component causes illuminates important principles such that preceded the disease event and without as multicausality, the dependence of the strength of component causes on the which the disease event either would not prevalence of complementary component causes, and interaction between com- have occun-ed at all or would not have oc- ponent causes. curred until some later time. Under this defi- Philosophers agree that causal propositions cannot be proved, and find flaws or nition it may be that no specific event, condi- practical limitations in all philosophies of causal inference. Hence, the role of logic, tion, or characteristic is sufiicient by itself to belief, and observation in evaluating causal propositions is not settled. Causal produce disease. This is not a definition, then, inference in epidemiology is better viewed as an exercise in measurement of an of a complete causal mechanism, but only a effect rather than as a criterion-guided process for deciding whether an effect is pres- component of it. A "sufficient cause," which ent or not. (4m JPub//cHea/f/i. 2005;95:S144-S150. doi:10.2105/AJPH.2004.059204) means a complete causal mechanism, can be defined as a set of minimal conditions and What do we mean by causation? Even among eral. -
Experiments & Observational Studies: Causal Inference in Statistics
Experiments & Observational Studies: Causal Inference in Statistics Paul R. Rosenbaum Department of Statistics University of Pennsylvania Philadelphia, PA 19104-6340 1 My Concept of a ‘Tutorial’ In the computer era, we often receive compressed • files, .zip. Minimize redundancy, minimize storage, at the expense of intelligibility. Sometimes scientific journals seemed to have been • compressed. Tutorial goal is: ‘uncompress’. Make it possible to • read a current article or use current software without going back to dozens of earlier articles. 2 A Causal Question At age 45, Ms. Smith is diagnosed with stage II • breast cancer. Her oncologist discusses with her two possible treat- • ments: (i) lumpectomy alone, or (ii) lumpectomy plus irradiation. They decide on (ii). Ten years later, Ms. Smith is alive and the tumor • has not recurred. Her surgeon, Steve, and her radiologist, Rachael de- • bate. Rachael says: “The irradiation prevented the recur- • rence – without it, the tumor would have recurred.” Steve says: “You can’t know that. It’s a fantasy – • you’re making it up. We’ll never know.” 3 Many Causal Questions Steve and Rachael have this debate all the time. • About Ms. Jones, who had lumpectomy alone. About Ms. Davis, whose tumor recurred after a year. Whenever a patient treated with irradiation remains • disease free, Rachael says: “It was the irradiation.” Steve says: “You can’t know that. It’s a fantasy. We’ll never know.” Rachael says: “Let’s keep score, add ’em up.” Steve • says: “You don’t know what would have happened to Ms. Smith, or Ms. Jones, or Ms Davis – you just made it all up, it’s all fantasy.