The Book of Why a Review by Lisa R

Total Page:16

File Type:pdf, Size:1020Kb

The Book of Why a Review by Lisa R BOOK REVIEW The Book of Why A review by Lisa R. Goldberg The Book of Why The holdup was the specter of a latent factor, perhaps some- The New Science of Cause and Effect thing genetic, that might cause both lung cancer and a crav- Judea Pearl and Dana Macken- ing for tobacco. If the latent factor were responsible for zie lung cancer, limiting cigarette smoking would not prevent Basic Books, 2018 the disease. Naturally, tobacco companies were fond of 432 pages this explanation, but it was also advocated by the promi- ISBN-13: 978-0465097609 nent statistician Ronald A. Fisher, co-inventor of the so- called gold standard of experimentation, the Randomized Judea Pearl is on a mission to Controlled Trial (RCT). change the way we interpret Subjects in an RCT on smoking and lung cancer would data. An eminent professor have been assigned to smoke or not on the flip of a coin. of computer science, Pearl has The study had the potential to disqualify a latent factor documented his research and opinions in scholarly books as the primary cause of lung cancer and elevate cigarettes and papers. Now, he has made his ideas accessible to to the leading suspect. Since a smoking RCT would have a broad audience in The Book of Why: The New Science been unethical, however, researchers made do with ob- of Cause and Effect, co-authored with science writer Dana servational studies showing association, and demurred on Mackenzie. With the release of this historically grounded the question of cause and effect for decades. and thought-provoking book, Pearl leaps from the ivory Was the problem simply that the tools available in the tower into the real world. 1950s and 1960s were too limited in scope? Pearl address- The Book of Why takes aim at perceived limitations of es that question in his three-step Ladder of Causation, observational studies, whose underlying data are found in which organizes inferential methods in terms of the prob- nature and not controlled by researchers. Many believe lems they can solve. The bottom rung is for model-free that an observational study can elucidate association but statistical methods that rely strictly on association or cor- not cause and effect. It cannot tell you why. relation. The middle rung is for interventions that allow Perhaps the most famous example concerns the impact for the measurement of cause and effect. The top rung is of smoking on health. By the mid 1950s, researchers had for counterfactual analysis, the exploration of alternative established a strong association between smoking and realities. lung cancer. Only in 1984, however, did the US govern- Early scientific inquiries about the relationship between ment mandate the phrase “smoking causes lung cancer.” smoking and lung cancer relied on the bottom rung, model-free statistical methods whose modern analogs Lisa Goldberg is a co-director of the Consortium for Data Analytics in Risk and dominate the analysis of observational studies today. In an adjunct professor of Economics and Statistics at University of California, one of The Book of Why’s many wonderful historical anec- Berkeley. She is a director of research at Aperio Group, LLC. Her email address dotes, the predominance of these methods is traced to the is [email protected]. work of Francis Galton, who discovered the principle of re- Communicated by Notices Book Review Editor Stephan Ramon Garcia. gression to the mean in an attempt to understand the pro- For permission to reprint this article, please contact: cess that drives heredity of human characteristics. Regres- [email protected]. sion to the mean involves association, and this led Galton DOI: https://doi.org/10.1090/noti1912 AUGUST 2019 NOTICES OF THE AMERICAN MATHEMATICAL SOCIETY 1093 Book Review and his disciple, Karl Pearson, to conclude that association only to my admiration for his courage and deter- was more central to science than causation. mination. Imagine the situation in 1921. A self- Pearl places deep learning and other modern data min- taught mathematician faces the hegemony of the ing tools on the bottom rung of the Ladder of Causation. statistical establishment alone. They tell him Bottom rung methods include AlphaGo, the deep learning “Your method is based on a complete misappre- program that defeated the world’s best human Go players hension of the nature of causality in the scientific in 2015 and 2016 [1]. For the benefit of those who remem- sense.” And he retorts, “Not so! My method is ber the ancient times before data mining changed every- important and goes beyond anything you can gen- thing, he explains, erate.” The successes of deep learning have been truly re- markable and have caught many of us by surprise. Pearl defines a causal model to be a directed acyclic graph Nevertheless, deep learning has succeeded primar- that can be paired with data to produce quantitative causal ily by showing that certain questions or tasks we estimates. The graph embodies the structural relationships thought were difficult are in fact not. that a researcher assumes are driving empirical results. The structure of the graphical model, including the identifica- The issue is that algorithms, unlike three-year-olds, do as tion of vertices as mediators, confounders, or colliders, they are told, but in order to create an algorithm capable can guide experimental design through the identification of causal reasoning, of minimal sets of control variables. Modern expositions ...we have to teach the computer how to selectively on graphical cause and effect models are [3] and [4]. break the rules of logic. Computers are not good at breaking rules, a skill at which children excel. Figure 2. Mutated causal model facilitating the calculation of the effect of smoking on lung cancer. The arrow from the Figure 1. Causal model of assumed relationships among confounding smoking gene to the act of smoking is deleted. smoking, lung cancer, and a smoking gene. Within this framework, Pearl defines the do operator, Methods for extracting causal conclusions from observa- which isolates the impact of a single variable from other tional studies are on the middle rung of Pearl’s Ladder of effects. The probability of 푌 do 푋, 푃[푌|do(푋)], is not Causation, and they can be expressed in a mathematical the same thing as the conditional probability of 푌 given language that extends classical statistics and emphasizes 푋. Rather 푃[푌|do(푋)] is estimated in a mutated causal graphical models. model, from which arrows pointing into the assumed cause are removed. Confounding is the difference between Various options exist for causal models: causal dia- 푃[푌|do(푋)] and 푃[푌|푋]. In the 1950s, researchers were grams, structural equations, logical statements, after the former but could estimate only the latter in obser- and so forth. I am strongly sold on causal dia- vational studies. That was Ronald A. Fisher’s point. grams for nearly all applications, primarily due to Figure 1 depicts a simplified relationship between smok- their transparency but also due to the explicit an- ing and lung cancer. Directed edges represent assumed swers they provide to many of the questions we causal relationships, and the smoking gene is represented wish to ask. by an empty circle, indicating that the variable was not ob- The use of graphical models to determine cause and effect servable when the connection between smoking and can- in observational studies was pioneered by Sewall Wright, cer was in question. Filled circles represent quantities that whose work on the effects of birth weight, litter size, length could be measured, like rates of smoking and lung cancer of gestation period, and other variables on the weight of a in a population. Figure 2 shows the mutated causal model 33-day-old guinea pig is in [2]. Pearl relates Wright’s per- that isolates the impact of smoking on lung cancer. sistence in response to the cold reception his work received The conclusion that smoking causes lung cancer was from the scientific community. eventually reached without appealing to a causal model. A My admiration for Wright’s precision is second crush of evidence, including the powerful sensitivity anal- 1094 NOTICES OF THE AMERICAN MATHEMATICAL SOCIETY VOLUME 66, NUMBER 7 Book Review ysis developed in [5], ultimately swayed opinion. Pearl ar- fundamentally changed our understanding of how we gues that his methods, had they been available, might have make decisons. Pearl draws on the work of Kahneman and resolved the issue sooner. Pearl illustrates his point in a Tversky in The Book of Why, and Pearl’s approach to analyz- hypothetical setting where smoking causes cancer only by ing counterfactuals might be best explained in terms of a depositing tar in lungs. The corresponding causal diagram question that Kahneman and Tversky posed in their study is shown in Figure 3. His front door formula corrects for the [10] of how we explore alternative realities. confounding of the unobservable smoking gene without How close did Hitler’s scientists come to develop- ever mentioning it. The bias-corrected impact of smoking, ing the atom bomb in World War II? If they had 푋, on lung cancer, 푌, can be expressed developed it in February 1945, would the outcome ′ ′ of the war have been different? 푃[푌|do(푋)] = ∑ 푃[푍|푋] ∑ 푃[푌|푋 , 푍]푃[푋 ]. 푍 푋′ —The Simulation Heuristic Pearl’s response to this question includes the probability of necessity for Germany and its allies to have won World II had they developed the atom bomb in 1945, given our his- torical knowledge that they did not have an atomic bomb in February 1945 and lost the war.
Recommended publications
  • A Philosophical Analysis of Causality in Econometrics
    A Philosophical Analysis of Causality in Econometrics Damien James Fennell London School of Economics and Political Science Thesis submitted to the University of London for the completion of the degree of a Doctor of Philosophy August 2005 1 UMI Number: U209675 All rights reserved INFORMATION TO ALL USERS The quality of this reproduction is dependent upon the quality of the copy submitted. In the unlikely event that the author did not send a complete manuscript and there are missing pages, these will be noted. Also, if material had to be removed, a note will indicate the deletion. Dissertation Publishing UMI U209675 Published by ProQuest LLC 2014. Copyright in the Dissertation held by the Author. Microform Edition © ProQuest LLC. All rights reserved. This work is protected against unauthorized copying under Title 17, United States Code. ProQuest LLC 789 East Eisenhower Parkway P.O. Box 1346 Ann Arbor, Ml 48106-1346 Abstract This thesis makes explicit, develops and critically discusses a concept of causality that is assumed in structural models in econometrics. The thesis begins with a development of Herbert Simon’s (1953) treatment of causal order for linear deterministic, simultaneous systems of equations to provide a fully explicit mechanistic interpretation for these systems. Doing this allows important properties of the assumed causal reading to be discussed including: the invariance of mechanisms to intervention and the role of independence in interventions. This work is then extended to basic structural models actually used in econometrics, linear models with errors-in-the-equations. This part of the thesis provides a discussion of how error terms are to be interpreted and sets out a way to introduce probabilistic concepts into the mechanistic interpretation set out earlier.
    [Show full text]
  • ABSTRACT CAUSAL PROGRAMMING Joshua Brulé
    ABSTRACT Title of dissertation: CAUSAL PROGRAMMING Joshua Brul´e Doctor of Philosophy, 2019 Dissertation directed by: Professor James A. Reggia Department of Computer Science Causality is central to scientific inquiry. There is broad agreement on the meaning of causal statements, such as \Smoking causes cancer", or, \Applying pesticides affects crop yields". However, formalizing the intuition underlying such statements and conducting rigorous inference is difficult in practice. Accordingly, the overall goal of this dissertation is to reduce the difficulty of, and ambiguity in, causal modeling and inference. In other words, the goal is to make it easy for researchers to state precise causal assumptions, understand what they represent, understand why they are necessary, and to yield precise causal conclusions with minimal difficulty. Using the framework of structural causal models, I introduce a causation coeffi- cient as an analogue of the correlation coefficient, analyze its properties, and create a taxonomy of correlation/causation relationships. Analyzing these relationships provides insight into why correlation and causation are often conflated in practice, as well as a principled argument as to why formal causal analysis is necessary. Next, I introduce a theory of causal programming that unifies a large number of previ- ously separate problems in causal modeling and inference. I describe the use and implementation of a causal programming language as an embedded, domain-specific language called `Whittemore'. Whittemore permits rigorously identifying and esti- mating interventional queries without requiring the user to understand the details of the underlying inference algorithms. Finally, I analyze the computational com- plexity in determining the equilibrium distribution of cyclic causal models.
    [Show full text]
  • Bk Brll 010665.Pdf
    12 THE BOOK OF WHY Figure I. How an “inference engine” combines data with causal knowl- edge to produce answers to queries of interest. The dashed box is not part of the engine but is required for building it. Arrows could also be drawn from boxes 4 and 9 to box 1, but I have opted to keep the diagram simple. the Data input, it will use the recipe to produce an actual Estimate for the answer, along with statistical estimates of the amount of uncertainty in that estimate. This uncertainty reflects the limited size of the data set as well as possible measurement errors or miss- ing data. To dig more deeply into the chart, I have labeled the boxes 1 through 9, which I will annotate in the context of the query “What is the effect of Drug D on Lifespan L?” 1. “Knowledge” stands for traces of experience the reasoning agent has had in the past, including past observations, past actions, education, and cultural mores, that are deemed relevant to the query of interest. The dotted box around “Knowledge” indicates that it remains implicit in the mind of the agent and is not explicated formally in the model. 2. Scientific research always requires simplifying assumptions, that is, statements which the researcher deems worthy of making explicit on the basis of the available Knowledge. While most of the researcher’s knowledge remains implicit in his or her brain, only Assumptions see the light of day and 9780465097609-text.indd 12 3/14/18 10:42 AM 26 THE BOOK OF WHY the direction from which to approach the mammoth; in short, by imagining and comparing the consequences of several hunting strategies.
    [Show full text]
  • Causal Screening to Interpret Graph Neural
    Under review as a conference paper at ICLR 2021 CAUSAL SCREENING TO INTERPRET GRAPH NEURAL NETWORKS Anonymous authors Paper under double-blind review ABSTRACT With the growing success of graph neural networks (GNNs), the explainability of GNN is attracting considerable attention. However, current works on feature attribution, which frame explanation generation as attributing a prediction to the graph features, mostly focus on the statistical interpretability. They may struggle to distinguish causal and noncausal effects of features, and quantify redundancy among features, thus resulting in unsatisfactory explanations. In this work, we focus on the causal interpretability in GNNs and propose a method, Causal Screening, from the perspective of cause-effect. It incrementally selects a graph feature (i.e., edge) with large causal attribution, which is formulated as the individual causal effect on the model outcome. As a model-agnostic tool, Causal Screening can be used to generate faithful and concise explanations for any GNN model. Further, by conducting extensive experiments on three graph classification datasets, we observe that Causal Screening achieves significant improvements over state-of-the-art approaches w.r.t. two quantitative metrics: predictive accuracy, contrastivity, and safely passes sanity checks. 1 INTRODUCTION Graph neural networks (GNNs) (Gilmer et al., 2017; Hamilton et al., 2017; Velickovic et al., 2018; Dwivedi et al., 2020) have exhibited impressive performance in a wide range of tasks. Such a success comes from the powerful representation learning, which incorporates the graph structure with node and edge features in an end-to-end fashion. With the growing interest in GNNs, the explainability of GNN is attracting considerable attention.
    [Show full text]
  • Artificial Intelligence Is Stupid and Causal Reasoning Won't Fix It
    Artificial Intelligence is stupid and causal reasoning wont fix it J. Mark Bishop1 1 TCIDA, Goldsmiths, University of London, London, UK Email: [email protected] Abstract Artificial Neural Networks have reached `Grandmaster' and even `super- human' performance' across a variety of games, from those involving perfect- information, such as Go [Silver et al., 2016]; to those involving imperfect- information, such as `Starcraft' [Vinyals et al., 2019]. Such technological developments from AI-labs have ushered concomitant applications across the world of business, where an `AI' brand-tag is fast becoming ubiquitous. A corollary of such widespread commercial deployment is that when AI gets things wrong - an autonomous vehicle crashes; a chatbot exhibits `racist' behaviour; automated credit-scoring processes `discriminate' on gender etc. - there are often significant financial, legal and brand consequences, and the incident becomes major news. As Judea Pearl sees it, the underlying reason for such mistakes is that \... all the impressive achievements of deep learning amount to just curve fitting". The key, Pearl suggests [Pearl and Mackenzie, 2018], is to replace `reasoning by association' with `causal reasoning' - the ability to infer causes from observed phenomena. It is a point that was echoed by Gary Marcus and Ernest Davis in a recent piece for the New York Times: \we need to stop building computer systems that merely get better and better at detecting statistical patterns in data sets { often using an approach known as \Deep Learning" { and start building computer systems that from the moment of arXiv:2008.07371v1 [cs.CY] 20 Jul 2020 their assembly innately grasp three basic concepts: time, space and causal- ity"[Marcus and Davis, 2019].
    [Show full text]
  • Part IV: Reminiscences
    Part IV: Reminiscences 31 Questions and Answers Nils J. Nilsson Few people have contributed as much to artificial intelligence (AI) as has Judea Pearl. Among his several hundred publications, several stand out as among the historically most significant and influential in the theory and practice of AI. With my few pages in this celebratory volume, I join many of his colleagues and former students in showing our gratitude and respect for his inspiration and exemplary career. He is a towering figure in our field. Certainly one key to Judea’s many outstanding achievements (beyond dedication and hard work) is his keen ability to ask the right questions and follow them up with insightful intuitions and penetrating mathematical analyses. His overarching question, it seems to me, is “how is it that humans can do so much with simplistic, unreliable, and uncertain information?” The very name of his UCLA laboratory, the Cognitive Systems Laboratory, seems to proclaim his goal: understanding and automating the most cognitive of all systems, namely humans. In this essay, I’ll focus on the questions and inspirations that motivated his ground-breaking research in three major areas: heuristics, uncertain reasoning, and causality. He has collected and synthesized his work on each of these topics in three important books [Pearl 1984; Pearl 1988; Pearl 2000]. 1 Heuristics Pearl is explicit about what inspired his work on heuristics [Pearl 1984, p. xi]: The study of heuristics draws its inspiration from the ever-amazing ob- servation of how much people can accomplish with that simplistic, un- reliable information source known as intuition.
    [Show full text]
  • Awards and Distinguished Papers
    Awards and Distinguished Papers IJCAI-15 Award for Research Excellence search agenda in their area and will have a first-rate profile of influential re- search results. e Research Excellence award is given to a scientist who has carried out a e award is named for John McCarthy (1927-2011), who is widely rec- program of research of consistently high quality throughout an entire career ognized as one of the founders of the field of artificial intelligence. As well as yielding several substantial results. Past recipients of this honor are the most giving the discipline its name, McCarthy made fundamental contributions illustrious group of scientists from the field of artificial intelligence: John of lasting importance to computer science in general and artificial intelli- McCarthy (1985), Allen Newell (1989), Marvin Minsky (1991), Raymond gence in particular, including time-sharing operating systems, the LISP pro- Reiter (1993), Herbert Simon (1995), Aravind Joshi (1997), Judea Pearl (1999), Donald Michie (2001), Nils Nilsson (2003), Geoffrey E. Hinton gramming languages, knowledge representation, commonsense reasoning, (2005), Alan Bundy (2007), Victor Lesser (2009), Robert Anthony Kowalski and the logicist paradigm in artificial intelligence. e award was estab- (2011), and Hector Levesque (2013). lished with the full support and encouragement of the McCarthy family. e winner of the 2015 Award for Research Excellence is Barbara Grosz, e winner of the 2015 inaugural John McCarthy Award is Bart Selman, Higgins Professor of Natural Sciences at the School of Engineering and Nat- professor at the Department of Computer Science, Cornell University. Pro- ural Sciences, Harvard University. Professor Grosz is recognized for her pio- fessor Selman is recognized for expanding our understanding of problem neering research in natural language processing and in theories and applica- complexity and developing new algorithms for efficient inference.
    [Show full text]
  • Relative Entropy, Probabilistic Inference and AI
    I I RELATIVE ENTROPY, PROBABILISTIC INFERENCE, AND AI John E. Shore I Computer Science and Systems Branch Code 7591 Information Technology Division I Naval Research Laboratory Washington, D. C. 20375 I I. INTRODUCTION For probability distributions q and p, the relative entropy I n "" q · H(q,p) = L.Jqilog -'. (1) i=l p, I is an information-theoretic measure of the dissimilarity between q = q 1, • • · ,qn and p = p1 1 • • • , pn (H is also called cross-entropy, discrimination information, directed divergence, !-divergence, K-L number, among other terms). Various properties of relative entropy have led to its widespread use in information theory. These properties suggest that relative entropy has a role I to play in systems that attempt to perform inference in terms of probability distributions. In this paper, I will review some basic properties of relative entropy as well as its role in probabilistic inference. I will also mention briefly a few existing and potential applications of relative entropy I to so-called artificial intelligence (AI). I II. INFORMATION, ENTROPY, AND RELATIVE-ENTROPY A. Information and Entropy Suppose that some proposition Xi has probability p(xi) of being true. Then the amount of I information that we would obtain if we learn that Xi is true is I ( xi) = - log p ( xi) . (2) !(xi) measures the uncertainty of xi. Except for the base of the logarithm, this common I definition arises essentially from the requirement that the information implicit in two independent propositions be additive. In particular, if xi and xi, i =I=j, are independent propositions with joint probability p(xi/\ x1) = p(xi )p (x; ), then the requirements I !(xi/\ xi)= !(xi)+ !(xi) (3) and I I(xd � 0 (4) are sufficient to establish (2) except for the arbitrary base of the logarithm, which determines the units.
    [Show full text]
  • Artificial Intelligence Pioneer Wins A.M. Turing Award 15 March 2012
    Artificial intelligence pioneer wins A.M. Turing Award 15 March 2012 said Vint Cerf, a Google executive who is considered one of the fathers of the Internet. "They have redefined the term 'thinking machine,'" said Cerf, who is also a Turing Award winner. The ACM said Pearl had created a "computational foundation for processing information under uncertainty, a core problem faced by intelligent systems. "His work serves as the standard method for handling uncertainty in computer systems, with Judea Pearl, winner of the 2011 A.M. Turing Award, is applications ranging from medical diagnosis, pictured with his wife Ruth at a Hanukkah celebration at homeland security and genetic counseling to the White House in 2007. natural language understanding and mapping gene expression data," it said. Judea Pearl, a pioneer in the field of artificial "His influence extends beyond artificial intelligence intelligence, has been awarded the prestigious and even computer science, to human reasoning 2011 A.M. Turing Award. and the philosophy of science." Pearl, 75, was being honored for "innovations that (c) 2012 AFP enabled remarkable advances in the partnership between humans and machines," the Association for Computing Machinery (ACM) said. The award, named for British mathematician Alan M. Turing and considered the "Nobel Prize in Computing," carries a $250,000 prize sponsored by computer chip giant Intel and Internet titan Google. Pearl is a professor of computer science at the University of California, Los Angeles. He is the father of Daniel Pearl, a journalist for The Wall Street Journal who was kidnapped and murdered in Pakistan in 2002. "(Judea Pearl's) accomplishments over the last 30 years have provided the theoretical basis for progress in artificial intelligence and led to extraordinary achievements in machine learning," 1 / 2 APA citation: Artificial intelligence pioneer wins A.M.
    [Show full text]
  • A Two Day ACM Turing Award Celebrations-Live Web Stream @
    ACM Coordinators Department of Information Technology Dr.R.Suganya & Mr.A.Sheik Abdullah Assistant Professors, The Department of Information Technology is A Two day ACM Turing Award Department of Information Technology established in the year 1999. It offers B.Tech (Information Technology) UG Programme since 1999 and M.E. Celebrations-Live Web Stream Thiagarjar College of Engineering (Computer Science and Information Security) PG Mail id: [email protected], [email protected] programme since 2014. The Department is supported by @ TCE PH: +919150245157, +919994216590 well experienced and qualified faculty members. The Department focuses on imparting practical and project based training to student through Outcome Based 23-24th June, 2017 Thiagarajar College of Engineering Education. The Department has different Special Interest Thiagarajar College of Engineering, an autonomous govt. Groups (SIGs) such as Data Engineering, Distributed aided Institution, affiliated to the Anna University is one Systems, Mobile Technologies, Information Security and among the several educational and philanthropic Management. These SIGs work on promoting research and institutions founded by Philanthropist and Industrialist development activities. The Department has sponsored research projects supported by funding organisations like Late. Shri.Karumuttu Thiagarajan Chettiar. It was UGC, AICTE, Industry supported Enterprise Mobility Lab established in the year 1957 and granted Autonomy in the by Motorola and Data Analytics Lab by CDAC, Chennai. year 1987. TCE is funded by Central and State Governments and Management. TCE offers 8 Undergraduate Programmes, 13 Postgraduate Programmes About the Award Celebrations Organized by and Doctoral Programmes in Engineering, Sciences and Architecture. TCE is an approved QIP centre for pursuing The ACM A.M.
    [Show full text]
  • AAAI Presidential Address: Steps Toward Robust Artificial Intelligence
    Articles AAAI Presidential Address: Steps Toward Robust Artificial Intelligence Thomas G. Dietterich I Recent advances in artificial intelli- et me begin by acknowledging the recent death of gence are encouraging governments and Marvin Minsky. Professor Minsky was, of course, corporations to deploy AI in high-stakes Lone of the four authors of the original Dartmouth settings including driving cars Summer School proposal to develop artificial intelligence autonomously, managing the power (McCarthy et al. 1955). In addition to his many contri- grid, trading on stock exchanges, and controlling autonomous weapons sys- butions to the intellectual foundations of artificial intel- tems. Such applications require AI ligence, I remember him most for his iconoclastic and methods to be robust to both the known playful attitude to research ideas. No established idea unknowns (those uncertain aspects of could long withstand his critical assaults, and up to his the world about which the computer can death, he continually urged us all to be more ambitious, reason explicitly) and the unknown to think more deeply, and to keep our eyes focused on unknowns (those aspects of the world the fundamental questions. that are not captured by the system’s In 1959, Minsky wrote an influential essay titled Steps models). This article discusses recent progress in AI and then describes eight Toward Artificial Intelligence (Minsky 1961), in which he ideas related to robustness that are summarized the state of AI research and sketched a path being pursued within the AI research forward. In his honor, I have extended his title to incor- community. While these ideas are a porate the topic that I want to discuss today: How can we start, we need to devote more attention make artificial intelligence systems that are robust in the to the challenges of dealing with the face of lack of knowledge about the world? known and unknown unknowns.
    [Show full text]
  • Introduction to Causal Inference
    Clinical Development & Analytics Statistical Methodology A Gentle Introduction to Causal Inference in View of the ICH E9 Addendum on Estimands Björn Bornkamp, Heinz Schmidli, Dong Xi ASA Biopharmaceutical Section Regulatory-Industry Statistics Workshop September 22, 2020 Disclaimer The views and opinions expressed in this presentation and on the slides are solely those of the presenter and not necessarily those of Novartis. Novartis does not guarantee the accuracy or reliability of the information provided herein. 2 Public Acknowledgements . Novartis . Simon Newsome, Baldur Magnusson, Angelika Caputo . Harvard University . Alex Ocampo . Johns Hopkins University . Daniel Scharfstein 3 Public Agenda 10:00 – 11:40 AM . Introduction to causal effects & potential outcomes (Heinz Schmidli) . Relation to questions and concepts encountered in randomized clinical trials (Björn Bornkamp) 12:00 – 1:30 PM . Standardization & inverse probability weighting (Dong Xi) 4 Public Part 1: Introduction to causal effects and potential outcomes Outline . Causal effects . Potential outcomes . Causal estimands . Causal inference . Clinical development . Conclusions 6 Public Causal effects Does smoking cause lung cancer? Cancer and Smoking The curious associations with lung cancer found in relation to smoking habits do not, in the minds of some of us, lend themselves easily to the simple conclusion that the products of combustion reaching the surface of the bronchus induce, though after a long interval, the development of a cancer. Ronald A. Fisher Nature 1958;182(4635):596. 7 Public Causal effects Directed Acyclic Graph (DAG) to express causal relationships Smoking causes cancer Patient characteristic X causes both smoking and cancer 8 Public Poll question 1 Do you believe that smoking causes lung cancer? .
    [Show full text]