Quasi-Experimental Design and Methods
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
DATA COLLECTION METHODS Section III 5
AN OVERVIEW OF QUANTITATIVE AND QUALITATIVE DATA COLLECTION METHODS Section III 5. DATA COLLECTION METHODS: SOME TIPS AND COMPARISONS In the previous chapter, we identified two broad types of evaluation methodologies: quantitative and qualitative. In this section, we talk more about the debate over the relative virtues of these approaches and discuss some of the advantages and disadvantages of different types of instruments. In such a debate, two types of issues are considered: theoretical and practical. Theoretical Issues Most often these center on one of three topics: · The value of the types of data · The relative scientific rigor of the data · Basic, underlying philosophies of evaluation Value of the Data Quantitative and qualitative techniques provide a tradeoff between breadth and depth, and between generalizability and targeting to specific (sometimes very limited) populations. For example, a quantitative data collection methodology such as a sample survey of high school students who participated in a special science enrichment program can yield representative and broadly generalizable information about the proportion of participants who plan to major in science when they get to college and how this proportion differs by gender. But at best, the survey can elicit only a few, often superficial reasons for this gender difference. On the other hand, separate focus groups (a qualitative technique related to a group interview) conducted with small groups of men and women students will provide many more clues about gender differences in the choice of science majors, and the extent to which the special science program changed or reinforced attitudes. The focus group technique is, however, limited in the extent to which findings apply beyond the specific individuals included in the groups. -
A Note on the Impact Evaluation of Public Policies: the Counterfactual Analysis
A note on the impact evaluation of public policies: the counterfactual analysis Massimo Loi and Margarida Rodrigues 2012 Report EUR 25519 EN 1 European Commission Joint Research Centre Institute for the Protection and Security of the Citizen Contact information Forename Surname Address: Joint Research Centre, Via Enrico Fermi 2749, TP 361, 21027 Ispra (VA), Italy E-mail: [email protected] Tel.: +39 0332 78 5633 Fax: +39 0332 78 5733 http://ipsc.jrc.ec.europa.eu/ http://www.jrc.ec.europa.eu/ Legal Notice Neither the European Commission nor any person acting on behalf of the Commission is responsible for the use which might be made of this publication. Europe Direct is a service to help you find answers to your questions about the European Union Freephone number (*): 00 800 6 7 8 9 10 11 (*) Certain mobile telephone operators do not allow access to 00 800 numbers or these calls may be billed. A great deal of additional information on the European Union is available on the Internet. It can be accessed through the Europa server http://europa.eu/. JRC74778 EUR 25519 EN ISBN 978-92-79-26425-2 ISSN 1831-9424 doi:10.2788/50327 Luxembourg: Publications Office of the European Union, 2012 © European Union, 2012 Reproduction is authorised provided the source is acknowledged. Printed in Italy Contents 1. Introduction ............................................................................................................................. 4 2. Basic concepts ........................................................................................................................ -
Statistical Inference: How Reliable Is a Survey?
Math 203, Fall 2008: Statistical Inference: How reliable is a survey? Consider a survey with a single question, to which respondents are asked to give an answer of yes or no. Suppose you pick a random sample of n people, and you find that the proportion that answered yes isp ˆ. Question: How close isp ˆ to the actual proportion p of people in the whole population who would have answered yes? In order for there to be a reliable answer to this question, the sample size, n, must be big enough so that the sample distribution is close to a bell shaped curve (i.e., close to a normal distribution). But even if n is big enough that the distribution is close to a normal distribution, usually you need to make n even bigger in order to make sure your margin of error is reasonably small. Thus the first thing to do is to be sure n is big enough for the sample distribution to be close to normal. The industry standard for being close enough is for n to be big enough so that 1 − p 1 − p n > 9 and n > 9 p p both hold. When p is about 50%, n can be as small as 10, but when p gets close to 0 or close to 1, the sample size n needs to get bigger. If p is 1% or 99%, then n must be at least 892, for example. (Note also that n here depends on p but not on the size of the whole population.) See Figures 1 and 2 showing frequency histograms for the number of yes respondents if p = 1% when the sample size n is 10 versus 1000 (this data was obtained by running a computer simulation taking 10000 samples). -
Data Extraction for Complex Meta-Analysis (Decimal) Guide
Pedder, H. , Sarri, G., Keeney, E., Nunes, V., & Dias, S. (2016). Data extraction for complex meta-analysis (DECiMAL) guide. Systematic Reviews, 5, [212]. https://doi.org/10.1186/s13643-016-0368-4 Publisher's PDF, also known as Version of record License (if available): CC BY Link to published version (if available): 10.1186/s13643-016-0368-4 Link to publication record in Explore Bristol Research PDF-document This is the final published version of the article (version of record). It first appeared online via BioMed Central at http://systematicreviewsjournal.biomedcentral.com/articles/10.1186/s13643-016-0368-4. Please refer to any applicable terms of use of the publisher. University of Bristol - Explore Bristol Research General rights This document is made available in accordance with publisher policies. Please cite only the published version using the reference above. Full terms of use are available: http://www.bristol.ac.uk/red/research-policy/pure/user-guides/ebr-terms/ Pedder et al. Systematic Reviews (2016) 5:212 DOI 10.1186/s13643-016-0368-4 RESEARCH Open Access Data extraction for complex meta-analysis (DECiMAL) guide Hugo Pedder1*, Grammati Sarri2, Edna Keeney3, Vanessa Nunes1 and Sofia Dias3 Abstract As more complex meta-analytical techniques such as network and multivariate meta-analyses become increasingly common, further pressures are placed on reviewers to extract data in a systematic and consistent manner. Failing to do this appropriately wastes time, resources and jeopardises accuracy. This guide (data extraction for complex meta-analysis (DECiMAL)) suggests a number of points to consider when collecting data, primarily aimed at systematic reviewers preparing data for meta-analysis. -
SAMPLING DESIGN & WEIGHTING in the Original
Appendix A 2096 APPENDIX A: SAMPLING DESIGN & WEIGHTING In the original National Science Foundation grant, support was given for a modified probability sample. Samples for the 1972 through 1974 surveys followed this design. This modified probability design, described below, introduces the quota element at the block level. The NSF renewal grant, awarded for the 1975-1977 surveys, provided funds for a full probability sample design, a design which is acknowledged to be superior. Thus, having the wherewithal to shift to a full probability sample with predesignated respondents, the 1975 and 1976 studies were conducted with a transitional sample design, viz., one-half full probability and one-half block quota. The sample was divided into two parts for several reasons: 1) to provide data for possibly interesting methodological comparisons; and 2) on the chance that there are some differences over time, that it would be possible to assign these differences to either shifts in sample designs, or changes in response patterns. For example, if the percentage of respondents who indicated that they were "very happy" increased by 10 percent between 1974 and 1976, it would be possible to determine whether it was due to changes in sample design, or an actual increase in happiness. There is considerable controversy and ambiguity about the merits of these two samples. Text book tests of significance assume full rather than modified probability samples, and simple random rather than clustered random samples. In general, the question of what to do with a mixture of samples is no easier solved than the question of what to do with the "pure" types. -
The Practice of Causal Inference in Cancer Epidemiology
Vol. 5. 303-31 1, April 1996 Cancer Epidemiology, Biomarkers & Prevention 303 Review The Practice of Causal Inference in Cancer Epidemiology Douglas L. Weedt and Lester S. Gorelic causes lung cancer (3) and to guide causal inference for occu- Preventive Oncology Branch ID. L. W.l and Comprehensive Minority pational and environmental diseases (4). From 1965 to 1995, Biomedical Program IL. S. 0.1. National Cancer Institute, Bethesda, Maryland many associations have been examined in terms of the central 20892 questions of causal inference. Causal inference is often practiced in review articles and editorials. There, epidemiologists (and others) summarize evi- Abstract dence and consider the issues of causality and public health Causal inference is an important link between the recommendations for specific exposure-cancer associations. practice of cancer epidemiology and effective cancer The purpose of this paper is to take a first step toward system- prevention. Although many papers and epidemiology atically reviewing the practice of causal inference in cancer textbooks have vigorously debated theoretical issues in epidemiology. Techniques used to assess causation and to make causal inference, almost no attention has been paid to the public health recommendations are summarized for two asso- issue of how causal inference is practiced. In this paper, ciations: alcohol and breast cancer, and vasectomy and prostate we review two series of review papers published between cancer. The alcohol and breast cancer association is timely, 1985 and 1994 to find answers to the following questions: controversial, and in the public eye (5). It involves a common which studies and prior review papers were cited, which exposure and a common cancer and has a large body of em- causal criteria were used, and what causal conclusions pirical evidence; over 50 studies and over a dozen reviews have and public health recommendations ensued. -
Statistics and Causal Inference (With Discussion)
Applied Statistics Lecture Notes Kosuke Imai Department of Politics Princeton University February 2, 2008 Making statistical inferences means to learn about what you do not observe, which is called parameters, from what you do observe, which is called data. We learn the basic principles of statistical inference from a perspective of causal inference, which is a popular goal of political science research. Namely, we study statistics by learning how to make causal inferences with statistical methods. 1 Statistical Framework of Causal Inference What do we exactly mean when we say “An event A causes another event B”? Whether explicitly or implicitly, this question is asked and answered all the time in political science research. The most commonly used statistical framework of causality is based on the notion of counterfactuals (see Holland, 1986). That is, we ask the question “What would have happened if an event A were absent (or existent)?” The following example illustrates the fact that some causal questions are more difficult to answer than others. Example 1 (Counterfactual and Causality) Interpret each of the following statements as a causal statement. 1. A politician voted for the education bill because she is a democrat. 2. A politician voted for the education bill because she is liberal. 3. A politician voted for the education bill because she is a woman. In this framework, therefore, the fundamental problem of causal inference is that the coun- terfactual outcomes cannot be observed, and yet any causal inference requires both factual and counterfactual outcomes. This idea is formalized below using the potential outcomes notation. -
Bayesian Causal Inference
Bayesian Causal Inference Maximilian Kurthen Master’s Thesis Max Planck Institute for Astrophysics Within the Elite Master Program Theoretical and Mathematical Physics Ludwig Maximilian University of Munich Technical University of Munich Supervisor: PD Dr. Torsten Enßlin Munich, September 12, 2018 Abstract In this thesis we address the problem of two-variable causal inference. This task refers to inferring an existing causal relation between two random variables (i.e. X → Y or Y → X ) from purely observational data. We begin by outlining a few basic definitions in the context of causal discovery, following the widely used do-Calculus [Pea00]. We continue by briefly reviewing a number of state-of-the-art methods, including very recent ones such as CGNN [Gou+17] and KCDC [MST18]. The main contribution is the introduction of a novel inference model where we assume a Bayesian hierarchical model, pursuing the strategy of Bayesian model selection. In our model the distribution of the cause variable is given by a Poisson lognormal distribution, which allows to explicitly regard discretization effects. We assume Fourier diagonal covariance operators, where the values on the diagonal are given by power spectra. In the most shallow model these power spectra and the noise variance are fixed hyperparameters. In a deeper inference model we replace the noise variance as a given prior by expanding the inference over the noise variance itself, assuming only a smooth spatial structure of the noise variance. Finally, we make a similar expansion for the power spectra, replacing fixed power spectra as hyperparameters by an inference over those, where again smoothness enforcing priors are assumed. -
Summary of Human Subjects Protection Issues Related to Large Sample Surveys
Summary of Human Subjects Protection Issues Related to Large Sample Surveys U.S. Department of Justice Bureau of Justice Statistics Joan E. Sieber June 2001, NCJ 187692 U.S. Department of Justice Office of Justice Programs John Ashcroft Attorney General Bureau of Justice Statistics Lawrence A. Greenfeld Acting Director Report of work performed under a BJS purchase order to Joan E. Sieber, Department of Psychology, California State University at Hayward, Hayward, California 94542, (510) 538-5424, e-mail [email protected]. The author acknowledges the assistance of Caroline Wolf Harlow, BJS Statistician and project monitor. Ellen Goldberg edited the document. Contents of this report do not necessarily reflect the views or policies of the Bureau of Justice Statistics or the Department of Justice. This report and others from the Bureau of Justice Statistics are available through the Internet — http://www.ojp.usdoj.gov/bjs Table of Contents 1. Introduction 2 Limitations of the Common Rule with respect to survey research 2 2. Risks and benefits of participation in sample surveys 5 Standard risk issues, researcher responses, and IRB requirements 5 Long-term consequences 6 Background issues 6 3. Procedures to protect privacy and maintain confidentiality 9 Standard issues and problems 9 Confidentiality assurances and their consequences 21 Emerging issues of privacy and confidentiality 22 4. Other procedures for minimizing risks and promoting benefits 23 Identifying and minimizing risks 23 Identifying and maximizing possible benefits 26 5. Procedures for responding to requests for help or assistance 28 Standard procedures 28 Background considerations 28 A specific recommendation: An experiment within the survey 32 6. -
Survey Experiments
IU Workshop in Methods – 2019 Survey Experiments Testing Causality in Diverse Samples Trenton D. Mize Department of Sociology & Advanced Methodologies (AMAP) Purdue University Survey Experiments Page 1 Survey Experiments Page 2 Contents INTRODUCTION ............................................................................................................................................................................ 8 Overview .............................................................................................................................................................................. 8 What is a survey experiment? .................................................................................................................................... 9 What is an experiment?.............................................................................................................................................. 10 Independent and dependent variables ................................................................................................................. 11 Experimental Conditions ............................................................................................................................................. 12 WHY CONDUCT A SURVEY EXPERIMENT? ........................................................................................................................... 13 Internal, external, and construct validity .......................................................................................................... -
Propensity Scores up to Head(Propensitymodel) – Modify the Number of Threads! • Inspect the PS Distribution Plot • Inspect the PS Model
Overview of the CohortMethod package Martijn Schuemie CohortMethod is part of the OHDSI Methods Library Cohort Method Self-Controlled Case Series Self-Controlled Cohort IC Temporal Pattern Disc. Case-control New-user cohort studies using Self-Controlled Case Series A self-controlled cohort A self-controlled design, but Case-control studies, large-scale regressions for analysis using fews or many design, where times preceding using temporals patterns matching controlss on age, propensity and outcome predictors, includes splines for exposure is used as control. around other exposures and gender, provider, and visit models age and seasonality. outcomes to correct for time- date. Allows nesting of the Estimation methods varying confounding. study in another cohort. Patient Level Prediction Feature Extraction Build and evaluate predictive Automatically extract large models for users- specified sets of featuress for user- outcomes, using a wide array specified cohorts using data in of machine learning the CDM. Prediction methods algorithms. Empirical Calibration Method Evaluation Use negative control Use real data and established exposure-outcomes pairs to reference sets ass well as profile and calibrate a simulations injected in real particular analysis design. data to evaluate the performance of methods. Method characterization Database Connector Sql Render Cyclops Ohdsi R Tools Connect directly to a wide Generate SQL on the fly for Highly efficient Support tools that didn’t fit range of databases platforms, the various SQLs dialects. implementations of regularized other categories,s including including SQL Server, Oracle, logistic, Poisson and Cox tools for maintaining R and PostgreSQL. regression. libraries. Supporting packages Under construction Technologies CohortMethod uses • DatabaseConnector and SqlRender to interact with the CDM data – SQL Server – Oracle – PostgreSQL – Amazon RedShift – Microsoft APS • ff to work with large data objects • Cyclops for large scale regularized regression Graham study steps 1. -
Generalized Scatter Plots
Generalized scatter plots Daniel A. Keim a Abstract Scatter Plots are one of the most powerful and most widely used b techniques for visual data exploration. A well-known problem is that scatter Ming C. Hao plots often have a high degree of overlap, which may occlude a significant Umeshwar Dayal b portion of the data values shown. In this paper, we propose the general a ized scatter plot technique, which allows an overlap-free representation of Halldor Janetzko and large data sets to fit entirely into the display. The basic idea is to allow the Peter Baka ,* analyst to optimize the degree of overlap and distortion to generate the best possible view. To allow an effective usage, we provide the capability to zoom ' University of Konstanz, Universitaetsstr. 10, smoothly between the traditional and our generalized scatter plots. We iden Konstanz, Germany. tify an optimization function that takes overlap and distortion of the visualiza bHewlett Packard Research Labs, 1501 Page tion into acccount. We evaluate the generalized scatter plots according to this Mill Road, Palo Alto, CA94304, USA. optimization function, and show that there usually exists an optimal compro mise between overlap and distortion . Our generalized scatter plots have been ' Corresponding author. applied successfully to a number of real-world IT services applications, such as server performance monitoring, telephone service usage analysis and financial data, demonstrating the benefits of the generalized scatter plots over tradi tional ones. Keywords: scatter plot; overlapping; distortion; interpolation; smoothing; interactions Introduction Motivation Large amounts of multi-dimensional data occur in many important appli cation domains such as telephone service usage analysis, sales and server performance monitoring.