Ch7 Sampling Techniques
Total Page:16
File Type:pdf, Size:1020Kb
Load more
Recommended publications
-
Hypothesis Testing with Two Categorical Variables 203
Chapter 10 distribute or Hypothesis Testing With Two Categoricalpost, Variables Chi-Square copy, Learning Objectives • Identify the correct types of variables for use with a chi-square test of independence. • Explainnot the difference between parametric and nonparametric statistics. • Conduct a five-step hypothesis test for a contingency table of any size. • Explain what statistical significance means and how it differs from practical significance. • Identify the correct measure of association for use with a particular chi-square test, Doand interpret those measures. • Use SPSS to produce crosstabs tables, chi-square tests, and measures of association. 202 Part 3 Hypothesis Testing Copyright ©2016 by SAGE Publications, Inc. This work may not be reproduced or distributed in any form or by any means without express written permission of the publisher. he chi-square test of independence is used when the independent variable (IV) and dependent variable (DV) are both categorical (nominal or ordinal). The chi-square test is member of the family of nonparametric statistics, which are statistical Tanalyses used when sampling distributions cannot be assumed to be normally distributed, which is often the result of the DV being categorical rather than continuous (we will talk in detail about this). Chi-square thus sits in contrast to parametric statistics, which are used when DVs are continuous and sampling distributions are safely assumed to be normal. The t test, analysis of variance, and correlation are all parametric. Before going into the theory and math behind the chi-square statistic, read the Research Examples for illustrations of the types of situations in which a criminal justice or criminology researcher would utilize a chi- square test. -
Combining Probability and Nonprobability Samples to Form Efficient Hybrid Estimates: an Evaluation of the Common Support Assumption Jill A
Combining Probability and Nonprobability Samples to form Efficient Hybrid Estimates: An Evaluation of the Common Support Assumption Jill A. Dever RTI International, Washington, DC Proceedings of the 2018 Federal Committee on Statistical Methodology (FCSM) Research Conference Abstract Nonprobability surveys, those without a defined random sampling scheme, are becoming more prevalent. These studies can offer faster results at less cost than many probability surveys, especially for targeting important subpopulations. This can be an attractive option given the continual challenge of doing more with less, as survey costs continue to rise and response rates to plummet. Nonprobability surveys alone, however, may not fit the needs (purpose) of Federal statistical agencies where population inference is critical. Nonprobability samples may best serve to enhance the representativeness of certain domains within probability samples. For example, if locating and interviewing a required number of subpopulation members is resource prohibitive, data from a targeted nonprobability survey may lower coverage bias exhibited in a probability survey. In this situation, the question is how to best combine information from both sources. Our research searches for an answer to this question through an evaluation of hybrid estimation methods currently in use that combine probability and nonprobability data. Methods that employ generalized analysis weights (i.e., one set of weights for all analyses) are the focus because they enable other survey researchers and policy makers to analyze the data. The goal is to identify procedures that maximize the strength of each data source to produce hybrid estimates with the low mean square error. The details presented here focus on the propensity score adjusted (PSA) nonprobability weights needed prior to combining the data sources, and the common support assumption critical to hybrid estimation and PSA weighting. -
A Critical Review of Studies Investigating the Quality of Data Obtained with Online Panels Based on Probability and Nonprobability Samples1
Callegaro c02.tex V1 - 01/16/2014 6:25 P.M. Page 23 2 A critical review of studies investigating the quality of data obtained with online panels based on probability and nonprobability samples1 Mario Callegaro1, Ana Villar2, David Yeager3,and Jon A. Krosnick4 1Google, UK 2City University, London, UK 3University of Texas at Austin, USA 4Stanford University, USA 2.1 Introduction Online panels have been used in survey research as data collection tools since the late 1990s (Postoaca, 2006). The potential great cost and time reduction of using these tools have made research companies enthusiastically pursue this new mode of data collection. However, 1 We would like to thank Reg Baker and Anja Göritz, Part editors, for their useful comments on preliminary versions of this chapter. Online Panel Research, First Edition. Edited by Mario Callegaro, Reg Baker, Jelke Bethlehem, Anja S. Göritz, Jon A. Krosnick and Paul J. Lavrakas. © 2014 John Wiley & Sons, Ltd. Published 2014 by John Wiley & Sons, Ltd. Callegaro c02.tex V1 - 01/16/2014 6:25 P.M. Page 24 24 ONLINE PANEL RESEARCH the vast majority of these online panels were built by sampling and recruiting respondents through nonprobability methods such as snowball sampling, banner ads, direct enrollment, and other strategies to obtain large enough samples at a lower cost (see Chapter 1). Only a few companies and research teams chose to build online panels based on probability samples of the general population. During the 1990s, two probability-based online panels were documented: the CentER data Panel in the Netherlands and the Knowledge Networks Panel in the United States. -
Esomar/Grbn Guideline for Online Sample Quality
ESOMAR/GRBN GUIDELINE FOR ONLINE SAMPLE QUALITY ESOMAR GRBN ONLINE SAMPLE QUALITY GUIDELINE ESOMAR, the World Association for Social, Opinion and Market Research, is the essential organisation for encouraging, advancing and elevating market research: www.esomar.org. GRBN, the Global Research Business Network, connects 38 research associations and over 3500 research businesses on five continents: www.grbn.org. © 2015 ESOMAR and GRBN. Issued February 2015. This Guideline is drafted in English and the English text is the definitive version. The text may be copied, distributed and transmitted under the condition that appropriate attribution is made and the following notice is included “© 2015 ESOMAR and GRBN”. 2 ESOMAR GRBN ONLINE SAMPLE QUALITY GUIDELINE CONTENTS 1 INTRODUCTION AND SCOPE ................................................................................................... 4 2 DEFINITIONS .............................................................................................................................. 4 3 KEY REQUIREMENTS ................................................................................................................ 6 3.1 The claimed identity of each research participant should be validated. .................................................. 6 3.2 Providers must ensure that no research participant completes the same survey more than once ......... 8 3.3 Research participant engagement should be measured and reported on ............................................... 9 3.4 The identity and personal -
SAMPLING DESIGN & WEIGHTING in the Original
Appendix A 2096 APPENDIX A: SAMPLING DESIGN & WEIGHTING In the original National Science Foundation grant, support was given for a modified probability sample. Samples for the 1972 through 1974 surveys followed this design. This modified probability design, described below, introduces the quota element at the block level. The NSF renewal grant, awarded for the 1975-1977 surveys, provided funds for a full probability sample design, a design which is acknowledged to be superior. Thus, having the wherewithal to shift to a full probability sample with predesignated respondents, the 1975 and 1976 studies were conducted with a transitional sample design, viz., one-half full probability and one-half block quota. The sample was divided into two parts for several reasons: 1) to provide data for possibly interesting methodological comparisons; and 2) on the chance that there are some differences over time, that it would be possible to assign these differences to either shifts in sample designs, or changes in response patterns. For example, if the percentage of respondents who indicated that they were "very happy" increased by 10 percent between 1974 and 1976, it would be possible to determine whether it was due to changes in sample design, or an actual increase in happiness. There is considerable controversy and ambiguity about the merits of these two samples. Text book tests of significance assume full rather than modified probability samples, and simple random rather than clustered random samples. In general, the question of what to do with a mixture of samples is no easier solved than the question of what to do with the "pure" types. -
Lecture 1: Why Do We Use Statistics, Populations, Samples, Variables, Why Do We Use Statistics?
1pops_samples.pdf Michael Hallstone, Ph.D. [email protected] Lecture 1: Why do we use statistics, populations, samples, variables, why do we use statistics? • interested in understanding the social world • we want to study a portion of it and say something about it • ex: drug users, homeless, voters, UH students Populations and Samples Populations, Sampling Elements, Frames, and Units A researcher defines a group, “list,” or pool of cases that she wishes to study. This is a population. Another definition: population = complete collection of measurements, objects or individuals under study. 1 of 11 sample = a portion or subset taken from population funny circle diagram so we take a sample and infer to population Why? feasibility – all MD’s in world , cost, time, and stay tuned for the central limits theorem...the most important lecture of this course. Visualizing Samples (taken from) Populations Population Group you wish to study (Mostly made up of “people” in the Then we infer from sample back social sciences) to population (ALWAYS SOME ERROR! “sampling error” Sample (a portion or subset of the population) 4 This population is made up of the things she wishes to actually study called sampling elements. Sampling elements can be people, organizations, schools, whales, molecules, and articles in the popular press, etc. The sampling element is your exact unit of analysis. For crime researchers studying car thieves, the sampling element would probably be individual car thieves – or theft incidents reported to the police. For drug researchers the sampling elements would be most likely be individual drug users. Inferential statistics is truly the basis of much of our scientific evidence. -
Stratified Sampling Using Cluster Analysis: a Sample Selection Strategy for Improved Generalizations from Experiments
Article Evaluation Review 1-31 ª The Author(s) 2014 Stratified Sampling Reprints and permission: sagepub.com/journalsPermissions.nav DOI: 10.1177/0193841X13516324 Using Cluster erx.sagepub.com Analysis: A Sample Selection Strategy for Improved Generalizations From Experiments Elizabeth Tipton1 Abstract Background: An important question in the design of experiments is how to ensure that the findings from the experiment are generalizable to a larger population. This concern with generalizability is particularly important when treatment effects are heterogeneous and when selecting units into the experiment using random sampling is not possible—two conditions commonly met in large-scale educational experiments. Method: This article introduces a model-based balanced-sampling framework for improv- ing generalizations, with a focus on developing methods that are robust to model misspecification. Additionally, the article provides a new method for sample selection within this framework: First units in an inference popula- tion are divided into relatively homogenous strata using cluster analysis, and 1 Department of Human Development, Teachers College, Columbia University, NY, USA Corresponding Author: Elizabeth Tipton, Department of Human Development, Teachers College, Columbia Univer- sity, 525 W 120th St, Box 118, NY 10027, USA. Email: [email protected] 2 Evaluation Review then the sample is selected using distance rankings. Result: In order to demonstrate and evaluate the method, a reanalysis of a completed experiment is conducted. This example compares samples selected using the new method with the actual sample used in the experiment. Results indicate that even under high nonresponse, balance is better on most covariates and that fewer coverage errors result. Conclusion: The article concludes with a discussion of additional benefits and limitations of the method. -
The Effect of Sampling Error on the Time Series Behavior of Consumption Data*
Journal of Econometrics 55 (1993) 235-265. North-Holland The effect of sampling error on the time series behavior of consumption data* William R. Bell U.S.Bureau of the Census, Washington, DC 20233, USA David W. Wilcox Board of Governors of the Federal Reserve System, Washington, DC 20551, USA Much empirical economic research today involves estimation of tightly specified time series models that derive from theoretical optimization problems. Resulting conclusions about underly- ing theoretical parameters may be sensitive to imperfections in the data. We illustrate this fact by considering sampling error in data from the Census Bureau’s Retail Trade Survey. We find that parameter estimates in seasonal time series models for retail sales are sensitive to whether a sampling error component is included in the model. We conclude that sampling error should be taken seriously in attempts to derive economic implications by modeling time series data from repeated surveys. 1. Introduction The rational expectations revolution has transformed the methodology of macroeconometric research. In the new style, the researcher typically begins by specifying a dynamic optimization problem faced by agents in the model economy. Then the researcher derives the solution to the optimization problem, expressed as a stochastic model for some observable economic variable(s). A trademark of research in this tradition is that each of the parameters in the model for the observable variables is endowed with a specific economic interpretation. The last step in the research program is the application of the model to actual economic data to see whether the model *This paper reports the general results of research undertaken by Census Bureau and Federal Reserve Board staff. -
13 Collecting Statistical Data
13 Collecting Statistical Data 13.1 The Population 13.2 Sampling 13.3 Random Sampling 1.1 - 1 • Polls, studies, surveys and other data collecting tools collect data from a small part of a larger group so that we can learn something about the larger group. • This is a common and important goal of statistics: Learn about a large group by examining data from some of its members. 1.1 - 2 Data collections of observations (such as measurements, genders, survey responses) 1.1 - 3 Statistics is the science of planning studies and experiments, obtaining data, and then organizing, summarizing, presenting, analyzing, interpreting, and drawing conclusions based on the data 1.1 - 4 Population the complete collection of all individuals (scores, people, measurements, and so on) to be studied; the collection is complete in the sense that it includes all of the individuals to be studied 1.1 - 5 Census Collection of data from every member of a population Sample Subcollection of members selected from a population 1.1 - 6 A Survey • The practical alternative to a census is to collect data only from some members of the population and use that data to draw conclusions and make inferences about the entire population. • Statisticians call this approach a survey (or a poll when the data collection is done by asking questions). • The subgroup chosen to provide the data is called the sample, and the act of selecting a sample is called sampling. 1.1 - 7 A Survey • The first important step in a survey is to distinguish the population for which the survey applies (the target population) and the actual subset of the population from which the sample will be drawn, called the sampling frame. -
Options for Conducting Web Surveys Matthias Schonlau and Mick P
Statistical Science 2017, Vol. 32, No. 2, 279–292 DOI: 10.1214/16-STS597 © Institute of Mathematical Statistics, 2017 Options for Conducting Web Surveys Matthias Schonlau and Mick P. Couper Abstract. Web surveys can be conducted relatively fast and at relatively low cost. However, Web surveys are often conducted with nonprobability sam- ples and, therefore, a major concern is generalizability. There are two main approaches to address this concern: One, find a way to conduct Web surveys on probability samples without losing most of the cost and speed advantages (e.g., by using mixed-mode approaches or probability-based panel surveys). Two, make adjustments (e.g., propensity scoring, post-stratification, GREG) to nonprobability samples using auxiliary variables. We review both of these approaches as well as lesser-known ones such as respondent-driven sampling. There are many different ways Web surveys can solve the challenge of gen- eralizability. Rather than adopting a one-size-fits-all approach, we conclude that the choice of approach should be commensurate with the purpose of the study. Key words and phrases: Convenience sample, Internet survey. 1. INTRODUCTION tion and invitation of sample persons to a Web sur- vey. No complete list of e-mail addresses of the general Web or Internet surveys1 have come to dominate the survey world in a very short time (see Couper, 2000; population exists from which one can select a sample Couper and Miller, 2008). The attraction of Web sur- and send e-mailed invitations to a Web survey. How- veys lies in the speed with which large numbers of ever, for many other important populations of interest people can be surveyed at relatively low cost, using (e.g., college students, members of professional asso- complex instruments that extend measurement beyond ciations, registered users of Web services, etc.), such what can be done in other modes (especially paper). -
Sampling Methods It’S Impractical to Poll an Entire Population—Say, All 145 Million Registered Voters in the United States
Sampling Methods It’s impractical to poll an entire population—say, all 145 million registered voters in the United States. That is why pollsters select a sample of individuals that represents the whole population. Understanding how respondents come to be selected to be in a poll is a big step toward determining how well their views and opinions mirror those of the voting population. To sample individuals, polling organizations can choose from a wide variety of options. Pollsters generally divide them into two types: those that are based on probability sampling methods and those based on non-probability sampling techniques. For more than five decades probability sampling was the standard method for polls. But in recent years, as fewer people respond to polls and the costs of polls have gone up, researchers have turned to non-probability based sampling methods. For example, they may collect data on-line from volunteers who have joined an Internet panel. In a number of instances, these non-probability samples have produced results that were comparable or, in some cases, more accurate in predicting election outcomes than probability-based surveys. Now, more than ever, journalists and the public need to understand the strengths and weaknesses of both sampling techniques to effectively evaluate the quality of a survey, particularly election polls. Probability and Non-probability Samples In a probability sample, all persons in the target population have a change of being selected for the survey sample and we know what that chance is. For example, in a telephone survey based on random digit dialing (RDD) sampling, researchers know the chance or probability that a particular telephone number will be selected. -
American Community Survey Accuracy of the Data (2018)
American Community Survey Accuracy of the Data (2018) INTRODUCTION This document describes the accuracy of the 2018 American Community Survey (ACS) 1-year estimates. The data contained in these data products are based on the sample interviewed from January 1, 2018 through December 31, 2018. The ACS sample is selected from all counties and county-equivalents in the United States. In 2006, the ACS began collecting data from sampled persons in group quarters (GQs) – for example, military barracks, college dormitories, nursing homes, and correctional facilities. Persons in sample in (GQs) and persons in sample in housing units (HUs) are included in all 2018 ACS estimates that are based on the total population. All ACS population estimates from years prior to 2006 include only persons in housing units. The ACS, like any other sample survey, is subject to error. The purpose of this document is to provide data users with a basic understanding of the ACS sample design, estimation methodology, and the accuracy of the ACS data. The ACS is sponsored by the U.S. Census Bureau, and is part of the Decennial Census Program. For additional information on the design and methodology of the ACS, including data collection and processing, visit: https://www.census.gov/programs-surveys/acs/methodology.html. To access other accuracy of the data documents, including the 2018 PRCS Accuracy of the Data document and the 2014-2018 ACS Accuracy of the Data document1, visit: https://www.census.gov/programs-surveys/acs/technical-documentation/code-lists.html. 1 The 2014-2018 Accuracy of the Data document will be available after the release of the 5-year products in December 2019.