Lesson 3: Sampling Plan 1. Introduction to Quantitative Sampling Sampling: Definition

Total Page:16

File Type:pdf, Size:1020Kb

Lesson 3: Sampling Plan 1. Introduction to Quantitative Sampling Sampling: Definition Quantitative approaches Quantitative approaches Plan Lesson 3: Sampling 1. Introduction to quantitative sampling 2. Sampling error and sampling bias 3. Response rate 4. Types of "probability samples" 5. The size of the sample 6. Types of "non-probability samples" 1 2 Quantitative approaches Quantitative approaches 1. Introduction to quantitative sampling Sampling: Definition Sampling = choosing the unities (e.g. individuals, famililies, countries, texts, activities) to be investigated 3 4 Quantitative approaches Quantitative approaches Sampling: quantitative and qualitative Population and Sample "First, the term "sampling" is problematic for qualitative research, because it implies the purpose of "representing" the population sampled. Population Quantitative methods texts typically recognize only two main types of sampling: probability sampling (such as random sampling) and Sample convenience sampling." (...) any nonprobability sampling strategy is seen as "convenience sampling" and is strongly discouraged." IIIIIIIIIIIIIIII Sampling This view ignores the fact that, in qualitative research, the typical way of IIIIIIIIIIIIIIII IIIII selecting settings and individuals is neither probability sampling nor IIIII convenience sampling." IIIIIIIIIIIIIIII IIIIIIIIIIIIIIII It falls into a third category, which I will call purposeful selection; other (= «!Miniature population!») terms are purposeful sampling and criterion-based selection." IIIIIIIIIIIIIIII This is a strategy in which particular settings, persons, or activieties are selected deliberately in order to provide information that can't be gotten as well from other choices." Maxwell , Joseph A. , Qualitative research design..., 2005 , 88 5 6 Quantitative approaches Quantitative approaches Population, Sample, Sampling frame Representative sample, probability sample Population = ensemble of unities from which the sample is Representative sample = Sample that reflects the population taken in a reliable way: the sample is a «!miniature population!» Sample = part of the population that is chosen for investigation. The choice may be based on Probability sample = Sample that has been randomly randomness or not. chosen. Therefore, every unity has a known probability to be chosen. Sampling frame = list of all the unities from which the choice is made. 7 8 Quantitative approaches Quantitative approaches Representativity: an empirical question 2. Sampling error, sampling bias The representativity of the sample cannot be assured by following a given method. If we use the correct methods (random choice, stratification etc.) we can only maximize the probability of producing a representative sample. It is an empirical question (and should be tested) if the sample is really representative of the population. For example: we would investigate if the percentage of women in the sample are not significantly different from those of the population (==> the sample is representative concerning gender). 9 10 Quantitative approaches Quantitative approaches Errors: different types Sampling error, sampling bias 1. Sampling error due to chance, size of sample Sampling error = Differences between the sample and the 2. Sampling bias not due to chance or size of population that are due to the sampling sample. E.g. non-response linked (the randomness). Sampling error can be to the specific theme of the diminished by increasing the size of the research sample 3. Data collection error e.g. bad question wording; bad interviewing Sampling bias = Differences between the sample and the 4. Data processing error e.g. wrong coding population that are not due to sampling 5. Data analysis error e.g. wrong statistical model; (the randomness); the sampling bias erroneous data analysis does not diminish with increased sample size. 6. Data interpretation error e.g. wrong interpretation of results 11 12 Quantitative approaches Quantitative approaches Sampling error/bias: example (I) Sampling error/bias: example (II) smokers non-smokers smokers non-smokers smokers non-smokers smokers non-smokers O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O Population : N = 200 Population : N = 200 Population : N = 200 Population : N = 200 Sample : N = 32 Sample : N = 32 Sample : N = 32 no error/bias a bit of error/bias a lot of error/bias P(s) = 0.5; p(s) = 0.5 P(s) = 0.5; p(s) = 0.47 P(s) = 0.5; p(s) = 0.33 13 14 Quantitative approaches Quantitative approaches Sampling error: decreases Possible reasons for sampling bias with increasing sample size Experiment with a coin • The sampling frame does not include all the elements of the Probability of throwing «!heads!»? population (example: telephone directory) • The choice is not really random (example: open telephone P «!in reality!» = 0.5 directory at a random page and choose the next 600 names) We do 5 tries with N =1,2,5,20 • Certain groups of respondents have a higher (lower) response rate (example: the very poor, the very rich, ther very active, With growing N, the p is approaching the P the people with an active interest in the question, the people critical of surveys) N = 1 -> p = 0, 1, 0, 1, 1 N = 2 -> p = 0, 0.5, 0.5, 1, 0 N = 5 -> p = 0.6, 0.2, 0.4, 0.8, 0.1 N = 20 -> p = 0.4, 0.35, 0.45, 0.35, 0.55 15 16 Quantitative approaches Quantitative approaches Sampling error vs.
Recommended publications
  • Combining Probability and Nonprobability Samples to Form Efficient Hybrid Estimates: an Evaluation of the Common Support Assumption Jill A
    Combining Probability and Nonprobability Samples to form Efficient Hybrid Estimates: An Evaluation of the Common Support Assumption Jill A. Dever RTI International, Washington, DC Proceedings of the 2018 Federal Committee on Statistical Methodology (FCSM) Research Conference Abstract Nonprobability surveys, those without a defined random sampling scheme, are becoming more prevalent. These studies can offer faster results at less cost than many probability surveys, especially for targeting important subpopulations. This can be an attractive option given the continual challenge of doing more with less, as survey costs continue to rise and response rates to plummet. Nonprobability surveys alone, however, may not fit the needs (purpose) of Federal statistical agencies where population inference is critical. Nonprobability samples may best serve to enhance the representativeness of certain domains within probability samples. For example, if locating and interviewing a required number of subpopulation members is resource prohibitive, data from a targeted nonprobability survey may lower coverage bias exhibited in a probability survey. In this situation, the question is how to best combine information from both sources. Our research searches for an answer to this question through an evaluation of hybrid estimation methods currently in use that combine probability and nonprobability data. Methods that employ generalized analysis weights (i.e., one set of weights for all analyses) are the focus because they enable other survey researchers and policy makers to analyze the data. The goal is to identify procedures that maximize the strength of each data source to produce hybrid estimates with the low mean square error. The details presented here focus on the propensity score adjusted (PSA) nonprobability weights needed prior to combining the data sources, and the common support assumption critical to hybrid estimation and PSA weighting.
    [Show full text]
  • A Critical Review of Studies Investigating the Quality of Data Obtained with Online Panels Based on Probability and Nonprobability Samples1
    Callegaro c02.tex V1 - 01/16/2014 6:25 P.M. Page 23 2 A critical review of studies investigating the quality of data obtained with online panels based on probability and nonprobability samples1 Mario Callegaro1, Ana Villar2, David Yeager3,and Jon A. Krosnick4 1Google, UK 2City University, London, UK 3University of Texas at Austin, USA 4Stanford University, USA 2.1 Introduction Online panels have been used in survey research as data collection tools since the late 1990s (Postoaca, 2006). The potential great cost and time reduction of using these tools have made research companies enthusiastically pursue this new mode of data collection. However, 1 We would like to thank Reg Baker and Anja Göritz, Part editors, for their useful comments on preliminary versions of this chapter. Online Panel Research, First Edition. Edited by Mario Callegaro, Reg Baker, Jelke Bethlehem, Anja S. Göritz, Jon A. Krosnick and Paul J. Lavrakas. © 2014 John Wiley & Sons, Ltd. Published 2014 by John Wiley & Sons, Ltd. Callegaro c02.tex V1 - 01/16/2014 6:25 P.M. Page 24 24 ONLINE PANEL RESEARCH the vast majority of these online panels were built by sampling and recruiting respondents through nonprobability methods such as snowball sampling, banner ads, direct enrollment, and other strategies to obtain large enough samples at a lower cost (see Chapter 1). Only a few companies and research teams chose to build online panels based on probability samples of the general population. During the 1990s, two probability-based online panels were documented: the CentER data Panel in the Netherlands and the Knowledge Networks Panel in the United States.
    [Show full text]
  • Options for Conducting Web Surveys Matthias Schonlau and Mick P
    Statistical Science 2017, Vol. 32, No. 2, 279–292 DOI: 10.1214/16-STS597 © Institute of Mathematical Statistics, 2017 Options for Conducting Web Surveys Matthias Schonlau and Mick P. Couper Abstract. Web surveys can be conducted relatively fast and at relatively low cost. However, Web surveys are often conducted with nonprobability sam- ples and, therefore, a major concern is generalizability. There are two main approaches to address this concern: One, find a way to conduct Web surveys on probability samples without losing most of the cost and speed advantages (e.g., by using mixed-mode approaches or probability-based panel surveys). Two, make adjustments (e.g., propensity scoring, post-stratification, GREG) to nonprobability samples using auxiliary variables. We review both of these approaches as well as lesser-known ones such as respondent-driven sampling. There are many different ways Web surveys can solve the challenge of gen- eralizability. Rather than adopting a one-size-fits-all approach, we conclude that the choice of approach should be commensurate with the purpose of the study. Key words and phrases: Convenience sample, Internet survey. 1. INTRODUCTION tion and invitation of sample persons to a Web sur- vey. No complete list of e-mail addresses of the general Web or Internet surveys1 have come to dominate the survey world in a very short time (see Couper, 2000; population exists from which one can select a sample Couper and Miller, 2008). The attraction of Web sur- and send e-mailed invitations to a Web survey. How- veys lies in the speed with which large numbers of ever, for many other important populations of interest people can be surveyed at relatively low cost, using (e.g., college students, members of professional asso- complex instruments that extend measurement beyond ciations, registered users of Web services, etc.), such what can be done in other modes (especially paper).
    [Show full text]
  • Chapter 3: Simple Random Sampling and Systematic Sampling
    Chapter 3: Simple Random Sampling and Systematic Sampling Simple random sampling and systematic sampling provide the foundation for almost all of the more complex sampling designs that are based on probability sampling. They are also usually the easiest designs to implement. These two designs highlight a trade-off inherent in all sampling designs: do we select sample units at random to minimize the risk of introducing biases into the sample or do we select sample units systematically to ensure that sample units are well- distributed throughout the population? Both designs involve selecting n sample units from the N units in the population and can be implemented with or without replacement. Simple Random Sampling When the population of interest is relatively homogeneous then simple random sampling works well, which means it provides estimates that are unbiased and have high precision. When little is known about a population in advance, such as in a pilot study, simple random sampling is a common design choice. Advantages: • Easy to implement • Requires little advance knowledge about the target population Disadvantages: • Imprecise relative to other designs if the population is heterogeneous • More expensive to implement than other designs if entities are clumped and the cost to travel among units is appreciable How it is implemented: • Select n sample units at random from N available in the population All units within the population must have the same probability of being selected, therefore each and every sample of size n drawn from the population has an equal chance of being selected. There are many strategies available for selecting a random sample.
    [Show full text]
  • R(Y NONRESPONSE in SURVEY RESEARCH Proceedings of the Eighth International Workshop on Household Survey Nonresponse 24-26 September 1997
    ZUMA Zentrum für Umfragen, Melhoden und Analysen No. 4 r(y NONRESPONSE IN SURVEY RESEARCH Proceedings of the Eighth International Workshop on Household Survey Nonresponse 24-26 September 1997 Edited by Achim Koch and Rolf Porst Copyright O 1998 by ZUMA, Mannheini, Gerinany All rights reserved. No part of tliis book rnay be reproduced or utilized in any form or by aiiy means, electronic or mechanical, including photocopying, recording, or by any inforniation Storage and retrieval System, without permission in writing froni the publisher. Editors: Achim Koch and Rolf Porst Publisher: Zentrum für Umfragen, Methoden und Analysen (ZUMA) ZUMA is a member of the Gesellschaft Sozialwissenschaftlicher Infrastruktureinrichtungen e.V. (GESIS) ZUMA Board Chair: Prof. Dr. Max Kaase Dii-ector: Prof. Dr. Peter Ph. Mohlcr P.O. Box 12 21 55 D - 68072.-Mannheim Germany Phone: +49-62 1- 1246-0 Fax: +49-62 1- 1246- 100 Internet: http://www.social-science-gesis.de/ Printed by Druck & Kopie hanel, Mannheim ISBN 3-924220-15-8 Contents Preface and Acknowledgements Current Issues in Household Survey Nonresponse at Statistics Canada Larry Swin und David Dolson Assessment of Efforts to Reduce Nonresponse Bias: 1996 Survey of Income and Program Participation (SIPP) Preston. Jay Waite, Vicki J. Huggi~isund Stephen 1'. Mnck Tlie Impact of Nonresponse on the Unemployment Rate in the Current Population Survey (CPS) Ciyde Tucker arzd Brian A. Harris-Kojetin An Evaluation of Unit Nonresponse Bias in the Italian Households Budget Survey Claudio Ceccarelli, Giuliana Coccia and Fahio Crescetzzi Nonresponse in the 1996 Income Survey (Supplement to the Microcensus) Eva Huvasi anci Acfhnz Marron The Stability ol' Nonresponse Rates According to Socio-Dernographic Categories Metku Znletel anci Vasju Vehovar Understanding Household Survey Nonresponse Through Geo-demographic Coding Schemes Jolin King Response Distributions when TDE is lntroduced Hikan L.
    [Show full text]
  • Lecture 8: Sampling Methods
    Lecture 8: Sampling Methods Donglei Du ([email protected]) Faculty of Business Administration, University of New Brunswick, NB Canada Fredericton E3B 9Y2 Donglei Du (UNB) ADM 2623: Business Statistics 1 / 30 Table of contents 1 Sampling Methods Why Sampling Probability vs non-probability sampling methods Sampling with replacement vs without replacement Random Sampling Methods 2 Simple random sampling with and without replacement Simple random sampling without replacement Simple random sampling with replacement 3 Sampling error vs non-sampling error 4 Sampling distribution of sample statistic Histogram of the sample mean under SRR 5 Distribution of the sample mean under SRR: The central limit theorem Donglei Du (UNB) ADM 2623: Business Statistics 2 / 30 Layout 1 Sampling Methods Why Sampling Probability vs non-probability sampling methods Sampling with replacement vs without replacement Random Sampling Methods 2 Simple random sampling with and without replacement Simple random sampling without replacement Simple random sampling with replacement 3 Sampling error vs non-sampling error 4 Sampling distribution of sample statistic Histogram of the sample mean under SRR 5 Distribution of the sample mean under SRR: The central limit theorem Donglei Du (UNB) ADM 2623: Business Statistics 3 / 30 Why sampling? The physical impossibility of checking all items in the population, and, also, it would be too time-consuming The studying of all the items in a population would not be cost effective The sample results are usually adequate The destructive nature of certain tests Donglei Du (UNB) ADM 2623: Business Statistics 4 / 30 Sampling Methods Probability Sampling: Each data unit in the population has a known likelihood of being included in the sample.
    [Show full text]
  • Final Abstracts in Order of Presentation
    Final Abstracts in Order of Presentation Sunday, September 20, 2015 9:30-11:30 a.m. Paper Session I Interactions of Survey Error and Ethnicity I Session Chair: Sunghee Lee Invited Presentation: Ethnic Minorities in Surveys: Applying the TSE Paradigm to Surveys Among Ethnic Minority Groups to Assess the Relationship Between Survey Design, Sample Frame and Survey Data Quality Joost Kappelhof1 Institute for Social Research/SCP1 Minority ethnic groups are difficult to survey mainly because of cultural differences, language barriers, socio-demographic characteristics and a high mobility (Feskens, 2009). As a result, ethnic minorities are often underrepresented in surveys (Groves & Couper, 1998; Stoop, 2005). At the same time, national and international policy makers need specific information about these groups, especially on issues such as socio-economic and cultural integration. Using the TSE framework, we will integrate existing international empirical literature on survey research among ethnic minorities. In particular, this paper will discuss four key topics in designing and evaluating survey research among ethnic minorities for policy makers. First of all, it discusses the reasons why ethnic minorities are underrepresented in survey. In this part an overview of the international empirical literature on reasons why it is difficult to conduct survey research among ethnic minorities will be placed in the TSE framework. Secondly, it reviews measures that can be undertaken to increase the representation of minorities in surveys and it discusses the consequences of these measures. In particular the relationship with survey design, sample frame and trade-off decisions in the TSE paradigm is discussed in combination with budget and time considerations.
    [Show full text]
  • Workshop on Probability-Based and Nonprobability Survey Research
    Workshop on Probability-Based and Nonprobability Survey Research Collaborative Research Center SFB 884 University of Mannheim June 25-26, 2018 Keynote: Jon A. Krosnick (Stanford University) Scientific Committee: Carina Cornesse Alexander Wenz Annelies Blom Location: SFB 884 – Political Economy of Reforms B6, 30-32 68131 Mannheim Room 008 (Ground Floor) Schedule Monday, June 25 08:30 – 09:10 Registration and coffee 09:10 – 09:30 Conference opening 09:30 – 10:30 Session 1: Professional Respondents and Response Quality o Professional respondents: are they a threat to probability-based online panels as well? (Edith D. de Leeuw) o Response quality in nonprobability and probability-based online panels (Carina Cornesse and Annelies Blom) 10:30 – 11:00 Coffee break 11:00 – 12:30 Session 2: Sample Accuracy o Comparing complex measurement instruments across probabilistic and non-probabilistic online surveys (Stefan Zins, Henning Silber, Tobias Gummer, Clemens Lechner, and Alexander Murray-Watters) o Comparing web nonprobability based surveys and telephone probability-based surveys with registers data: the case of Global Entrepreneurship Monitor in Luxembourg (Cesare A. F. Riillo) o Does sampling matter? Evidence from personality and politics (Mahsa H. Kashani and Annelies Blom) 12:30 – 13:30 Lunch 1 13:30 – 15:00 Session 3: Conceptual Issues in Probability-Based and Nonprobability Survey Research o The association between population representation and response quality in probability-based and nonprobability online panels (Alexander Wenz, Carina Cornesse, and Annelies Blom) o Probability vs. nonprobability or high-information vs. low- information? (Andrew Mercer) o Non-probability based online panels: market research practitioners perspective (Wojciech Jablonski) 15:00 – 15:30 Coffee break 15:30 – 17:00 Session 4: Practical Considerations in Online Panel Research o Replenishment of the Life in Australia Panel (Benjamin Phillips and Darren W.
    [Show full text]
  • Survey Design
    1 Survey Design Target population: The scope of ACES is to capture investment by all domestic, private, non-farm businesses, including agricultural non-farm business and businesses without employees. Investment made after applying for an Employer Identification Number (EIN) from the Internal Revenue Service (IRS) but before having any payroll or receipts is also included. Major exclusions are foreign operations of U.S. businesses, businesses in the U.S. territories, government operations (including the U.S. Postal Service), agricultural production companies and private households. Sampling frame: ACES collects information at the company level. The records of how the company invests are maintained at the headquarters level, and not at the location of each physical operating location. Companies may elect to have divisions within the company report, but the sampling unit and tabulation unit will be the company. A company’s importance to the survey depends on their employment, payroll, and their business activity. The greater the number of employees or the larger the payroll, the more likely, in general, a company is to be selected in the sample. The influence that the amount of payroll has on the likelihood of selection is adjusted by the business activity such that two companies with similar payroll but in different business activities will not have the same likelihood of selection. This is done to improve the quality of the estimates from any particular ACES specific industry code for business activities. Estimates are from two distinct samples from distinct frames. The first frame collects in-scope companies with employees. Companies sampled from this frame will receive an ACE-1 form, and the frame and sample are the ACE-1 frame and sample, respectively.
    [Show full text]
  • STANDARDS and GUIDELINES for STATISTICAL SURVEYS September 2006
    OFFICE OF MANAGEMENT AND BUDGET STANDARDS AND GUIDELINES FOR STATISTICAL SURVEYS September 2006 Table of Contents LIST OF STANDARDS FOR STATISTICAL SURVEYS ....................................................... i INTRODUCTION......................................................................................................................... 1 SECTION 1 DEVELOPMENT OF CONCEPTS, METHODS, AND DESIGN .................. 5 Section 1.1 Survey Planning..................................................................................................... 5 Section 1.2 Survey Design........................................................................................................ 7 Section 1.3 Survey Response Rates.......................................................................................... 8 Section 1.4 Pretesting Survey Systems..................................................................................... 9 SECTION 2 COLLECTION OF DATA................................................................................... 9 Section 2.1 Developing Sampling Frames................................................................................ 9 Section 2.2 Required Notifications to Potential Survey Respondents.................................... 10 Section 2.3 Data Collection Methodology.............................................................................. 11 SECTION 3 PROCESSING AND EDITING OF DATA...................................................... 13 Section 3.1 Data Editing ........................................................................................................
    [Show full text]
  • Unit 16: Census and Sampling
    Unit 16: Census and Sampling Summary of Video There are some questions for which an experiment can’t help us find the answer. For example, suppose we wanted to know what percentage of Americans smoke cigarettes, or what per- centage of supermarket chicken is contaminated with salmonella bacteria. There is no experi- ment that can be done to answer these types of questions. We could test every chicken on the market, or ask every person if they smoke. This is a census, a count of each and every item in a population. It seems like a census would be a straightforward way to get the most accurate, thorough information. But taking an accurate census is more difficult than you might think. The U.S. Constitution requires a census of the U.S population every ten years. In 2010, more than 308 million Americans were counted. However, the Census Bureau knows that some people are not included in this count. Undercounting certain segments of the population is a problem that can affect the representation given to a certain region as well as the federal funds it receives. What is particularly problematic is that not all groups are undercounted at the same rate. For example, the 2010 census had a hard time trying to reach renters. The first step in the U.S. Census is mailing a questionnaire to every household in the country. In 2010 about three quarters of the questionnaires were returned before the deadline. A census taker visits those households that do not respond by mail, but still not everyone is reached.
    [Show full text]
  • Second Stage Sampling for Conflict Areas: Methods and Implications Kristen Himelein, Stephanie Eckman, Siobhan Murray and Johannes Bauer 1
    Second Stage Sampling for Conflict Areas: Methods and Implications Kristen Himelein, Stephanie Eckman, Siobhan Murray and Johannes Bauer 1 Abstract: The collection of survey data from war zones or other unstable security situations is vulnerable to error because conflict often limits the options for implementation. Although there are elevated risks throughout the process, we focus here on challenges to frame construction and sample selection. We explore several alternative sampling approaches considered for the second stage selection of households for a survey in Mogadishu, Somalia. The methods are evaluated on precision, the complexity of calculations, the amount of time necessary for preparatory office work and the field implementation, and ease of implementation and verification. Unpublished manuscript prepared for the Annual Bank Conference on Africa on June 8 – 9, 2015 in Berkeley, California. Do not cite without authors’ permission. Acknowledgments: The authors would like to thank Utz Pape, from the World Bank, and Matthieu Dillais from Altai Consulting for their comments on this draft, and Hannah Mautner and Ruben Bach of IAB for their research assistance. 1 Kristen Himelein is a senior economist / statistician in the Poverty Global Practice at the World Bank. Stephanie Eckman is a senior researcher at the Institute for Employment Research (IAB) in Nuremberg, Germany. Siobhan Murray is a technical specialist in the Development Economics Research Group in the World Bank. Johannes Bauer is research fellow at the Institute for Sociology, Ludwigs-Maximilians University Munich and at the Institute for Employment Research (IAB). All views are those of the authors and do not reflect the views of their employers including the World Bank or its member countries.
    [Show full text]